Lilypad is developing a serverless, distributed compute network that enables internet-scale data processing, AI, ML & other arbitrary computation from blockchains, while unleashing idle processing power & unlocking a new marketplace for compute.
- FAQs
- Lilypad docs
- Quick Start - Lilypad "Hello World"
- Module Boilerplate - A boilerplate for getting started with building a Lilypad module
- Lilypad Discord - Come ask questions!
- AI Model Marketplace
- Local JS CLI Wrapper API - Run jobs on the Lilypad network with a locally hosted API.
- Smart contract jobs - Trigger jobs on the Lilypad network from a smart contract
- Run Lilypad on a client - Integrate Lilypad into a frontend to run a SDXL module to generate images
- create-lilypad-module - Scaffolding tool for building Lilypad modules
A list of frontend applications that utilize Lilypad Network, provided by the Lilypad team and community.
App | Description |
---|---|
Sir Croaksworth's Roast DApp | Humorous web application that roasts users' cryptocurrency transactions. |
A list of modules that can run on the Lilypad Network, provided by the Lilypad team and community.
Modules supported by the Lilypad API are also compatible with our CLI.
Module | Model | Repo |
---|---|---|
DeepScaleR 1.5B | "deepscaler:1.5b" |
github.com/DevlinRocha/lilypad-deepscaler-1.5b:c078cab5d190fb6ea0e3ea92a33b74935fa7c4bb |
DeepSeek R1 7B | "deepseek-r1:7b" |
github.com/DevlinRocha/lilypad-deepseek-r1-7b:9775e21b351420be521080b951f833dabcacc5c5 |
Llava 7B | "llava:7b" |
github.com/DevlinRocha/lilypad-llava-7b:d85058d4335acb5f7f09d139eca9a009ec69df02 |
Mistral 7B | "mistral:7b" |
github.com/DevlinRocha/lilypad-mistral-7b:aac363a81acc8774414dd18a3702fb81af5171a9 |
OpenThinker 7B | "openthinker:7b" |
github.com/DevlinRocha/lilypad-openthinker-7b:97a48dfbdbba14a61557691a5ad0809313e26535 |
Phi-4 14B | "phi4:14b" |
github.com/DevlinRocha/lilypad-phi4-14b:ac0008abbd8d50fc3183984672141c22143aaf92 |
Phi-4 Mini 3.8B | "phi4-mini:3.8b" |
github.com/DevlinRocha/lilypad-phi4-mini-3.8b:b6d4033ce009eb35686cd18d9cc48501afac774c |
Qwen2.5 7B | "qwen2.5:7b" |
github.com/DevlinRocha/lilypad-qwen2.5-7b:8d50d99731215350218097ad3313f668ebde8933 |
Qwen2.5 Coder 7B | "qwen2.5-coder:7b" |
github.com/DevlinRocha/lilypad-qwen2.5-coder-7b:c5ea2904d2ec02d397de16dc96f74e0ff4b0d9b8 |
Use With Our API
- Replace
YOUR_API_KEY
with your API key from the Anura website. - Replace the
model
field value of your request object with the "Model" column from our modules table below.
curl -X POST "https://anura-testnet.lilypad.tech/api/v1/chat/completions" \
-H "Content-Type: application/json" \
-H "Accept: text/event-stream" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "MODEL_NAME:MODEL_VERSION",
"messages": [{
"role": "system",
"content": "you are a helpful AI assistant"
},
{
"role": "user",
"content": "what order do frogs belong to?"
}],
"temperature": 0.6
}'
Use With Our CLI
- Replace
GITHUB_USERNAME/MODULE_REPO:TAG
with the "Repo" column from our modules table below. - Replace the
model
field value of your request object with the "Model" column from our modules table below.
When using our CLI, make sure that you Base64 encode your request.
lilypad run github.com/GITHUB_USERNAME/MODULE_REPO:TAG \
-i request="$(echo -n '{
"model": "MODEL_NAME:MODEL_VERSION",
"messages": [{
"role": "system",
"content": "you are a helpful AI assistant"
},
{
"role": "user",
"content": "what order do frogs belong to?"
}],
"temperature": 0.6
}' | base64 -w 0)"
Field | Description | Type |
---|---|---|
model |
Model ID used to generate the response (e.g. deepseek-r1:7b). Required. | string |
messages |
A list of messages comprising the conversation so far. Required. | array |
Optional Fields
Our API modules support the following optional fields for the request.
Field | Description | Default |
---|---|---|
frequency_penalty |
Number between -2.0 and 2.0 . Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim. |
0 |
max_tokens |
The maximum number of tokens that can be generated in the chat completion. | |
presence_penalty |
Number between -2.0 and 2.0 . Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics. |
0 |
response_format |
An object specifying the format that the model must output. Learn more. | |
seed |
Makes a best effort to sample deterministically, such that repeated requests with the same seed and parameters should return the same result. Determinism is not guaranteed, and you should refer to the system_fingerprint response parameter to monitor changes in the backend. |
|
stop |
Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence. | null |
stream |
If set to true, the model response data will be streamed to the client as it is generated using server-sent events. | false |
stream_options |
Options for streaming response. Only set this when you set stream: true . Learn more. |
null |
temperature |
What sampling temperature to use, between 0 and 2 . Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. We recommend altering this or top_p but not both. |
1 |
top_p |
An alternative to sampling with temperature , called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We recommend altering this or temperature but not both. |
1 |
tools |
A list of tools the model may call. Currently, only functions are supported as a tool. Use this to provide a list of functions the model may generate JSON inputs for. A max of 128 functions are supported. Learn more. |
Modules created by the Lilypad team are listed in the Lilypad docs under "Lilypad Modules".
Module | Run Command | Arguments |
---|---|---|
lilysay | lilypad run lilysay:0.5.2 |
-i Message="" |
cowsay | lilypad run cowsay:v0.0.4 |
-i Message="" |
Stable Diffusion Turbo Pipeline | lilypad run github.com/noryev/module-sdxl-ipfs:main |
-i prompt="" |
DeepSeek R1 1.5B | lilypad run github.com/narbs91/lilypad-ollama-deepseek-r1-1-5b:main |
-i Prompt="" |
Llama 2 | lilypad run github.com/noryev/module-llama2:main |
-i PROMPT="" |
Modules from the community have not been tested by the Lilypad team. Run these modules at your own risk.
Module | Run Command | Arguments |
---|---|---|
DeepSeek R1 7B | lilypad run github.com/rhochmayr/ollama-deepseek-r1-7b:1.0.0 |
-i Prompt="" |
Llama 3.2 3B | lilypad run github.com/rhochmayr/ollama-llama3.2-3b:1.0.0 |
-i PROMPT="" |
Qwen 2.5 Coder 7B | lilypad run github.com/rhochmayr/ollama-qwen2.5-coder-7b:1.0.0 |
-i Prompt="" |
Spark-TTS-0.5B | lilypad run github.com/jamiebones/Spark-TTS-0.5B.git:v1.6 |
-i input="" |
- AI learning buddy - an AI learning assistant using RAG and the Lilypad Anura inference API
- AI Oncologist Agent - AI agent for oncology research and diagnostics
- Llama3 Chatbot - AI conversations powered by Llama 3 on Lilypad
- RAG Support Agent - AI-powered agent that retrieves and summarizes relevant information
- VS Code Helper Extension - AI-powered code assistance
- Lilypad Node Monitoring - Node Monitoring with Uptime Kuma and Telegram Notifications
- Waterlily.ai - An ethical generative AI-Art dApp leveraging FVM smart contracts and decentralized edge compute platform Bacalhau
- Defikicks - A decentralized, community-governed Data DAO on Filecoin that democratizes DeFi data aggregation and TVL calculations
- Lilypad Kamu - Custom Lilypad module for SQL streaming through kamu.dev
- Project C - Personalized course generation using Lilypad
- AI Capwyn - ML tooling for security audits
- Lilypad Hub - Hub for Lilypad modules
- Obsidian Lilypad - Pull Desci nodes, stream together external apis and more
- Decenter AI - PaaS infrastructure for training cost effective AI models through decentralized parallel training methods
- Lilylatte - DataDAO focused on web3 market research
- TinyHops - A workflow specification and runner for executing Lilypad enabled jobs in sequence
- Saturn Observatory - Monthly analytical reports using historical data from the Filecoin Saturn network
- Lilypad JavaScript Wrapper - JavaScript wrapper for interacting with the Lilypad CLI using Node.js
- Lilywrite - AI generated poems using Lilypad
- Rejuvenate AI - Community blockchain based project built to promote healthy living and achieve healthy locations
- GreenBadge - Onchain funding platform
- Gadus CLI - CLI package for running 3D surface reconstruction
- Deehr Market - Provides patients with ownership of their electronic health records
- CréditDécentrale - Decentralized credit score system using zkSNARKs, Decentralized Computation and Storage
- Cypher Deposit - Secure crypto withdrawals via anonymous transfers
- LLMBench - Verifiable on-chain Large Language Models drift benchmarking
- Uncensored GPT-5 - Uncensored 100% decentralized GPT AI chat running on the blockchain
- ZK Microphone - Trusted audio using zk-SNARKS
- CarpAI - Fastchat LLM inference module for Lilypad v1
- Brian - Decentralized web3 AI assistant for executing onchain transactions via prompt
- Avasoul.ai - Private-Model Generation and Service Platform
- EasyCraft AI - AI and blockchain based supply chain management implementation matching factory capacity with customer orders for economic efficiency
- Decentralized Yield Data Collector - Decentralized yield data aggregator
- Pensieve - Decentralized file storage
- Daggle - A swiss knife for Bacalhau, everything you need in an easily accessible dashboard
- InferAI - Single click deployment for ML models on a decentralized compute and storage powered by Filecoin Virtual Machine, Bacalhau & Libp2p
- TENTAI - SDK for AI models running on decentralized computing
- Rock Paper Ninja - A web3 take on the classic game of Rock Paper Scissors
- Dogtoken ai erc404 - ERC404 token with custom Lilypad stable diffusion to generate unique images for holders
- Filecoin ML Engine - Onchain ML engine to create dApp ML agents on top of Filecoin
- NexTown - A decentralized AI computation and storage platform allowing users to run AI models via LilyPad
- CipherCraft - A decentralized hub for federated model training on access controlled private datasets