Try it yourself
In this tutorial we will build DevDuck, our multi-agent system that combines Cerebras fast inference with Docker Compose for isolated AI development environments.1
Initialization
First, make sure that you have Docker installed, which you can download from the official website.
2
Clone the repository
Next, open a terminal and run the following commands:
3
Set up your environment
Once you’ve cloned the repository, create a
.env file with your Cerebras API key, which you can obtain from the Cerebras platform. DevDuck uses the ultra-fast gpt-oss-120b model through Cerebras for all agents.4
Running DevDuck
All that’s left to do is run the program. To build and start DevDuck, run:The compose setup spins up our agents and Docker’s MCP gateway which manages the MCP tools for working with the Node.js sandbox containers.
5
Using DevDuck
The application will be available at http://0.0.0.0:8000. DevDuck has a coordinator agent that manages two specialized Cerebras-powered sub-agents for different programming tasks.To get started, simply ask for a task and the program will automatically route to the correct agent. The DevDuck coordinator hands off to the appropriate Cerebras agent, which then executes your request in seconds.
Example Usage
🔢 Prime Number Checker
🔢 Prime Number Checker
🌊 Fibonacci Sequence Generator
🌊 Fibonacci Sequence Generator
📊 Array Statistics Calculator
📊 Array Statistics Calculator
💡 Quick Tips for Sandbox Execution
💡 Quick Tips for Sandbox Execution
- Use the word “EXECUTE” explicitly
- Keep functions simple and self-contained
- Provide specific test data
- Avoid external dependencies
Available Models
DevDuck uses Cerebras models for ultra-fast code generation:| Model | Parameters | Best For |
|---|---|---|
| llama-3.3-70b | 70B | Best for complex reasoning, long-form content, and tasks requiring deep understanding |
| qwen-3-32b | 32B | Balanced performance for general-purpose applications |
| llama3.1-8b | 8B | Fastest option for simple tasks and high-throughput scenarios |
| gpt-oss-120b | 120B | Largest model for the most demanding tasks (default in DevDuck) |
| zai-glm-4.7 | 357B | Advanced 357B parameter model with strong reasoning capabilities |
CEREBRAS_CHAT_MODEL environment variable to switch between models.
