Source Code:
src/gaia/cli.pyPlatform Support
Windows 11
Full GUI and CLI support with installer and desktop shortcuts
Linux
Full GUI and CLI support via source installation (Ubuntu/Debian)
Quick Start
- Windows
- Linux
- Follow the Getting Started Guide to install
gaiaCLI andlemonadeLLM server - Double click the GAIA-CLI desktop icon to launch the command-line shell
- GAIA automatically starts Lemonade Server when needed, or start manually:
Initialization
Init Command
New users start here! The
gaia init command is the easiest way to get GAIA running.| Option | Type | Default | Description |
|---|---|---|---|
--profile, -p | string | chat | Profile to initialize (minimal, chat, code, rag, all) |
--minimal | flag | false | Shortcut for --profile minimal |
--skip-models | flag | false | Skip model downloads (only install Lemonade) |
--skip-lemonade | flag | false | Skip Lemonade installation check (for CI with pre-installed Lemonade) |
--force-reinstall | flag | false | Force reinstall even if compatible version exists |
--force-models | flag | false | Force re-download models (deletes then re-downloads each model) |
--yes, -y | flag | false | Skip confirmation prompts (non-interactive) |
--verbose | flag | false | Enable verbose output |
--remote | flag | false | Use remote Lemonade server (skip local install/start, still checks version) |
| Profile | Models | Description | Approx Size |
|---|---|---|---|
minimal | Qwen3-0.6B | Fast setup with lightweight model | ~400 MB |
chat | Qwen3-Coder-30B, nomic-embed, Qwen3-VL-4B | Interactive chat with RAG and vision | ~25 GB |
code | Qwen3-Coder-30B | Autonomous coding assistant | ~18 GB |
rag | Qwen3-Coder-30B, nomic-embed, Qwen3-VL-4B | Document Q&A with retrieval and vision | ~25 GB |
all | All models | All models for all agents | ~26 GB |
All profiles also include the lightweight
Qwen3-0.6B model used by gaia llm for quick queries.- Checks Lemonade Server - Detects if installed and verifies version compatibility
- Installs/Upgrades Lemonade - Downloads and installs from GitHub releases (Windows/Linux only). Automatically uninstalls old version if version mismatch detected.
- Checks Server - Ensures Lemonade server is running (prompts to start if not)
- Downloads Models - Pulls required models for the selected profile
- Verifies Setup - Tests each model with inference to detect corrupted downloads
Install Command
Install individual GAIA components.| Option | Type | Description |
|---|---|---|
--lemonade | flag | Install Lemonade Server |
--yes, -y | flag | Skip confirmation prompts |
If a different version of Lemonade is already installed, you’ll be prompted to uninstall first.
Uninstall Command
Uninstall GAIA components.| Option | Type | Description |
|---|---|---|
--lemonade | flag | Uninstall Lemonade Server |
--models | flag | Delete all downloaded models from HuggingFace cache |
--yes, -y | flag | Skip confirmation prompts |
Kill Command
Stop running GAIA services.| Option | Type | Description |
|---|---|---|
--lemonade | flag | Kill Lemonade Server and child processes |
--port | integer | Kill process on specific port |
On Windows,
--lemonade also kills orphaned llama-server.exe and lemonade-tray.exe processes.Core Commands
LLM Direct Query
The fastest way to interact with AI models - no server management required.
| Option | Type | Default | Description |
|---|---|---|---|
--model | string | Client default | Specify the model to use |
--max-tokens | integer | 512 | Maximum tokens to generate |
--no-stream | flag | false | Disable streaming response |
Chat Command
Start an interactive conversation or send a single message with conversation history.- No message: Starts interactive chat session
- Message provided: Sends single message and exits
| Option | Type | Default | Description |
|---|---|---|---|
--query, -q | string | - | Single query to execute |
--model | string | Qwen3-Coder-30B-A3B-Instruct-GGUF | Model name to use |
--max-steps | integer | 10 | Maximum conversation steps |
--index, -i | path(s) | - | PDF document(s) to index for RAG |
--watch, -w | path(s) | - | Directories to monitor for new documents |
--chunk-size | integer | 500 | Document chunk size for RAG |
--max-chunks | integer | 3 | Maximum chunks to retrieve for RAG |
--stats | flag | false | Show performance statistics |
--streaming | flag | false | Enable streaming responses |
--show-prompts | flag | false | Display prompts sent to LLM |
--debug | flag | false | Enable debug output |
--list-tools | flag | false | List available tools and exit |
| Command | Description |
|---|---|
/clear | Clear conversation history |
/history | Show conversation history |
/system | Show current system prompt configuration |
/model | Show current model information |
/prompt | Show complete formatted prompt sent to LLM |
/stats | Show performance statistics (tokens/sec, latency, token counts) |
/help | Show available commands |
quit, exit, bye | End the chat session |
Prompt Command
Send a single prompt to a GAIA agent.| Option | Type | Default | Description |
|---|---|---|---|
--model | string | Qwen3-0.6B-GGUF | Model to use for the agent |
--max-tokens | integer | 512 | Maximum tokens to generate |
--stats | flag | false | Show performance statistics |
Specialized Agents
Code Agent
Code Development
AI-powered code generation, analysis, and linting for Python/TypeScript
- Intelligent Language Detection (Python/TypeScript)
- Code Generation (functions, classes, unit tests)
- Autonomous Workflow (planning → implementation → testing → verification)
- Automatic Test Generation
- Iterative Error Correction
- Code Analysis with AST
- Linting & Formatting
Blender Agent
3D Scene Creation
Natural language 3D modeling and scene manipulation
- Natural Language 3D Modeling
- Interactive Planning
- Object Management
- Material Assignment
- MCP Integration
SD Command
Image Generation
Generate images using Stable Diffusion on Ryzen AI
| Option | Type | Default | Description |
|---|---|---|---|
prompt | string | - | Text description of the image to generate |
-i, --interactive | flag | false | Run in interactive mode |
--sd-model | string | SD-Turbo | Model: SD-Turbo (fast, default), SDXL-Turbo, SDXL-Base-1.0 (photorealistic), SD-1.5 |
--size | string | auto | Image size: 512x512, 768x768, 1024x1024 (auto-selected per model) |
--steps | integer | auto | Inference steps (auto: 4 for Turbo, 20 for Base) |
--cfg-scale | float | auto | CFG scale (auto: 1.0 for Turbo, 7.5 for Base) |
--output-dir | path | .gaia/cache/sd/images | Directory to save images |
--seed | integer | random | Seed for reproducibility |
--no-open | flag | false | Skip prompt to open image in viewer |
Talk Command
Voice Interaction
Speech-to-speech conversation with optional document Q&A
| Option | Type | Default | Description |
|---|---|---|---|
--model | string | Qwen3-0.6B-GGUF | Model to use |
--max-tokens | integer | 512 | Maximum tokens to generate |
--no-tts | flag | false | Disable text-to-speech |
--audio-device-index | integer | auto-detect | Audio input device index |
--whisper-model-size | string | base | Whisper model [tiny, base, small, medium, large] |
--silence-threshold | float | 0.5 | Silence threshold in seconds |
--stats | flag | false | Show performance statistics |
--index, -i | path | - | PDF document for voice Q&A |
API Server
API Server
OpenAI-compatible REST API for VSCode and IDE integrations
Quick Start
- Start Lemonade with extended context:
- Start GAIA API server:
- Test the server:
Commands
- Start
- Status
- Stop
--host- Server host (default: localhost)--port- Server port (default: 8080)--background- Run in background--debug- Enable debug logging
MCP Client
MCP Client
Connect GAIA agents to external MCP servers
~/.gaia/mcp_servers.json by default, or to a custom config file using --config.
Commands
gaia mcp add
Add an MCP server to configuration.
<server-name>- Unique identifier for the server (e.g., “time”, “memory”)"<command>"- Shell command to start the MCP server (must be quoted)
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
gaia mcp list
List all configured MCP servers.
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
gaia mcp remove
Remove an MCP server from configuration.
<server-name>- Name of the server to remove
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
gaia mcp tools
List tools available from a configured MCP server.
<server-name>- Name of the server to query
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
gaia mcp test-client
Test connection to a configured MCP server.
<server-name>- Name of the server to test
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
MCP Bridge
MCP Bridge
Expose GAIA agents as MCP servers
Quick Start
Install MCP support:Commands
| Command | Description |
|---|---|
start | Start the MCP bridge server |
status | Check MCP server status |
stop | Stop background MCP bridge server |
test | Test MCP bridge functionality |
agent | Test MCP orchestrator agent |
docker | Start Docker MCP server |
Model Management
Download Command
Download all models required for GAIA agents with streaming progress.| Option | Type | Default | Description |
|---|---|---|---|
--agent | string | all | Agent to download models for |
--list | flag | false | List required models without downloading |
--timeout | integer | 1800 | Timeout per model in seconds |
--host | string | localhost | Lemonade server host |
--port | integer | 8000 | Lemonade server port |
Pull Command
To download individual models, use the Lemonade Server CLI directly:Evaluation Commands
Evaluation Framework
Systematic testing, benchmarking, and model comparison
- Ground Truth Generation
- Automated Evaluation
- Batch Experimentation
- Performance Analysis
- Transcript Testing
Visualize Command
Launch interactive web-based visualizer for comparing evaluation results.| Option | Type | Default | Description |
|---|---|---|---|
--port | integer | 3000 | Visualizer server port |
--experiments-dir | path | ./output/experiments | Experiments directory |
--evaluations-dir | path | ./output/evaluations | Evaluations directory |
--workspace | path | current directory | Base workspace directory |
--no-browser | flag | false | Don’t auto-open browser |
--host | string | localhost | Host address |
- Interactive Comparison (side-by-side)
- Key Metrics Dashboard
- Quality Analysis
- Real-time Updates
- Responsive Design
Node.js must be installed. Dependencies are automatically installed on first run.
Utility Commands
Stats Command
View performance statistics from the most recent model run.Test Commands
Run various tests for development and troubleshooting.- TTS Tests
- ASR Tests
Test Types:Test streaming:Generate audio file:
tts-preprocessing- Test TTS text preprocessingtts-streaming- Test TTS streaming playbacktts-audio-file- Test TTS audio file generation
--test-text- Text to use for TTS tests--output-audio-file- Output file path (default: output.wav)
YouTube Utilities
Download transcripts from YouTube videos.--download-transcript- YouTube URL to download transcript from--output-path- Output file path (defaults to transcript_.txt)
Kill Command
Terminate processes running on specific ports.| Option | Type | Description |
|---|---|---|
--port | integer | Port number to kill process on |
--lemonade | flag | Kill Lemonade server (port 8000) |
- Find the process ID (PID) bound to the specified port
- Forcefully terminate that process
- Provide feedback about success or failure
Global Options
All commands support these global options:| Option | Type | Default | Description |
|---|---|---|---|
--logging-level | string | INFO | Logging verbosity [DEBUG, INFO, WARNING, ERROR, CRITICAL] |
-v, --version | flag | - | Show program’s version and exit |
Troubleshooting
Connection Errors
Connection Errors
If you get connection errors, ensure Lemonade server is running:
Model Issues
Model Issues
Check available system memory (16GB+ recommended)Verify model compatibility:Pre-download models:Install additional models: See Features Guide
Audio Issues
Audio Issues
List available devices:Verify microphone permissions in Windows settingsTry different audio device indices if default doesn’t work
Performance
Performance
For optimal NPU performance:
- Disable discrete GPUs in Device Manager
- Ensure NPU drivers are up to date
- Monitor system resources during execution
See Also
Code Agent
Python/TypeScript development
Blender Agent
3D scene creation
Voice Interaction
Speech-to-speech conversation
API Server
OpenAI-compatible REST API
MCP Integration
Model Context Protocol
Evaluation Framework
Testing and benchmarking