Source Code:
src/gaia/cli.pyPlatform Support
Windows 11
Full GUI and CLI support
Linux
Full GUI and CLI support via source installation (Ubuntu/Debian)
Quick Start
- Windows
- Linux
- Follow the Quickstart to install GAIA
- Open PowerShell and run
gaiato launch the Agent UI, orgaia --clifor terminal chat - GAIA automatically starts Lemonade Server when needed, or start manually:
Top-Level Flags
Runninggaia with no subcommand launches the Agent UI by default:
| Flag | Description |
|---|---|
--ui | Launch the Agent UI (browser-based chat interface) — this is the default |
--ui-port <port> | Port for the Agent UI server (default: 4200) |
--ui-dist <path> | Serve an alternative prebuilt Agent UI bundle (e.g., a PR preview) instead of the one shipped in the installed package |
--cli | Launch interactive CLI chat |
--base-url <url> | Use a remote Lemonade server (e.g. https://host:8000/api/v1). Also accepted as a global option on most subcommands (e.g. gaia chat --base-url ...). |
gaia chat --ui continues to work as an alias. The Agent UI requires an AMD Ryzen AI Max (Strix Halo) or an AMD Radeon GPU with ≥ 24 GB VRAM. If your device is not supported, a dismissible warning banner will appear in the UI.Initialization
Init Command
New users start here! The
gaia init command is the easiest way to get GAIA running.| Option | Type | Default | Description |
|---|---|---|---|
--profile, -p | string | chat | Profile to initialize (minimal, sd, chat, code, rag, mcp, vlm, all) |
--minimal | flag | false | Shortcut for --profile minimal |
--skip-models | flag | false | Skip model downloads (only install Lemonade) |
--skip-lemonade | flag | false | Skip Lemonade installation check (for CI with pre-installed Lemonade) |
--force-reinstall | flag | false | Force reinstall even if compatible version exists |
--force-models | flag | false | Force re-download models (deletes then re-downloads each model) |
--yes, -y | flag | false | Skip confirmation prompts (non-interactive) |
--verbose | flag | false | Enable verbose output |
--remote | flag | false | Use remote Lemonade Server (skip local install/start, download models via API). Auto-detected when LEMONADE_BASE_URL is set to a non-localhost URL. |
--profile flag accepts one of the following values (see
src/gaia/cli.py for the
canonical choices list):
| Profile | Models | Description | Approx Size |
|---|---|---|---|
minimal | Qwen3-0.6B | Fast setup with lightweight model | ~400 MB |
sd | SD-Turbo | Image generation with Stable Diffusion | ~8 GB |
chat | Qwen3.5-35B, nomic-embed, Qwen3-VL-4B | Interactive chat with RAG and vision | ~25 GB |
code | Qwen3.5-35B | Autonomous coding assistant | ~18 GB |
rag | Qwen3.5-35B, nomic-embed, Qwen3-VL-4B | Document Q&A with retrieval and vision | ~25 GB |
mcp | Qwen3.5-35B, nomic-embed, Qwen3-VL-4B | Model Context Protocol bridge server with vision | ~25 GB |
vlm | Qwen3-VL-4B | Image understanding and analysis | ~3 GB |
all | All models | All models for all agents | ~26 GB |
The
talk, blender, jira, and docker agents all rely on the chat profile
(Qwen3.5-35B family) — there is no dedicated profile for each. Run
gaia init --profile chat (or all) to prepare their dependencies.All profiles also include the lightweight
Qwen3-0.6B model used by gaia llm for quick queries.- Local — Lemonade runs on the same machine (default)
- Remote — Lemonade runs on another machine; enable with
--remoteor by settingLEMONADE_BASE_URL
- Local Mode
- Remote Mode
- Checks Lemonade Server - Detects if installed and verifies version compatibility
- Installs/Upgrades Lemonade - Downloads and installs from GitHub releases (Windows/Linux only). Automatically uninstalls old version if version mismatch detected.
- Starts Server - Ensures Lemonade server is running, prompts to start if not
- Downloads Models - Pulls required models for the selected profile
- Verifies Setup - Tests each model with inference to detect corrupted downloads
Install Command
Install individual GAIA components.| Option | Type | Description |
|---|---|---|
--lemonade | flag | Install Lemonade Server |
--yes, -y | flag | Skip confirmation prompts |
If a different version of Lemonade is already installed, you’ll be prompted to uninstall first.
Uninstall Command
Tiered cleanup of GAIA components. By default, the OS-native uninstall (Add or Remove Programs, drag to Trash,apt remove) only removes the app — user data in ~/.gaia/ is preserved. This command lets you escalate cleanup as far as you want.
| Option | Type | Description |
|---|---|---|
--venv | flag | Remove the Python environment (~/.gaia/venv/) |
--purge | flag | Remove venv + chats + documents + config + logs |
--purge-lemonade | flag | Also remove Lemonade Server (requires --purge) |
--purge-models | flag | Also remove downloaded Lemonade models (requires --purge) |
--purge-hf-cache | flag | Also remove ~/.cache/huggingface/ (requires --purge) |
--dry-run | flag | Show what would be removed without deleting anything |
--yes, -y | flag | Skip confirmation prompts |
Kill Command
Stop running GAIA services.| Option | Type | Description |
|---|---|---|
--lemonade | flag | Kill Lemonade Server and child processes |
--port | integer | Kill process on specific port |
On Windows,
--lemonade also kills orphaned llama-server.exe and lemonade-tray.exe processes.Agent Command
Manage custom agents installed under~/.gaia/agents/. Supports exporting every custom agent into a single .zip bundle and importing a bundle produced on another machine.
| Option | Type | Description |
|---|---|---|
--output <path> | string | Destination .zip file (default: ~/.gaia/export.zip) |
| Option | Type | Description |
|---|---|---|
path | string | Path to the .zip bundle to import (positional, required) |
--yes, -y | flag | Skip the interactive trust prompt (required in non-interactive/CI contexts) |
Importing a bundle runs third-party Python code on your machine.
gaia agent import shows the agent IDs in the bundle and requires explicit y/yes confirmation (or --yes) before proceeding.Core Commands
LLM Direct Query
The fastest way to interact with AI models - no server management required.
| Option | Type | Default | Description |
|---|---|---|---|
--model | string | Client default | Specify the model to use |
--max-tokens | integer | 512 | Maximum tokens to generate |
--no-stream | flag | false | Disable streaming response |
Chat Command
Start an interactive conversation or send a single message with conversation history.- No message: Starts interactive chat session
- Message provided: Sends single message and exits
| Option | Type | Default | Description |
|---|---|---|---|
--query, -q | string | - | Single query to execute |
--model | string | auto-selected by agent | Override the model used by ChatAgent (None means let the agent pick; see ChatAgentConfig.model_id) |
--index, -i | path(s) | - | PDF document(s) to index for RAG |
--watch, -w | path(s) | - | Directories to monitor for new documents |
--chunk-size | integer | 500 | Document chunk size for RAG |
--max-chunks | integer | 3 | Maximum chunks to retrieve for RAG |
--max-indexed-files | integer | 100 | Maximum number of files to keep indexed. Evicts least-recently-accessed files when LRU eviction is enabled; rejects new files when disabled. |
--max-total-chunks | integer | 10000 | Maximum total chunks across all indexed files. Same eviction behavior as --max-indexed-files. |
--allowed-paths | path(s) | - | Restrict RAG/file tools to these directories (security sandbox) |
--stats, --show-stats | flag | false | Show performance statistics |
--stream | flag | false | Enable streaming responses |
--show-prompts | flag | false | Display prompts sent to LLM |
--debug | flag | false | Enable debug output |
--list-tools | flag | false | List available tools and exit |
--ui | flag | false | Launch the Chat Web UI (browser-based interface on port 4200) |
--ui-port | integer | 4200 | Port for the Agent UI server (used with --ui) |
--ui-dist | path | - | Serve an alternative prebuilt Agent UI bundle (e.g., a PR preview) instead of the one shipped in the package |
| Command | Description |
|---|---|
/clear | Clear conversation history |
/history | Show conversation history |
/system | Show current system prompt configuration |
/model | Show current model information |
/prompt | Show complete formatted prompt sent to LLM |
/stats | Show performance statistics (tokens/sec, latency, token counts) |
/help | Show available commands |
quit, exit, bye | End the chat session |
Prompt Command
Send a single prompt to a GAIA agent.| Option | Type | Default | Description |
|---|---|---|---|
--model | string | Qwen3-0.6B-GGUF | Model to use for the agent |
--max-tokens | integer | 512 | Maximum tokens to generate |
--stats | flag | false | Show performance statistics |
Specialized Agents
Code Agent
Code Development
AI-powered code generation, analysis, and linting for Python/TypeScript
- Intelligent Language Detection (Python/TypeScript)
- Code Generation (functions, classes, unit tests)
- Autonomous Workflow (planning → implementation → testing → verification)
- Automatic Test Generation
- Iterative Error Correction
- Code Analysis with AST
- Linting & Formatting
Blender Agent
3D Scene Creation
Natural language 3D modeling and scene manipulation
- Natural Language 3D Modeling
- Interactive Planning
- Object Management
- Material Assignment
- MCP Integration
SD Command
Image Generation
Generate images using Stable Diffusion on Ryzen AI
| Option | Type | Default | Description |
|---|---|---|---|
prompt | string | - | Text description of the image to generate |
-i, --interactive | flag | false | Run in interactive mode |
--sd-model | string | SD-Turbo | Model: SD-Turbo (fast, default), SDXL-Turbo, SDXL-Base-1.0 (photorealistic), SD-1.5 |
--size | string | auto | Image size: 512x512, 768x768, 1024x1024 (auto-selected per model) |
--steps | integer | auto | Inference steps (auto: 4 for Turbo, 20 for Base) |
--cfg-scale | float | auto | CFG scale (auto: 1.0 for Turbo, 7.5 for Base) |
--output-dir | path | .gaia/cache/sd/images | Directory to save images |
--seed | integer | random | Seed for reproducibility |
--no-open | flag | false | Skip prompt to open image in viewer |
Talk Command
Voice Interaction
Speech-to-speech conversation with optional document Q&A
| Option | Type | Default | Description |
|---|---|---|---|
--model | string | Qwen3-0.6B-GGUF | Model to use |
--max-tokens | integer | 512 | Maximum tokens to generate |
--no-tts | flag | false | Disable text-to-speech |
--audio-device-index | integer | auto-detect | Audio input device index |
--whisper-model-size | string | base | Whisper model [tiny, base, small, medium, large] |
--silence-threshold | float | 0.5 | Silence threshold in seconds |
--mic-threshold | float | 0.003 | Microphone amplitude threshold for voice detection (lower = more sensitive) |
--stats | flag | false | Show performance statistics |
--index, -i | path | - | PDF document for voice Q&A |
API Server
API Server
OpenAI-compatible REST API for VSCode and IDE integrations
Quick Start
- Start Lemonade with extended context:
- Start GAIA API server:
- Test the server:
Commands
- Start
- Status
- Stop
--host- Server host (default: localhost)--port- Server port (default: 8080).--background- Run in background--debug- Enable debug logging--show-prompts- Log the prompts sent to the LLM for every request (useful for debugging)--streaming- Stream tokens to clients via SSE (OpenAI-style)--step-through- Pause between agent steps for manual inspection (development aid)
MCP Client
MCP Client
Connect GAIA agents to external MCP servers
~/.gaia/mcp_servers.json by default, or to a custom config file using --config.
Commands
gaia mcp add
Add an MCP server to configuration.
<server-name>- Unique identifier for the server (e.g., “time”, “memory”)"<command>"- Shell command to start the MCP server (must be quoted)
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
gaia mcp list
List all configured MCP servers.
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
gaia mcp remove
Remove an MCP server from configuration.
<server-name>- Name of the server to remove
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
gaia mcp tools
List tools available from a configured MCP server.
<server-name>- Name of the server to query
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
gaia mcp test-client
Test connection to a configured MCP server.
<server-name>- Name of the server to test
--config PATH- Custom config file path (default:~/.gaia/mcp_servers.json)
MCP Bridge
MCP Bridge
Expose GAIA agents as MCP servers
Quick Start
Install MCP support:Commands
| Command | Description |
|---|---|
start | Start the MCP bridge server |
status | Check MCP server status |
stop | Stop background MCP bridge server |
test | Test MCP bridge functionality (optionally against a specific --tool) |
agent | Invoke the MCP orchestrator agent (positional request arg + --domain, --context) |
docker | Start the Docker MCP server (default port 8080) |
gaia mcp start options
| Option | Default | Description |
|---|---|---|
--host | localhost | Bind address |
--port | 8765 | Port for the MCP bridge |
--auth-token | none | If set, require Authorization: Bearer <token> on every request |
--no-streaming | off | Disable SSE streaming; reply synchronously |
--background | off | Detach; use gaia mcp stop to terminate |
--log-file | stdout | Write logs to this path (useful with --background) |
--verbose | off | Verbose logging |
--ctx-size | 32768 | Context window hint passed to Lemonade for loaded models |
gaia mcp test options
| Option | Default | Description |
|---|---|---|
--query | required | Text to send to the server |
--tool | gaia.chat | Which MCP tool to invoke when testing |
gaia mcp agent
request and, optionally, a --domain (e.g. jira,
docker) and free-form --context string to steer the orchestrator.
gaia mcp docker options
--port defaults to 8080.
→ Full MCP Integration Guide
Model Management
Download Command
Download all models required for GAIA agents with streaming progress.| Option | Type | Default | Description |
|---|---|---|---|
--agent | string | all | Agent to download models for |
--list | flag | false | List required models without downloading |
--timeout | integer | 1800 | Timeout per model in seconds |
--host | string | localhost | Lemonade server host |
--port | integer | 8000 | Lemonade server port |
--clear-cache | flag | false | Remove cached download progress/metadata before running |
Pull Command
To download individual models, use the Lemonade Server CLI directly:Evaluation Commands
Evaluation Framework
Systematic testing, benchmarking, and model comparison
- Ground Truth Generation
- Automated Evaluation
- Batch Experimentation
- Performance Analysis
- Transcript Testing
Visualize Command
Launch interactive web-based visualizer for comparing evaluation results.| Option | Type | Default | Description |
|---|---|---|---|
--port | integer | 3000 | Visualizer server port |
--experiments-dir | path | ./output/experiments | Experiments directory |
--evaluations-dir | path | ./output/evaluations | Evaluations directory |
--workspace | path | current directory | Base workspace directory |
--no-browser | flag | false | Don’t auto-open browser |
--host | string | localhost | Host address |
- Interactive Comparison (side-by-side)
- Key Metrics Dashboard
- Quality Analysis
- Real-time Updates
- Responsive Design
Node.js must be installed. Dependencies are automatically installed on first run.
Utility Commands
Stats Command
View performance statistics from the most recent model run.Test Commands
Run various tests for development and troubleshooting.- TTS Tests
- ASR Tests
Test Types:Test streaming:Generate audio file:
tts-preprocessing- Test TTS text preprocessingtts-streaming- Test TTS streaming playbacktts-audio-file- Test TTS audio file generation
--test-text- Text to use for TTS tests--output-audio-file- Output file path (default: output.wav)
YouTube Utilities
Download transcripts from YouTube videos.--download-transcript- YouTube URL to download transcript from--output-path- Output file path (defaults to transcript_.txt)
Cache Command
Inspect or clear GAIA’s on-disk caches (document-Q7 metadata, chat history, context7 library docs, etc.). Caches live under~/.gaia/cache/ (Windows:
%LOCALAPPDATA%/gaia/cache/).
| Action | Description |
|---|---|
status | Print size and location of each cache directory |
clear | Remove cache contents (scoped by flags below) |
clear:
| Flag | Description |
|---|---|
--all | Remove every cache directory (safe: preserves models under ~/.gaia/models/) |
--context7 | Clear only the Context7 MCP library-docs cache |
Kill Command
Terminate processes running on specific ports.| Option | Type | Description |
|---|---|---|
--port | integer | Port number to kill process on |
--lemonade | flag | Kill Lemonade server (port 8000) |
- Find the process ID (PID) bound to the specified port
- Forcefully terminate that process
- Provide feedback about success or failure
Diagnostics Command
Bundle system info and logs into a tarball for bug reports.| Option | Type | Default | Description |
|---|---|---|---|
--output | path | ~/.gaia/diagnostics-<YYYYMMDD-HHMMSS>.tgz | Destination path for the tarball |
--no-logs | flag | false | Omit log files (useful for public bug reports) |
- System info snapshot (
uname -a, distro, relevant env vars, all TCP listeners viass -tlnp) - State files from
~/.gaia/(config, session data — no chat content) - Log files:
~/.gaia/gaia.logand~/.gaia/electron-main.log(omitted with--no-logs)
Global Options
All commands support these global options:| Option | Type | Default | Description |
|---|---|---|---|
--logging-level | string | INFO | Logging verbosity [DEBUG, INFO, WARNING, ERROR, CRITICAL] |
-v, --version | flag | - | Show program’s version and exit |
Troubleshooting
Connection Errors
Connection Errors
If you get connection errors, ensure Lemonade server is running:
Model Issues
Model Issues
Check available system memory (16GB+ recommended)Verify model compatibility:Pre-download models:Install additional models: See Features Guide
Audio Issues
Audio Issues
List available devices:Verify microphone permissions in Windows settingsTry different audio device indices if default doesn’t work
Performance
Performance
For optimal NPU performance:
- Disable discrete GPUs in Device Manager
- Ensure NPU drivers are up to date
- Monitor system resources during execution
See Also
Code Agent
Python/TypeScript development
Blender Agent
3D scene creation
Voice Interaction
Speech-to-speech conversation
API Server
OpenAI-compatible REST API
MCP Integration
Model Context Protocol
Evaluation Framework
Testing and benchmarking