Source Code:
src/gaia/Platform Support Overview
- Windows 11: Full GUI and CLI support with all features
- Linux (Ubuntu/Debian): Full GUI and CLI support with all features
| Use-Case Example | Function | Description | Platform Support |
|---|---|---|---|
| LLM Direct | Direct LLM queries via CLI | Direct model interaction using the new gaia llm command | Windows, Linux |
| Chat Agent | Interactive conversations with history | Interactive chat sessions with conversation context and commands | Windows, Linux |
| Code Agent | AI-powered code generation and analysis | Autonomous workflow for code generation, project scaffolding, and error fixing | Windows, Linux |
| Blender Agent | 3D content creation and manipulation | Specialized agent for Blender automation and workflow | Windows, Linux |
| Summarization | Document and transcript summarization | AI-powered summarization with multiple output formats | Windows, Linux |
| Evaluation Suite | Model evaluation and benchmarking | Comprehensive evaluation framework with groundtruth generation | Windows, Linux |
| Voice Interaction | Speech-to-speech conversation | Voice-based AI interaction with TTS and ASR | Windows, Linux |
| Web Dashboards | Browser-based interfaces for agents | JavaScript apps with simple configuration | Windows, Linux |
LLM Direct Usage
Thegaia llm command provides direct access to language models without requiring server setup. This is the simplest way to interact with AI models:
Chat Agent
The Chat agent provides an interactive conversational interface with conversation history and various utility commands:- Conversation History: Maintains context across messages with configurable history length
- Document Q&A: RAG support for querying PDF documents with —index option
- Interactive Commands: Built-in commands for session management and debugging
- Streaming Responses: Real-time response streaming for better user experience
- Model Flexibility: Support for different LLM models with automatic prompt formatting
- Single Message Mode: Non-interactive mode for scripting and automation
/clear- Clear conversation history/history- Show conversation history/system- Show current system prompt configuration/model- Show current model information/prompt- Show complete formatted prompt sent to LLM/stats- Show performance statistics (tokens/sec, latency, etc.)/help- Show available commandsquit,exit, orbye- End the chat session
Blender Agent
The Blender agent provides specialized functionality for 3D content creation and workflow automation. It is now fully integrated into the CLI for easy access:- Scene Management: Clear scenes and get scene information
- Object Creation: Create cubes, spheres, cylinders, cones, and torus objects
- Material Assignment: Set RGBA colors and materials for objects
- Object Modification: Modify position, rotation, and scale of existing objects
- Interactive Planning: Multi-step scene creation with automatic planning
Summarization
The Summarization feature provides AI-powered document and transcript processing with multiple output formats:- Multiple Input Types: Process transcripts, emails, and documents
- Flexible Output Formats: JSON, PDF, email, or both
- Configurable Styles: Executive summaries, detailed summaries, action items, key decisions, participants, and topics
- Batch Processing: Process entire directories of documents
- Template Support: Use predefined configuration templates
- Model Flexibility: Support for both local (Lemonade) and cloud (OpenAI/Claude) models
json- Structured JSON with all summary componentspdf- Professional PDF reportsemail- Email-ready format with recipientsboth- Generate both JSON and PDF
executive- High-level executive summarydetailed- Comprehensive detailed summaryaction_items- Extracted action itemskey_decisions- Important decisions madeparticipants- Meeting participants and rolestopics- Main topics discussed
Evaluation Framework
GAIA includes a comprehensive evaluation framework for testing and comparing AI model performance:- Synthetic Data Generation: Create realistic test scenarios
- Ground Truth Creation: Generate evaluation standards
- Batch Experimentation: Test multiple models systematically
- Automated Evaluation: Score and compare model outputs
- Interactive Visualization: Web-based results explorer
- Cost Tracking: Monitor API costs and token usage
- Performance Metrics: Detailed timing and quality analysis
Voice Interaction (Talk)
The Talk feature enables voice-based conversations with AI models:- Speech Recognition: Whisper ASR for voice input
- Text-to-Speech: Kokoro TTS for natural voice output
- Document Q&A: RAG support for hands-free document queries
- Real-time Processing: Streaming audio pipeline
- Device Selection: Configure audio input devices
- Model Flexibility: Choose ASR model sizes
Desktop Applications
GAIA provides desktop applications with graphical interfaces for interacting with AI capabilities. These applications are built with Electron and provide native desktop experiences. For available applications and setup instructions, see the Apps Documentation. Platform Availability: Windows and LinuxSupported LLMs
The following is a list of the currently supported LLMs with GAIA using Ryzen AI Hybrid (NPU+iGPU) mode usinggaia-windows-setup.exe. To request support for a new LLM, please contact the AMD GAIA team.
| LLM | Checkpoint | Backend | Data Type |
|---|---|---|---|
| Phi-3.5 Mini Instruct | amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid | oga | int4 |
| Phi-3 Mini Instruct | amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid | oga | int4 |
| Llama-2 7B Chat | amd/Llama-2-7b-chat-hf-awq-g128-int4-asym-fp16-onnx-hybrid | oga | int4 |
| Llama-3.2 1B Instruct | amd/Llama-3.2-1B-Instruct-awq-g128-int4-asym-fp16-onnx-hybrid | oga | int4 |
| Llama-3.2 3B Instruct | amd/Llama-3.2-3B-Instruct-awq-g128-int4-asym-fp16-onnx-hybrid | oga | int4 |
| Qwen 1.5 7B Chat | amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid | oga | int4 |
| Mistral 7B Instruct | amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-fp16-onnx-hybrid | oga | int4 |
| LLM | Checkpoint | Device | Backend | Data Type |
|---|
- oga - Onnx Runtime GenAI
Installing Additional Models
GAIA uses Lemonade Server for model management. Additional models can be installed and managed through:- System Tray Icon: Access the Lemonade model manager from the system tray
- Web UI: Manage models through the Lemonade web interface