Installation and Setup
Can I reinstall GAIA?
Can I reinstall GAIA?
Yes, you can reinstall GAIA. The installer provides an option to remove your existing installation before reinstalling.
How do I run GAIA in silent/headless mode?
How do I run GAIA in silent/headless mode?
Run the installer from command-line with parameters for CI/CD or silent installations:Available parameters:
/S- Silent installation (no UI)/D=<path>- Set installation directory (must be last parameter)
What are the system requirements?
What are the system requirements?
GAIA is designed for AMD Ryzen AI systems:
Driver Requirements:
| Component | Requirement |
|---|---|
| Processor | AMD Ryzen AI 300-series (for optimal performance) |
| RAM | 16GB minimum, 64GB recommended |
| Storage | 20GB free space |
| OS | Windows 11 Pro 24H2 or Ubuntu 22.04+ |
- Radeon iGPU:
32.0.22029.1019or later - NPU:
32.0.203.314or later
What platforms does GAIA support?
What platforms does GAIA support?
- Windows 11: ✅ Fully supported with complete UI and CLI functionality
- Linux (Ubuntu/Debian): ✅ Fully supported with complete UI and CLI functionality
How do I install additional models?
How do I install additional models?
Additional models can be installed through Lemonade Server’s model management interface:
- System Tray Icon: Access the Lemonade model manager from the system tray
- Web UI: Manage models through the Lemonade web interface
Demo and Capabilities
Discover the capabilities of Ryzen AI with GAIA - an innovative generative AI application that runs private, local LLMs on the Neural Processing Unit (NPU).
What is GAIA, and how does it integrate with Ryzen AI?
What is GAIA, and how does it integrate with Ryzen AI?
GAIA is AMD’s generative AI application that runs local, private LLMs on Ryzen AI’s NPU hardware. It leverages the power of the NPU for faster, more efficient processing, allowing users to keep their data local without relying on cloud infrastructure.GAIA uses the Lemonade Server to load and run LLM inference optimally on AMD hardware.Key Benefits:
- Local execution (no cloud dependency)
- Enhanced privacy
- Optimized for AMD NPU hardware
- Lower power consumption
How does the agent RAG pipeline work?
How does the agent RAG pipeline work?
The RAG (Retrieval-Augmented Generation) pipeline combines an LLM with a knowledge base:Components:
- Agent: Capable of retrieving relevant information
- Reasoning: Plans and executes multi-step tasks
- Tool Use: Accesses external tools and APIs
- Interactive Chat: Real-time conversation interface
What LLMs are supported?
What LLMs are supported?
GAIA supports various local LLMs optimized for Ryzen AI NPU hardware, including:Hybrid Mode (NPU + iGPU) - Ryzen AI 300 series:
- Phi-3.5 Mini Instruct
- Phi-3 Mini Instruct
- Llama-2 7B Chat
- Llama-3.2 1B/3B Instruct
- Qwen 1.5 7B Chat
- Mistral 7B Instruct
How does the NPU enhance LLM performance?
How does the NPU enhance LLM performance?
The NPU (Neural Processing Unit) in Ryzen AI is specialized for AI workloads, specifically the matrix multiplications (GEMMs) in the model:Benefits:
- Faster Inference: Optimized hardware acceleration
- Lower Power: More efficient than CPU/iGPU
- System Offload: Reduces load on main processor
Can this scale to larger LLMs or enterprise applications?
Can this scale to larger LLMs or enterprise applications?
Absolutely. While GAIA showcases local, private implementation:
- Architecture scales to larger models
- NPU optimization ensures efficient scaling
- Enterprise deployments supported
- Hybrid cloud/local configurations possible
What are the benefits of running LLMs locally on the NPU?
What are the benefits of running LLMs locally on the NPU?
Privacy:
- No data leaves your machine
- Complete control over sensitive information
- Reduced latency (no cloud communication)
- Faster response times with NPU acceleration
- No ongoing cloud API costs
- Lower power consumption
NPU vs iGPU: What's the difference?
NPU vs iGPU: What's the difference?
NPU (Neural Processing Unit):
- Optimized specifically for AI inference
- Lower power consumption
- Faster for LLM workloads
- AI-focused architecture
- General-purpose graphics/compute
- Higher power consumption for AI
- Slower inference for LLMs
- Graphics-focused architecture
Demo Components and Workflow
The GAIA demo consists of two main components working together to provide powerful local AI capabilities.
System Architecture
1. GAIA Backend Powered by the Ryzen AI platform through Lemonade Server:- NPU/iGPU Acceleration: Leverages Ryzen AI hardware
- Multiple Models: Supports various LLMs including Llama-3.2-3B-Instruct-Hybrid
- Agent System: Specialized tasks and workflows
- WebSocket Streaming: Real-time response delivery
- Dual Interfaces: Both CLI and GUI available
- Repository Vectorization: Fetches and indexes code repositories
- Local Vector Storage: Fast, local similarity search
- Fast Indexing: ~10 seconds for 40,000 lines of code (typical laptop)
- Ready for Queries: Instant access to indexed knowledge
Query Process
1
User Input
Query sent to GAIA (e.g., “How do I install dependencies?”)
2
Embedding Generation
For RAG queries, input transformed into embeddings
3
Context Retrieval
Relevant content retrieved from local repositories/documents
4
LLM Processing
Context passed to LLM via Lemonade Server for generation
5
Response Streaming
Generated response streamed back in real-time through GAIA interfaces
Use Cases and Applications
What applications or industries could benefit?
What applications or industries could benefit?
GAIA’s local AI capabilities are ideal for industries requiring high performance and privacy:Healthcare:
- Patient data privacy compliance
- Medical record analysis
- Clinical decision support
- Sensitive financial data processing
- Compliance and audit trails
- Risk analysis
- Internal document Q&A
- Code analysis and generation
- Knowledge management
- Writing assistance
- Code generation
- Document summarization
- Automated support systems
- Intent classification
- Response generation
How does this address data privacy concerns?
How does this address data privacy concerns?
GAIA emphasizes running LLMs locally, meaning:
- All data remains on your device
- No cloud transmission of sensitive information
- Complete control over data and models
- Compliance-friendly for regulated industries
- High-performance AI without privacy trade-offs
What toolset do I need to replicate this?
What toolset do I need to replicate this?
To run GAIA, you need:Hardware:
- Ryzen AI processor (300-series recommended)
- Lemonade Server for managing LLMs
- GAIA framework (available on Windows and Linux)
- Command-line interface (CLI) for scripting and automation
- Graphical user interface (GUI) for interactive use
See Also
Development Guide
Setup and installation instructions
Features
Complete feature overview
CLI Reference
Command-line interface guide
Supported Models
Available LLM models