Quickstart
Last updated
Last updated
Be sure to complete the before continuing with this guide.
Before starting AIOS, you need to make sure you have installed the LLM backends that you would like to run. Here are the LLM providers for supported backends for AIOS.
Anthropic
Claude 3.5 Sonnet
❌
claude-3-5-sonnet-20241022
anthropic
ANTHROPIC_API_KEY
Anthropic
Claude 3.5 Haiku
❌
claude-3-5-haiku-20241022
anthropic
ANTHROPIC_API_KEY
Anthropic
Claude 3 Opus
❌
claude-3-opus-20240229
anthropic
ANTHROPIC_API_KEY
Anthropic
Claude 3 Sonnet
❌
claude-3-sonnet-20240229
anthropic
ANTHROPIC_API_KEY
Anthropic
Claude 3 Haiku
❌
claude-3-haiku-20240307
anthropic
ANTHROPIC_API_KEY
Deepseek
Deepseek-reasoner (R1)
❌
deepseek-reasoner
deepseek
DEEPSEEK_API_KEY
Deepseek
Deepseek-chat (V3)
❌
deepseek-chat
deepseek
DEEPSEEK_API_KEY
OpenAI
GPT-4
❌
gpt-4
openai
OPENAI_API_KEY
OpenAI
GPT-4 Turbo
❌
gpt-4-turbo
openai
OPENAI_API_KEY
OpenAI
GPT-4o
❌
gpt-4o
openai
OPENAI_API_KEY
OpenAI
GPT-4o mini
❌
gpt-4o-mini
openai
OPENAI_API_KEY
OpenAI
GPT-3.5 Turbo
❌
gpt-3.5-turbo
openai
OPENAI_API_KEY
Gemini 1.5 Flash
❌
gemini-1.5-flash
GEMINI_API_KEY
Gemini 1.5 Flash-8B
❌
gemini-1.5-flash-8b
GEMINI_API_KEY
Gemini 1.5 Pro
❌
gemini-1.5-pro
GEMINI_API_KEY
Gemini 1.0 Pro
❌
gemini-1.0-pro
GEMINI_API_KEY
Groq
Llama 3.2 90B Vision
✅
llama-3.2-90b-vision-preview
groq
GROQ_API_KEY
Groq
Llama 3.2 11B Vision
✅
llama-3.2-11b-vision-preview
groq
GROQ_API_KEY
Groq
Llama 3.1 70B
✅
llama-3.1-70b-versatile
groq
GROQ_API_KEY
Groq
Llama Guard 3 8B
✅
llama-guard-3-8b
groq
GROQ_API_KEY
Groq
Llama 3 70B
✅
llama3-70b-8192
groq
GROQ_API_KEY
Groq
Llama 3 8B
✅
llama3-8b-8192
groq
GROQ_API_KEY
Groq
Mixtral 8x7B
✅
mixtral-8x7b-32768
groq
GROQ_API_KEY
Groq
Gemma 7B
✅
gemma-7b-it
groq
GROQ_API_KEY
Groq
Gemma 2B
✅
gemma2-9b-it
groq
GROQ_API_KEY
Groq
Llama3 Groq 70B
✅
llama3-groq-70b-8192-tool-use-preview
groq
GROQ_API_KEY
Groq
Llama3 Groq 8B
✅
llama3-groq-8b-8192-tool-use-preview
groq
GROQ_API_KEY
ollama
✅
model-name
ollama
-
vLLM
✅
model-name
vllm
-
HuggingFace
✅
model-name
huggingface
HF_HOME
Set up configuration file directly (Recommended)
You need API keys for services like OpenAI, Anthropic, Groq and HuggingFace. The simplest way to configure them is to edit the aios/config/config.yaml
.
[!TIP] It is important to mention that, we strongly recommend using the
aios/config/config.yaml
file to set up your API keys. This method is straightforward and helps avoid potential sychronization issues with environment variables.
A simple example to set up your API keys in aios/config/config.yaml
is shown below:
To obtain these API keys:
Deepseek API: https://api-docs.deepseek.com/
OpenAI API: https://platform.openai.com/api-keys
Google Gemini API: https://makersuite.google.com/app/apikey
Groq API: https://console.groq.com/keys
HuggingFace Token: https://huggingface.co/settings/tokens
Anthropic API: https://console.anthropic.com/keys
Configure LLM Models
You can configure which LLM models to use in the same aios/config/config.yaml
file. Here's an example configuration:
Using Ollama Models:
First, download ollama from https://ollama.com/
Start the ollama server in a separate terminal:
Pull your desired models from https://ollama.com/library:
Using vLLM Models:
Start the vLLM server in a separate terminal:
vLLM currently only supports Linux and GPU-enabled environments. If you don't have a compatible environment, please choose other backend options. To enable the tool calling feature of vllm, refer to https://docs.vllm.ai/en/latest/features/tool_calling.html
Using HuggingFace Models: You can configure HuggingFace models with specific GPU memory allocation:
After you setup the required keys, you can run the following command to launch the AIOS kernel.
And then you can start a client to interact with the AIOS kernel using Terminal or WebUI.
Ollama supports both CPU-only and GPU environments. For more details about ollama usage, visit
Install vLLM following their