Learn about the multiple LLM providers supported by NOMOS
NOMOS supports multiple LLM providers, allowing you to choose the best model for your use case.
GPT-4o, GPT-4o-mini, and more
Claude 3.5 Sonnet, Haiku, and Opus
Gemini 2.0 Flash, Pro, and more
Mistral Large, Medium, and Small
Local models including Llama, Qwen, and more
Open source models via HuggingFace
Use the BaseLLM
class to implement your own provider
NOMOS is continuously expanding support for new LLM providers
You can specify LLM configuration in your YAML config file:
You can pass additional parameters to LLM providers:
API Key Not Found
Ensure environment variables are set correctly in your shell profile or .env.local
file
Model Not Available
Check that the model name is correct and available in your region
Rate Limits
Implement retry logic or use different models with higher rate limits
Local Models (Ollama)
Ensure Ollama is running (ollama serve
) and the model is pulled (ollama pull model-name
)
NOMOS includes built-in error handling and retry mechanisms:
Use smaller models for simple tasks to reduce latency and costs
Lower values (0.1-0.3) for consistent responses
Limit response length to control costs and latency
Ollama for development or when data privacy is important
For the most up-to-date list of available models, refer to the official documentation:
Official Claude models documentation
Complete OpenAI models reference
Vertex AI Generative AI models
Mistral AI models overview
Browse available local models
Explore HuggingFace model hub