.toml file for configuration. The file is provided during initial installation. For clarity, below is a set of configurations that can be modified according to your needs.
Models configuration (Required)
The Context Engine requires at least:- One embedding model
- One chat model capable of agentic (multi-step) execution
Model types
1. Embedding model (Required)
Drives chunking and retrieval. May require tuning if replaced. Example:text-embedding-large-3
2. Super-fast model (Optional)
A lightweight model for very quick operations. If not provided, the agentic model will handle these tasks, but indexing may become slower. Example:gpt-4.1-nano
Recommendation: If no fast or super-fast model is available, it’s best to disable code description during indexing to avoid slowdowns.
3. Fast chat model (Optional)
Used for quick, non-reasoning tasks. If not provided, the agentic model can be used instead, with some performance impact. Example:gpt-4.1-mini
4. Agentic-capable chat model (Required)
Runs the core agent workflows and multi-step reasoning. Example:claude-sonnet-4-5
Required models summary
| Purpose | Model Example | Required |
|---|---|---|
| Embeddings | text-embedding-large-3 | Yes |
| Agentic-capable chat | e.g.,claude-sonnet-4-5 | Yes |
| Fast chat model | e.g.,gpt-4.1-mini | Optional |
| Super-fast model | e.g.,gpt-4.1-nano | Optional |
Example model configurations
OpenAI-only example
| Role | Model | Embedding |
|---|---|---|
| Embedding | openai/text-embedding-large-3 | Yes |
| Super-fast model | openai/gpt-4.1-nano | Optional |
| Fast chat model | openai/gpt-4.1-mini | Optional |
| Agentic-capable chat model | openai/gpt-5.1 | Yes |
Bedrock-only example
| Role | Model | Embedding |
|---|---|---|
| Embedding | bedrock/amazon.titan-embed-text-v2:0 | Yes |
| Super-fast model | bedrock/anthropic.claude-haiku-4-5-20251001-v1:0 | Optional |
| Fast chat model | bedrock/anthropic.claude-haiku-4-5-20251001-v1:0 | Optional |
| Agentic-capable chat model | bedrock/global.anthropic.claude-sonnet-4-5-20250929-v1:0 | Yes |
Default configuration
Bedrock configuration
Environment variable that must be set to enable bedrock routing- AI_MODELS_FILE=ai_models_bedrock.py
LLM gateway support
Our product supports integration with custom LLM gateways, provided they implement the same API interfaces as the official OpenAI or Anthropic model endpoints. Environment variable that must be set in all deployments to enable LLM Gateway routing- AI_MODELS_FILE=ai_models_onprem.py
Monitoring config (recommended)
- Sentry - We use Sentry for error tracking. When enabled, we only emit code crashes and traces.
- Mixpanel - Used for analytics only. No user data is sent.
- CLIENT_DISPLAY_NAME - Name of your org e.g.
"qodo_ai"used as identifier for the above tools