Guides
BYOK Providers
Deep dive into each supported provider.
Choosing a Provider
All providers produce embedding vectors for semantic search. The key difference is dimension support — your account locks to one dimension on first key save.
Recommended: 1024 dimensions — supported by OpenAI, Gemini, Cohere, and Voyage.
OpenAI
- Model:
text-embedding-3-small(1536d default, supports custom 256-3072) - LLM: Yes (gpt-4)
- Supports
baseUrlfor OpenRouter, Together, Fireworks
Google Gemini
- Model:
gemini-embedding-001(3072d default, supports custom 256-3072) - LLM: Yes (gemini-2.5-pro)
- Free tier: unlimited embedding calls
Cohere
- Model:
embed-v4.0(1024d default, supports 256-1024) - LLM: No (embedding only)
- Free tier: 1,000 calls/month
- Supports query/document input types for better search quality
Voyage AI
- Model:
voyage-3(1024d, supports custom) - LLM: No (embedding only)
- Free tier: 200M tokens, 3 RPM limit
Azure OpenAI
- Requires base URL:
https://your-resource.openai.azure.com - Key format: API key + base URL in the save form
- Same models as OpenAI
Mistral
- Model:
mistral-embed(256d fixed, cannot be changed) - LLM: Yes (mistral-large-latest)
- Only compatible with accounts locked at 256d
Ollama (Local)
- Model:
nomic-embed-text(768d) - API key = endpoint URL (e.g.,
http://localhost:11434/v1) - Free, self-hosted
Qwen (Self-Hosted)
- Model:
Qwen3-Embedding-8B(4096d) - API key = TEI endpoint URL
- Free, self-hosted via HuggingFace TEI