Skip to content

Providers

Remind supports multiple LLM and embedding providers. You need one of each — an LLM for consolidation and an embedding provider for retrieval.

Provider matrix

ProviderLLMEmbeddingsAPI key required
AnthropicYesNoYes
OpenAIYesYesYes
Azure OpenAIYesYesYes
OllamaYesYesNo (local)

Anthropic

LLM only. Pair with OpenAI or Ollama for embeddings.

json
{
  "llm_provider": "anthropic",
  "embedding_provider": "openai",
  "anthropic": {
    "api_key": "sk-ant-...",
    "model": "claude-sonnet-4-20250514"
  }
}
bash
ANTHROPIC_API_KEY=sk-ant-...
LLM_PROVIDER=anthropic
EMBEDDING_PROVIDER=openai
SettingDefault
api_key
modelclaude-sonnet-4-20250514

OpenAI

LLM and embeddings.

json
{
  "llm_provider": "openai",
  "embedding_provider": "openai",
  "openai": {
    "api_key": "sk-...",
    "model": "gpt-4.1",
    "embedding_model": "text-embedding-3-small"
  }
}
bash
OPENAI_API_KEY=sk-...
LLM_PROVIDER=openai
EMBEDDING_PROVIDER=openai
SettingDefault
api_key
base_urlnull (OpenAI default)
modelgpt-4.1
embedding_modeltext-embedding-3-small

Azure OpenAI

LLM and embeddings via Azure.

json
{
  "llm_provider": "azure_openai",
  "embedding_provider": "azure_openai",
  "azure_openai": {
    "api_key": "...",
    "base_url": "https://your-resource.openai.azure.com",
    "api_version": "2024-02-15-preview",
    "deployment_name": "gpt-4",
    "embedding_deployment_name": "text-embedding-3-small",
    "embedding_size": 1536
  }
}
bash
AZURE_OPENAI_API_KEY=...
AZURE_OPENAI_API_BASE_URL=https://your-resource.openai.azure.com
AZURE_OPENAI_API_VERSION=2024-02-15-preview
AZURE_OPENAI_DEPLOYMENT_NAME=gpt-4
AZURE_OPENAI_EMBEDDING_DEPLOYMENT_NAME=text-embedding-3-small
LLM_PROVIDER=azure_openai
EMBEDDING_PROVIDER=azure_openai
SettingDefault
api_key
base_url
api_version
deployment_name
embedding_deployment_name
embedding_size1536

Ollama (local)

Fully local, no API keys. Install Ollama first:

bash
ollama pull llama3.2           # LLM
ollama pull nomic-embed-text   # Embeddings
json
{
  "llm_provider": "ollama",
  "embedding_provider": "ollama",
  "ollama": {
    "url": "http://localhost:11434",
    "llm_model": "llama3.2",
    "embedding_model": "nomic-embed-text"
  }
}
bash
OLLAMA_URL=http://localhost:11434
OLLAMA_LLM_MODEL=llama3.2
OLLAMA_EMBEDDING_MODEL=nomic-embed-text
LLM_PROVIDER=ollama
EMBEDDING_PROVIDER=ollama
SettingDefault
urlhttp://localhost:11434
llm_modelllama3.2
embedding_modelnomic-embed-text

Mixing providers

You can use different providers for LLM and embeddings:

json
{
  "llm_provider": "anthropic",
  "embedding_provider": "openai"
}

Common combos:

  • Anthropic LLM + OpenAI embeddings — Best consolidation quality + fast embeddings
  • OpenAI for both — Simplest setup, single API key
  • Ollama for both — Fully local, no API keys, no data leaves your machine
  • Anthropic LLM + Ollama embeddings — Quality consolidation, local embeddings

Released under the Apache 2.0 License.