LLM configuration
Ollama
To use Letta with Ollama, set the environment variable
OLLAMA_BASE_URL=http://localhost:11434
. ⚠️ Make sure to use tags when downloading Ollama models!
Don’t do
ollama pull dolphin2.2-mistral
, instead doollama pull dolphin2.2-mistral:7b-q6_K
.If you don’t specify a tag, Ollama may default to using a highly compressed model variant (e.g. Q4). We highly recommend NOT using a compression level below Q5 when using GGUF (stick to Q6 or Q8 if possible). In our testing, certain models start to become extremely unstable (when used with MemGPT) below Q6.
Setup Ollama
- Download + install Ollama and the model you want to test with
- Download a model to test with by running
ollama pull <MODEL_NAME>
in the terminal (check the Ollama model library for available models)
For example, if we want to use Dolphin 2.2.1 Mistral, we can download it by running:
Enabling Ollama as a provider
To enable the Ollama provider, you must set the OLLAMA_BASE_URL
environment variable. When this is set, Letta will use available LLM and embedding models running on Ollama.