Prerequisites
- macOS/Linux: Install Ollama from ollama.ai
- Windows: Ollama is available as a preview for Windows — see ollama.ai/download
- Minimum 8 GB RAM for 7B models; 16 GB recommended for 13B+ models
Install Ollama and pull a model
Connect MIRA to Ollama
Select Ollama in the Engine tab
Press ⌘, → Engine tab → in the Provider & Model section, click the Ollama (local) button.
Enter a model name
Type the name of any model you have pulled with
ollama pull into the Model ID field.
MIRA shows a few suggestions as chips, but any model from ollama list works — just type the exact name.| Suggested model | RAM needed |
|---|---|
llama3.2 | 8 GB |
llama3.1 | 8 GB |
mistral | 8 GB |
qwen2.5-coder | 8 GB |
deepseek-r1 | 8 GB |
Set the Base URL (optional)
The Base URL field appears automatically for Ollama. Default:
http://localhost:11434. Change it only if Ollama is running on a different host or port.Performance tips
- Use a Mac with Apple Silicon (M1/M2/M3/M4) — Ollama uses Metal GPU acceleration
- Reduce Context Budget in the Engine → NAE settings when using smaller models (e.g. 20 000–40 000 for 7B models)
- Keep other heavy applications closed when running 13B+ models
Troubleshooting
| Error | Fix |
|---|---|
ECONNREFUSED localhost:11434 | Ollama is not running — run ollama serve |
| Model returns garbled output | Pull the model again: ollama pull <name> to ensure a complete download |
| Out of memory | Use a smaller model or increase system swap |
Edit this page — Open a pull
request