What Is Ollama?
Ollama is an open-source tool that lets you download and run large language models locally on your computer — no cloud, no API keys, no data leaving your machine.
Why Use Ollama?
- Privacy: Your conversations never leave your Mac
- No API costs: Run unlimited queries at zero cost
- Offline access: Works without an internet connection
- Speed: Local inference on Apple Silicon is surprisingly fast
Getting Started
# Install Ollama
brew install ollama
# Download and run a model
ollama run llama3
# List available models
ollama list
Popular Models on Ollama
| Model | Size | Best For |
|---|---|---|
| Llama 3 8B | 4.7 GB | General-purpose chat |
| Mistral 7B | 4.1 GB | Fast, efficient responses |
| Code Llama | 3.8 GB | Code generation |
| Phi-3 | 2.3 GB | Lightweight tasks |
| Gemma 2 | 5.4 GB | Google’s open model |
Ollama + Elvean
Elvean connects to Ollama natively — just point it at your local Ollama server and all your downloaded models appear automatically. Switch between local and cloud models mid-conversation with @mentions.
Elvean brings all these concepts together in one native Mac app — local models, cloud APIs, agentic tools, and more.
Learn more about Elvean