← AI Glossary

What Is Ollama?

Ollama is an open-source tool that lets you download and run large language models locally on your computer — no cloud, no API keys, no data leaving your machine.

Why Use Ollama?

  • Privacy: Your conversations never leave your Mac
  • No API costs: Run unlimited queries at zero cost
  • Offline access: Works without an internet connection
  • Speed: Local inference on Apple Silicon is surprisingly fast

Getting Started

# Install Ollama
brew install ollama

# Download and run a model
ollama run llama3

# List available models
ollama list
ModelSizeBest For
Llama 3 8B4.7 GBGeneral-purpose chat
Mistral 7B4.1 GBFast, efficient responses
Code Llama3.8 GBCode generation
Phi-32.3 GBLightweight tasks
Gemma 25.4 GBGoogle’s open model

Ollama + Elvean

Elvean connects to Ollama natively — just point it at your local Ollama server and all your downloaded models appear automatically. Switch between local and cloud models mid-conversation with @mentions.

Elvean brings all these concepts together in one native Mac app — local models, cloud APIs, agentic tools, and more.

Learn more about Elvean