What it does
The shape of Ollama, in plain English.
Ollama makes it simple to download and run open-source large language models like LLaMA, Mistral, and CodeLlama on your own machine with a clean CLI and OpenAI-compatible API.
Why we like it
The parts that make us reach for it.
- Running LLMs locally and privately
- Offline AI development
- Prototyping without API costs
- Privacy-sensitive applications
- Learning how LLMs work
When to use it
Match the tool to the job.
Each block below is a different day in the life of Ollama.
coding
Ship features, refactor code, and review diffs without leaving your editor.
research
Synthesise across long PDFs, papers, and transcripts — cite as you go.
agents
Keep a loop of reasoning + tool-use that doesn't spin forever.
What to watch out for
Where it gets in your way.
Not deal-breakers — just worth knowing before you commit.
- Requires decent hardware (8GB+ RAM)
- Slower than cloud APIs on consumer hardware
- Model capabilities behind frontier models
- No built-in fine-tuning
Under the hood
Feature checklist.
The bill
How much this will cost you.
Completely free and open-source. You provide the hardware.