Ollama
7 articles about ollama.
Learning Path for Local LLMs - From Ollama to Desktop Agents
A practical learning path for running local LLMs: start with Ollama basics, learn prompting, understand quantization, build workflows, then automate your desktop.
How to Cut AI Agent Costs 50-70% with Model Routing
Route simple tasks to local Ollama models, complex ones to Claude. Combine that with aggressive state summarization and context pruning to keep token usage down without losing important information.
LLM Observability for Desktop Agents - Beyond Logging Model Outputs
Traditional LLM observability focuses on model outputs. For desktop agents, watching what the agent actually does on screen - logging actions, not just reasoning - matters more.
VS Code Claude Extension vs Terminal with Ollama - Why the Terminal Route Wins
The VS Code Claude extension is locked to Anthropic's API. Running Claude Code in the terminal with Ollama gives you local models, more control, and zero API costs.
Using Ollama for Local Vision Monitoring on Apple Silicon
Local vision models through Ollama handle real-time monitoring tasks like watching your parked car. Apple Silicon M-series makes local inference fast enough for practical use.
Build a Local-First AI Agent with Ollama - No API Keys, No Cloud, No Signup
How to run an AI desktop agent entirely on your Mac using Ollama for local inference. No API keys needed, no data leaves your machine, works offline.
Local LLMs Are Not Just for Inference Anymore - Real Workflows on Your Machine
The shift to local LLMs is moving beyond chat and inference into real desktop automation. Browser control, CRM updates, document generation - all without cloud APIs.