Llm

10 articles about llm.

Accessibility Tree Dumps Overflow LLM Context Windows - How to Fix It

·3 min read

Raw accessibility tree data can consume 24KB or more per dump, flooding AI agent context windows. The fix: write to temp files and return concise summaries instead.

accessibility-treecontext-windowllmmacosoptimizationdesktop-agent

AI Pricing Is Unsustainable - API Costs Are Rising with Agent Usage

·3 min read

Building desktop automation tools, API costs went from $30 to $200 per month as agent usage scaled. The current AI pricing model is unsustainable for always-on agents.

pricingapi-costsai-agentsustainabilityllmbudget

Stop Re-Explaining Context to Your AI - Use File-Based Context Instead

·2 min read

Most people spend 20-30% of their AI interaction time re-explaining context. File-based context systems like CLAUDE.md eliminate this by loading context automatically every session.

contextllmfile-basedproductivityclaude-md

Spawning 5+ Claude Agents in Parallel Makes Your API Bill a Second Rent Payment

·2 min read

Without a proper LLM control plane, parallel agents burn tokens on repeated context. Route simple tasks locally, batch API calls, and prune aggressively.

llmparallel-agentsapi-costscontrol-planebudgeting

Building an LLM-Powered Data Janitor for Browser-Extracted Memories

·2 min read

How to build an LLM-powered review skill that classifies browser-extracted memories into keep, delete, merge, and fix categories - with self-ranking via hit rate.

llmdata-cleaningbrowsermemoriesai-agentautomation

Why Scoped 50K Context Agents Outperform One Million Token Context

·2 min read

One million token context windows sound impressive, but scoped agents with 50K context each consistently outperform a single giant context for real development work.

context-windowparallel-agentsscoped-agentsllmproductivity

Opus Token Burn Rate - Watching It Write, Delete, and Rewrite 200-Line Functions

·3 min read

Opus does not just burn tokens - it vaporizes them. The write-delete-rewrite cycle where Opus creates 200 lines, decides it does not like them, and starts over.

opustokensclaude-codeai-codingcostllm

Stop Fighting the Context Limit - Scope Each Agent to One Small Task

·2 min read

Instead of cramming everything into one LLM context window, scope each AI agent to a single small task. Fix this crash. Add this button. One job, one agent.

context-limitai-agentscopingproductivityllmworkflow

Your AI Agent Needs a Control Plane - LLM Routing, Token Budgets, and Fallbacks

·3 min read

Why AI agents need infrastructure for routing between Claude and local models, tracking token budgets, retrying with fallback, and audit logging.

llmcontrol-planeroutingtoken-budgetinfrastructure

How LLMs Can Control Your Computer - Voice-Driven, Local, No API Keys

·4 min read

A look at how large language models power desktop automation agents that control your actual computer through voice commands, running fully local with no cloud dependency.

llmdesktop-agentvoice-controllocal-firstopen-source

Browse by Topic