Skin in the Game Separates Agents from Assistants
Skin in the Game Separates Agents from Assistants
Ask an AI assistant to research a topic and it will make 50 API calls without hesitation. Give the same task to an agent that can see its running cost, and it makes 12 calls then delivers a "good enough" result. The difference is not intelligence - it is incentives.
The Cost-Blind Problem
Most AI assistants operate in a cost vacuum. They do not know what each action costs, and they have no reason to care. The result is predictable - they over-research, over-process, and over-generate. A simple question gets a 2000-word response because there is no cost signal telling the model to be concise.
What Changes with Visibility
When you add running cost as an observable variable in the agent's context:
- Research tasks terminate earlier with "sufficient" results
- The agent chooses cheaper models for simple sub-tasks
- Responses get more concise
- The agent asks clarifying questions instead of guessing (because a wrong guess wastes more money than a question)
Agents vs Assistants
This is the real distinction between an agent and an assistant:
- An assistant optimizes for output quality with no resource constraints
- An agent optimizes for outcome value minus resource cost
An assistant will spend $5 to improve a result by 2%. An agent will recognize that 2% improvement is not worth $5 and stop. The agent is not lazier - it is making better decisions because it has more information.
Building Cost Awareness
The implementation is simpler than you might think:
- Track cumulative API costs per task
- Include the running total in the agent's system prompt
- Set a budget ceiling and let the agent decide how to allocate it
- After each action, append the cost to the agent's context
The key insight is that you do not need to teach the agent to be frugal. You just need to make costs visible. The model's existing reasoning capabilities handle the rest.
The Nassim Taleb Connection
Taleb's "skin in the game" principle applies perfectly. Systems make better decisions when they bear the consequences of those decisions. AI agents that see their own bill naturally develop better judgment about when to stop researching, which model to use, and whether an additional API call is worth it.
Fazm is an open source macOS AI agent. Open source on GitHub.