The $20 Problem: As AI Inference Costs Plummet 50x, Knowledge Work Artifacts Become Commodities
Key Takeaways
- ▸LLM inference costs have declined 50x in three years, making knowledge work artifacts functionally free to produce at scale
- ▸A complete operating model with board presentation that formerly required a full day of analyst work now takes 30 minutes at a $20/month cost
- ▸This represents a 'desktop publishing moment' for knowledge work—commoditizing production while shifting value to judgment, strategy, and curation
Summary
With inference costs for large language models dropping by a factor of 50 over three years—from $20 to $0.40 per million tokens—knowledge work artifacts like presentations, analyses, and code are rapidly becoming commodities. A recent example demonstrated how a complex operating model that would take a Goldman Sachs analyst a full day to produce can now be generated in thirty minutes using Claude, complete with a board-ready presentation, for just $20 per month. This represents a fundamental shift in the economics of knowledge work, similar to how desktop publishing democratized layout design in the 1990s but shifted value from production capability to design judgment. The critical question emerging in strategy meetings is no longer about artifact production—it's about what organizations are actually selling when the cost of creating those artifacts approaches zero.
- Organizations across industries have not yet adapted their business models or strategy conversations to account for this fundamental cost shift
- Price competition between AI providers (DeepSeek's 90% cuts, OpenAI's 80% response) is accelerating the downward cost trajectory
Editorial Opinion
The collapse of artifact production costs represents one of the most disruptive economic shifts in knowledge work since the spreadsheet. This isn't theoretical—it's happening now, at scale, and most organizations haven't grasped the implications for their service offerings and value propositions. The real opportunity lies not in racing to automate faster, but in being the first to clearly answer what you're selling when production becomes free.


