Tokenmaxxing Is the Budget Game, Played With AI Tokens
Silicon Valley's new productivity metric rewards consumption over outcomes. Anyone who's managed a fiscal year-end budget has seen this movie before.
A few weeks ago, a Meta employee built an internal leaderboard called “Claudeonomics” that ranked 85,000 employees by how many AI tokens each one consumed. The top user burned through 281 billion tokens in 30 days. Some employees were running agents idle for hours just to climb the rankings. Titles got handed out: “Token Legend,” “Cache Wizard,” “Session Immortal.”
The leaderboard came down two days after the press found out.
I watched this story unfold and my first thought wasn’t about AI at all. I’ve seen this exact behavior every Q4 for twenty-plus years. If you’ve ever run a P&L or managed a departmental budget, you know the pattern. Your allocation for the fiscal year is $2M. You spent $1.4M. If you return the remaining $600K, next year your budget gets cut to $1.5M. So in October, you start spending on hardware you don’t need yet, conferences that are marginally useful, contracts pulled forward from Q1. The incentive structure makes that spending rational. Use it or lose it.
Tokenmaxxing is the same game with a different currency. Meta and Shopify now include AI usage in performance reviews. Jensen Huang said he’d be “deeply alarmed” if an engineer with a $500K salary didn’t consume at least $250K in tokens within a year. Meta’s CTO has publicly encouraged engineers to spend the equivalent of their salaries on tokens, claiming 10x productivity gains.
The message to the rank and file is clear: if you’re not burning tokens, you’re falling behind. And just like the budget game, people respond to the incentive, not the intent. Running five parallel agents overnight produces nothing that ships, but it produces a ranking. The leaderboard rewards what it can see.
Anyone who has worked in engineering recognizes this problem. We measured lines of code in the 90s, then story points in the 2010s, now tokens. The mistake keeps finding a new unit.
But the parallel to corporate budgets breaks in one important place.
Corporate budgets are finite and zero-sum. One department’s gain is another’s loss. Spending more doesn’t compound into anything. You buy the hardware you didn’t need, it sits in a closet, year over.
AI capability isn’t like that. An engineer who learns how to prompt well, who builds muscle memory with agentic workflows, who figures out where AI saves real time versus where it creates rework downstream, carries compounding knowledge. The learning accumulates. Even some of the wasted tokens produce real learning, because you need to try things before you know what works.
So tokenmaxxing sits in the middle. Whether it helps depends on what you’re actually measuring.
Business spending on AI tokens is up 13x since January 2025. Part of that is genuine capability expansion. Engineers discovering that a well-structured agentic workflow takes three days off a release cycle. Product teams prototyping faster than they can spec. Those are real gains worth paying for. The rest is pure theater. From the outside, both look identical on a token dashboard.
One pattern I had to correct in myself was confusing adoption with value. Years ago I pushed a team hard to adopt a new CI/CD pipeline. We got to 100% adoption in six weeks. I reported it up as a win. What I didn’t measure was that a third of the team had wired their old manual process into the new pipeline to make the dashboards green. They were technically using the tool. Nothing about how they worked had actually changed.
Tokenmaxxing carries the same risk at a much larger scale. When you track consumption without tracking what it produces, you’re measuring how fast the engine runs while ignoring whether the car is moving.
I think Huang’s underlying logic is sound. If a $500K engineer plus $250K in compute gets you the output of two engineers, that’s a reasonable bet. Maybe the best available right now. The companies that figure out how to direct that spend, not just how to increase it, are the ones who pull ahead. “Figure out how to direct it” is the hard part though, and it’s not a question a leaderboard can answer.
The companies I’d bet on are the ones where an engineer can tell you, specifically, what changed. Not “I used AI more this quarter.” Something concrete, like a review cycle that went from four days to one because AI handles the first pass on test coverage. Or a team that finally writes architecture decision records because the draft takes an hour instead of a week.
That’s the difference between spending a budget because it exists and investing it because you know what the return looks like. The cost question is real, and right now it’s almost secondary. Meta’s top token user could have cost the company more than $1.4M in 30 days. At some point, finance teams will ask what that bought. Orgs that can answer clearly get to keep their budgets.
Smart finance leaders eventually learned that “use it or lose it” was destroying capital discipline. The best orgs moved to rolling budgets, zero-based budgeting, or some model that separated “we need this resource” from “we’ll lose it if we don’t spend it.” AI adoption needs the same maturity. Treat token usage as a cost input, not a success metric, and measure what shipped, what improved, what decisions got made faster. Give teams room to experiment heavily because the curve is steep and the upside is genuine, but don’t reward the engineer with the biggest overnight agent bill.
Tokenmaxxing is the right instinct with the wrong scoreboard. Pushing hard on AI adoption and investing aggressively are the right moves. Counting consumption instead of outcomes is where it becomes the budget game again.
Anyone who has survived a few fiscal years knows how that one ends. The money gets spent, and the org doesn’t come out any smarter.



