Reid Hoffman on Tokenmaxxing and AI Work Metrics

Reid Hoffman says token usage can drive AI experimentation, but the real story is how tokenmaxxing turns adoption into status and spend.

Reid Hoffman on Tokenmaxxing and AI Work Metrics

The Reid Hoffman tokenmaxxing debate is not really about whether employees should use more AI. It is about what happens when companies turn experimentation into a visible score and start confusing activity with value.

The second a company turns behavior into a dashboard, the behavior is dead. What replaces it is performance.

So when I saw the Meta tokenmaxxing story, employees chasing titles like Token Legend and Cache Wizard, my first reaction was not wow, the future of work. It was: ah yes, AI Peloton for office politics. Then I read the TechCrunch piece on the Reid Hoffman tokenmaxxing debate, and annoyingly, he made the part of the argument that actually matters.

He said token usage is a useful dashboard for getting people to experiment with AI, but not a perfect measure of productivity. That sounds obvious. It is not obvious inside companies. Inside companies, the second a dashboard exists, people start treating it like scripture.

I learned this the dumb way. Years ago I ran a small product team and started tracking support response times. Within a week, replies got faster and solutions got worse. Bellissimo. We had optimized the number and made the customer experience more annoying. Tokenmaxxing has the same smell. Not productivity. Compliance theater with a nice analytics layer.

And to be fair to Hoffman, he is not saying highest token count wins. He is saying companies need people across functions actually using this stuff. Finance, recruiting, legal, product, sales. On that, I am with him. If half your company still treats AI like a toy for engineers and interns, you are going to get smoked by teams that build real intuition early.

But the second you make experimentation visible, ranked, and culturally loaded, you change the behavior. Now you are not rewarding judgment. You are rewarding visible enthusiasm.

That is the whole game.

What the Reid Hoffman tokenmaxxing debate is really measuring

The funniest thing about tokenmaxxing is that everyone pretends it is a productivity metric when it is mostly a social metric.

What companies want is not a clean measure of output. They want proof that employees got the memo. AI is the mandate now. The religion. The annual-plan bullet point. The thing you are supposed to be seen touching. If you are a manager trying to drag an org into a new workflow, a usage dashboard is incredibly seductive because it gives you a neat little picture of who looks bought in.

That is why the Reid Hoffman tokenmaxxing debate matters more than the meme. The question is not whether more token usage is good or bad. That is baby-brain framing. The real question is what companies are actually measuring when they celebrate usage. And the answer is usually some cocktail of curiosity, obedience, panic, and metric gaming.

I have seen this movie with different props. First it was inbox zero. Then Slack responsiveness. Then calendar Tetris, where people somehow convinced themselves that being triple-booked meant they were important. Tokenmaxxing is just the AI version. Same neurosis, nicer branding.

The word itself is a tell. Tokenmaxxing comes from the same internet swamp as looksmaxxing, sleepmaxxing, and all the other flavors of optimization cosplay. Once that language enters the office, management is not just rolling out software. It is adopting an identity. It is saying: serious people here are visibly AI-pilled.

And once that happens, the metric stops being neutral. It becomes a status signal.

Meta built the perfect bad incentive machine

The Meta example is so absurd it loops back around to useful.

According to Fortune, Meta had an internal dashboard called Claudeonomics, tracking the top 250 AI token users out of a workforce of more than 85,000. It handed out fantasy-league titles like Token Legend and Cache Wizard. If you had told me this in 2019, I would have assumed you got too high at a founder retreat in Tahoe.

But of course this happened. It was almost inevitable.

The best detail was that neither Mark Zuckerberg nor Andrew Bosworth ranked in the top 250. I laughed. Then I stopped laughing, because that is exactly how power works. The people at the top never have to perform devotion to the new tool the same way everyone else does. When you are the CEO, your job is to declare the religion, not kneel the hardest.

That is what these leaderboards really expose. Not who gets value from AI, but who feels pressure to be seen using it.

Fortune also reported that the dashboard was shut down after two days once the outside world noticed. Two days. That tells you everything. If this were just a boring internal productivity tool, nobody would have panicked. Instead it looked a little too much like what it was: a corporate status game with token billing attached.

Then there were the numbers. In a 30-day period, employee usage reportedly exceeded 60 trillion tokens. The top user averaged 281 billion tokens. Using the cheapest quoted Claude pricing, that one person alone could have represented more than $1.4 million in cost.

One person.

At that point we are not talking about experimentation. We are talking about either a runaway agent farm, a broken incentive, or somebody accidentally opening a portal to the sun.

Fortune also noted that some employees had AI agents running for hours to maximize token usage. And every founder who has ever watched a metric get gamed immediately understands what happened. If you tell smart, ambitious people that a number matters, they will optimize the number. Not the outcome. The number.

That is not evil. That is just systems.

And honestly, I do not even blame the employees. If your company is loudly saying AI-native behavior matters, performance reviews are being rewritten around AI-driven impact, and there is a visible token leaderboard floating around, you would be irrational not to respond. Token Legend starts as a joke and ends as career insurance.

That is the uncomfortable part. The KPI may be garbage, but the signal is real.

High token usage can mean almost anything

This is where the online discourse gets lazy. People talk about token counts like they obviously mean one thing. They do not.

High token usage can mean a designer is genuinely learning how to prototype faster. It can mean an engineer has six agents chewing through a codebase overnight. It can mean somebody writes bad prompts and needs ten retries to get one useful answer. It can mean someone found a loophole and let Claude burn money like a Roman candle.

Hoffman basically admitted this in the TechCrunch story, which is why I do not think he is being naive. He said token usage is not a perfect example of productivity and that some people will use lots of tokens in random or exploratory ways. Exactly. That caveat is not a footnote. It is the whole story.

If the metric needs that much context to mean anything, then the metric alone is mostly theater.

I only care about one version of this conversation: does higher AI usage actually produce better work? Did support quality improve? Did legal review cycles get shorter? Did sales reps spend less time doing admin sludge? Did product teams ship faster without making the product worse? Show me that. Otherwise spare me the dashboard screenshots and the we are an AI-first organization chest beating.

Because startups are full of metrics that feel precise while hiding low-quality behavior. Daily active users can be junk. Time in app can be junk. Meetings booked can be junk. Tokens are exactly the same. Useful only when tied to outcomes. Dangerous when treated like outcomes.

There is another wrinkle people keep glossing over: AI agents can burn tokens for hours with almost no human labor involved. So when someone tops a leaderboard, what exactly are we praising? Skill? Creativity? Initiative? Their willingness to let autonomous loops run until finance starts sweating through a Patagonia vest?

That ambiguity matters.

I had my own mini version of this earlier this year. I was in the middle of a product sprint, half-dead, bouncing between New York and Lisbon, sleeping badly, eating like an idiot, and I started using AI for way too many tiny decisions because my brain was cooked. Draft this. Summarize that. Rewrite the copy again. On paper I looked insanely productive. In reality I was outsourcing medium thinking so I could avoid hard thinking.

That is the personal version of tokenmaxxing.

Good AI use sharpens judgment. Bad AI culture replaces judgment with volume.

Reid Hoffman discussing Tokenmaxxing and AI work metrics at a tech conference, engaging audience with insights.

Follow the money and the joke stops being funny

Tokenmaxxing is hilarious right up until someone in finance opens the bill.

According to the Ramp AI Index, more than half of businesses now pay for AI services, up sharply from a year earlier. AI spend also reportedly quadrupled from February 2025 to February 2026. So this is not some weird Silicon Valley side quest anymore. This is becoming standard operating expense.

And AI pricing is messier than old-school SaaS. Seat-based software is boring in a way I have come to appreciate with age. You know how many employees you have. You know the contract. You know roughly what next month looks like. Usage-based AI pricing is chaos by comparison. Harder to forecast. Easier to hide inside workflows. Much easier to blow up when one team discovers agents and decides to experiment.

Ramp found that for the biggest spenders, AI costs jump 50% or more in about one out of every four months.

That is not a rounding error. That is a planning problem.

I have sat through enough budget conversations to know the script. First the company says AI adoption is strategic. Then teams start stacking tools on top of tools, ChatGPT Enterprise, Claude, coding agents, meeting summarizers, internal copilots, random niche workflows someone in ops swears are mission-critical. Nobody wants to be the person asking whether all this spend is producing anything measurable, because that sounds anti-innovation. Then six months later the CFO shows up with murder in their eyes.

And there is a structural reason for this. When software vendors make money from seats, they want adoption. When AI vendors make money from tokens, they want adoption and consumption. The more your employees poke the machine, the better the revenue graph looks.

So tokenmaxxing is not just a quirky worker behavior. It is a business model feature wearing a culture hat.

That does not mean the labs are wrong when they say experimentation matters. They are right. Real workflow change does come from people trying things. But if you think the pricing model does not shape the messaging, mamma mia, I have a bridge in Naples to sell you.

Labs love your experimentation

This is the part everyone politely avoids saying out loud.

Every executive telling employees to experiment more with AI is also, intentionally or not, feeding the revenue engine of the labs selling the models. That does not make the advice wrong. It just means we should stop pretending the incentives are clean.

Ramp noted that Anthropic said the number of business customers spending over $1 million annually doubled to 1,000. That is not a cute little trend. That is a serious enterprise revenue machine. Meanwhile the big AI labs are raising absurd amounts of money because training and inference are brutally expensive and everybody knows it.

So yes, of course they want more enterprise usage. Of course they want every team routing more tasks through models. Why would they not?

The awkward part is quality. If higher token burn comes from better workflows, great. If it comes from retries, lower reliability, sloppy prompting, or agents wandering around unsupervised like drunk tourists, that is a very different story. More tokens can mean more value. More tokens can also mean more waste.

Tech loves simple metrics because simple metrics make leadership decks look clean. But the system underneath is messy. The employee gets rewarded for visible adoption. The manager gets to say their team is AI-forward. The vendor gets more revenue. The board gets a story. The only person left asking annoying questions is finance, and nobody invites finance to the afterparty.

I am joking, but barely.

Because there is a real strategic risk here: companies may confuse AI fluency with AI consumption. Those are not the same thing. A fluent team knows when to use a model, when to use a smaller model, when to stop an agent loop, when to verify the output, and when to just write the damn SQL query themselves.

That restraint is not sexy.

It might also be the moat.

America turned tokens into a meme. China is turning them into infrastructure

The U.S. version of this debate is extremely American. Slack jokes. Leaked dashboards. VCs arguing about whether office workers should become prompt goblins. A little Black Mirror, a little Succession, a little please clap for my AI adoption strategy.

Meanwhile China is treating tokens like industrial infrastructure.

According to Fortune, China now has an official word for token: ciyuan. The term was introduced by Liu Liehong, head of the National Data Administration, who described tokens as the settlement unit linking technological supply with commercial demand. That phrase alone tells you the difference in framing. In the U.S., tokens are becoming a workplace flex. In China, they are being framed as an economic layer.

Fortune also reported that China is processing 140 trillion tokens per day, up from 100 billion at the start of 2024. The jump is so absurd it almost reads like a typo, but the broader point stands: they are thinking at the level of systems, not office vibes.

That includes the model ecosystem. Alibaba’s Qwen, ByteDance’s Doubao, MiniMax, Zhipu AI, Biren, this is not one company doing one flashy thing. It is a broader buildout. Doubao reportedly hit 100 million daily active users over Lunar New Year. That is not a leaderboard. That is an economy.

I was in Shanghai years ago for a startup event, and what struck me was not that people were more optimistic than in San Francisco. It was that they were more operational. Less what is the discourse and more what infrastructure are we building. Maybe that is why this contrast bugs me. In America we keep turning everything into a cultural signal. Somewhere else, people are trying to define the unit economics of the future stack.

That does not mean China’s approach is magically better. They have their own capital intensity, cost pressure, and political weirdness. But the framing is still smarter. Less obsessed with whether your PM is a Token Legend. More focused on what tokenized AI usage means for platforms, payments, infrastructure, and industrial deployment.

And if I am being blunt, the U.S. debate can feel embarrassingly small by comparison.

We are arguing over the office fantasy league while another market is trying to standardize the scoreboard.

The real flex will be token yield

What the TechCrunch story gets right, and what a lot of the backlash misses, is that token usage is not productivity, but it is not meaningless either. It reveals who is experimenting, who is adapting, who is gaming the system, who is panicking, and which companies are quietly turning AI spend into a status competition.

That is useful information.

It is just not the information most executives pretend it is.

The companies that win will not be the ones with the highest token counts. They will be the ones that teach people when not to spend them. Which model to use. How to prompt tightly. When to trust the output. When to verify it. When to kill the agent. When to do the work yourself because involving AI adds more overhead than value.

That is the adult version of AI adoption. Naturally, it is much less sexy on an internal dashboard.

My bet is that within a year, smart operators stop flexing token usage and start obsessing over token yield instead.

Useful work per token. Time saved per token. Revenue influenced per token. Bugs fixed per token. Shipped product per token. Pick your denominator, but make it real. Because once every company has an AI dashboard, the advantage stops being access and starts being judgment.

And judgment, mi dispiace, does not fit neatly into a leaderboard.

Sources

Related reading