Latest
Wayfair, Subconscious, Baseten, and Cloudflare Bring Live AI Agent Competition to Boston Tech WeekWayfair, Subconscious, Baseten, and Cloudflare Bring Live AI Agent Competition to Boston Tech Week|Jellyfish Says “Tokenmaxxing” Is Burning Cash Faster Than It Creates CodeJellyfish Says “Tokenmaxxing” Is Burning Cash Faster Than It Creates Code|AI Realized and AiGovOps Foundation Bring AI Governance Into the Operator EraAI Realized and AiGovOps Foundation Bring AI Governance Into the Operator Era|Supermoon’s Openclaw Workshop Signals the Next Operational Shift in AI FoundingSupermoon’s Openclaw Workshop Signals the Next Operational Shift in AI Founding|Boston AI Tech Week Opens With a Political Signal to the AI EconomyBoston AI Tech Week Opens With a Political Signal to the AI Economy|Vector Raises $10M Series A to Drag B2B Advertising Out of the FogVector Raises $10M Series A to Drag B2B Advertising Out of the Fog|Boston’s Tech Community Wants Sweat Equity Before Social EquityBoston’s Tech Community Wants Sweat Equity Before Social Equity|Ecomm’s CAC Problem and How to Fix It Signals a New AI-Commerce Reality in BostonEcomm’s CAC Problem and How to Fix It Signals a New AI-Commerce Reality in Boston|Mind Robotics Raises $400M as Industrial AI Stops Playing DefenseMind Robotics Raises $400M as Industrial AI Stops Playing Defense|Origin Lab Raises $8M Seed to Turn Video Games Into AI InfrastructureOrigin Lab Raises $8M Seed to Turn Video Games Into AI Infrastructure|Wayfair, Subconscious, Baseten, and Cloudflare Bring Live AI Agent Competition to Boston Tech WeekWayfair, Subconscious, Baseten, and Cloudflare Bring Live AI Agent Competition to Boston Tech Week|Jellyfish Says “Tokenmaxxing” Is Burning Cash Faster Than It Creates CodeJellyfish Says “Tokenmaxxing” Is Burning Cash Faster Than It Creates Code|AI Realized and AiGovOps Foundation Bring AI Governance Into the Operator EraAI Realized and AiGovOps Foundation Bring AI Governance Into the Operator Era|Supermoon’s Openclaw Workshop Signals the Next Operational Shift in AI FoundingSupermoon’s Openclaw Workshop Signals the Next Operational Shift in AI Founding|Boston AI Tech Week Opens With a Political Signal to the AI EconomyBoston AI Tech Week Opens With a Political Signal to the AI Economy|Vector Raises $10M Series A to Drag B2B Advertising Out of the FogVector Raises $10M Series A to Drag B2B Advertising Out of the Fog|Boston’s Tech Community Wants Sweat Equity Before Social EquityBoston’s Tech Community Wants Sweat Equity Before Social Equity|Ecomm’s CAC Problem and How to Fix It Signals a New AI-Commerce Reality in BostonEcomm’s CAC Problem and How to Fix It Signals a New AI-Commerce Reality in Boston|Mind Robotics Raises $400M as Industrial AI Stops Playing DefenseMind Robotics Raises $400M as Industrial AI Stops Playing Defense|Origin Lab Raises $8M Seed to Turn Video Games Into AI InfrastructureOrigin Lab Raises $8M Seed to Turn Video Games Into AI Infrastructure
Back to articles

Jellyfish Says “Tokenmaxxing” Is Burning Cash Faster Than It Creates Code

Jellyfish analyzed 12,000 developers across 200 companies and found extreme AI token usage drives sharply diminishing returns in software engineering.

Software teams spent the past 18 months treating AI coding tools like casino chips with autocomplete. More prompts. More tokens. More tabs. More “velocity.” Venture-backed startups called it acceleration while enterprise leaders called it transformation, and finance teams quietly started pricing antacids into quarterly planning. Now Jellyfish has dropped data into the middle of the debate, and the findings land like a steel chair in a board meeting.

Boston-based Jellyfish, the engineering intelligence platform founded by Andrew Lau, David Gourley, and Philip Braden, published a new research report titled Is ‘tokenmaxxing’ cost effective? New data from Jellyfish explains. The company analyzed usage patterns from 12,000 developers across 200 companies during Q1 2026, and the conclusion was blunt: higher AI token consumption increases output, but the economics collapse fast at the extreme end of usage. That matters because enterprise AI adoption is entering a different phase now. The first wave rewarded experimentation while the second wave rewards accountability. Software organizations are no longer asking whether developers should use AI coding tools. They are asking whether the productivity gains justify the operational cost structure forming underneath them.

The AI coding market suddenly looks less like a sprint and more like a procurement negotiation with a GPU addiction.

What Happened

Jellyfish’s research focused on what the company calls “tokenmaxxing,” a term describing aggressive usage of AI coding assistants and large language models inside software development workflows. The dataset covered 12,000 developers across 200 companies during Q1 2026, with Nicholas Arcolano, Ph.D., Jellyfish’s research lead, analyzing token usage alongside developer output measured through merged pull requests.

The median developer reportedly used about 51M AI tokens per month while developers in the 90th percentile consumed roughly 380M tokens monthly. Using published Claude API pricing, Jellyfish estimated the median monthly cost at $52.38 compared to $691.14 for heavy users. Then the report connected token spend to output, and the economics started getting uncomfortable.

Developers in the bottom 20% of token spending reportedly used about $3 worth of tokens during the quarter and shipped 11 merged pull requests, while developers in the top 20% spent roughly $1,822 and shipped 23 merged pull requests. Output doubled while costs exploded. Jellyfish calculated cost per merged pull request increasing from $0.28 to $89.32 between low and high token usage groups, transforming AI coding from a productivity story into a margin story. This is the moment where engineering leaders stop sounding like futurists and start sounding like portfolio managers.

Why This Matters

The AI coding boom created a strange cultural reflex inside software teams where usage itself became a proxy for sophistication. A developer consuming 300M tokens started sounding impressive the same way somebody bragging about cloud spend sounded impressive in 2017. Then invoices arrived.

The Jellyfish report matters because it introduces operational gravity into a market drowning in enthusiasm. AI coding tools like Cursor, Claude, GitHub Copilot, and Windsurf generated enormous excitement by compressing iteration cycles and reducing repetitive engineering work. Those gains are real, but software organizations still need a framework for measuring efficiency instead of activity. That distinction separates sustainable AI adoption from what amounts to computational binge drinking.

Jellyfish is positioning itself directly inside that emerging category: AI productivity observability. The company already built its reputation helping engineering organizations quantify delivery performance and business impact, and this research extends that thesis into AI infrastructure economics. Translation for enterprise buyers: somebody now has to explain why token spend is climbing faster than engineering throughput.

Market Context

The timing behind Jellyfish’s report is not accidental. AI coding adoption accelerated dramatically across enterprise software organizations throughout 2025 and early 2026 as Anthropic, OpenAI, GitHub, Google, and a wave of AI-native coding startups pushed increasingly powerful coding copilots into daily engineering workflows. Productivity headlines exploded across LinkedIn while engineers posted screenshots showing entire functions generated in seconds. Founders talked about smaller engineering teams shipping larger roadmaps, and investors started whispering about “10x developers” again like the industry had learned absolutely nothing from the last cycle.

Underneath the hype sat a quieter trend: token consumption scaling at enterprise velocity. Large language models are computationally expensive, and heavy inference usage compounds quickly across hundreds or thousands of developers. The more context windows expand and autonomous coding agents emerge, the more operational costs rise alongside them. That dynamic explains why infrastructure providers suddenly occupy strategic positions in AI markets. NVIDIA sells the picks and shovels while OpenAI and Anthropic meter intelligence like utilities, and engineering organizations absorb the bill in real time.

Jellyfish’s research introduces a missing layer into that ecosystem: measurement. Not measurement of possibility. Measurement of efficiency. That subtle difference is becoming one of the defining themes in enterprise AI adoption.

Competitive Landscape

Jellyfish operates inside the engineering intelligence market, competing broadly with platforms focused on developer productivity analytics, workflow visibility, and engineering operations. The company raised roughly $114.5M across funding rounds including a $71M Series C announced in 2022 backed by Accel, Insight Partners, Tiger Global, and Wing Venture Capital. Jellyfish customers publicly referenced in company materials include Toast, SessionM, Jobvite, Bazaarvoice, and Digital Guardian.

The competitive pressure now extends beyond traditional engineering management software. GitHub Copilot owns developer mindshare while Cursor dominates technical conversation among startup engineers. Anthropic’s Claude models increasingly power coding workflows because of their large context windows and reasoning performance, while OpenAI continues pushing coding-oriented GPT capabilities deeper into enterprise environments. That creates a secondary market opportunity for governance, analytics, and AI cost observability platforms.

Somebody has to measure the consequences after the demos end. Jellyfish appears determined to own part of that layer.

What This Signals

The most important insight inside the Jellyfish report is not that AI coding costs money because sophisticated operators already understood that. The real signal is that software engineering is entering an optimization era. During the early cloud boom, companies celebrated migration. Years later, FinOps emerged because organizations realized unlimited infrastructure scaling eventually creates terrifying invoices. AI development appears headed toward the same maturation cycle.

First comes adoption. Then comes optimization. Then comes governance. The companies that survive long term usually arrive at stage three before the bill becomes existential.

Jellyfish’s report also hints at a broader philosophical shift around AI productivity itself. More generation does not automatically equal more leverage. Bigger output does not guarantee stronger engineering outcomes. Sometimes a machine simply produces more volume with greater confidence. Anybody who survived the “growth at all costs” era should recognize the pattern immediately.

Frequently Asked Questions

What is Jellyfish?

Jellyfish is a Boston-based engineering intelligence and developer productivity platform founded in 2017 by Andrew Lau, David Gourley, and Philip Braden.

What is “tokenmaxxing”?

“Tokenmaxxing” refers to aggressively maximizing AI token usage inside coding workflows to increase software development output.

How many developers were included in Jellyfish’s research?

Jellyfish analyzed data from 12,000 developers across 200 companies during Q1 2026.

What did Jellyfish discover about AI coding costs?

Jellyfish found that high AI token usage increased output but created sharply rising marginal costs, with cost per merged pull request climbing from $0.28 to $89.32.

Who authored the Jellyfish research?

The report was authored by Nicholas Arcolano, Ph.D., a research leader at Jellyfish focused on AI and engineering analytics.

Why does the Jellyfish report matter to enterprises?

The report highlights growing operational and financial pressure tied to enterprise AI coding adoption, especially around token efficiency, infrastructure cost management, and engineering productivity measurement.