BCG Executive Urges Firms to Increase AI Token Spending
Business Insider reports that Sylvain Duranton, global leader of BCG X at Boston Consulting Group, told Business Insider that companies should begin measuring their AI token usage and understand the risks of holding back too much on consumption. Duranton is quoted saying, "I think you need to start the pump," and he discussed an early phase where organisations may push heavy usage before tightening allocation. Business Insider frames the debate around a "Tokenmaxxing" culture and notes pressure on employees to increase AI token consumption as agents and new workflows change how organisations consume model inference.
What happened
Business Insider reports that Sylvain Duranton, global leader of BCG X at Boston Consulting Group, told Business Insider that companies should start measuring their AI token usage and understand the risk of being too restrictive. Duranton is quoted directly: "I think you need to start the pump." Business Insider also frames a discussion about a "Tokenmaxxing" culture and says there is pressure on employees to raise AI token consumption as AI agents expand usage.
Editorial analysis - technical context
In practical terms, AI tokens are the atomic billing unit for large language model inference, and rising adoption of agentic workflows materially increases token consumption. For practitioners, early token spend is both an observable cost signal and a leading indicator of which workflows rely on higher-context, higher-cost models versus cheaper alternatives. Instrumenting token meters, per-workflow attribution, and model-level cost tracking are common technical controls teams deploy to translate token usage into usable engineering metrics.
Industry context
Companies that accelerate token consumption to discover value often face a tradeoff between faster experimentation and heightened run-rate costs. Observed patterns in comparable enterprise AI rollouts include rapid prototyping on larger models, followed by optimization efforts that introduce model choice, caching, and hybrid on-premise/LLM pipelines to control spend. Finance-ML collaboration and guardrail tooling frequently emerge as scaling bottlenecks when token budgets become meaningful to P&Ls.
What to watch
- •growth in reported token spend and month-over-month usage trends inside enterprises
- •rollout of observability and chargeback mechanisms linking tokens to teams and products
- •procurement moves toward multi-model pricing, context-window negotiation, or fixed-commit contracts
- •reporting from vendors and consultancies on token economics and model cost-per-call benchmarks
Reported-source note
All direct quotes and paraphrased remarks about Duranton and the "Tokenmaxxing" framing are reported by Business Insider.
Scoring Rationale
The story is a notable advisory from a major consultancy that highlights operational and budgeting issues practitioners face when adopting LLMs. It is not a technical breakthrough or market-moving announcement, but it signals common scaling challenges for enterprise AI.
Practice interview problems based on real data
1,500+ SQL & Python problems across 15 industry datasets — the exact type of data you work with.
Try 250 free problems

