GitHub Copilot Moves to Token-Based Billing

Microsoft is transitioning GitHub Copilot from a requests-based subscription model to a token-based consumption billing system, rolling out in early June with an announcement expected April 23. The company paused new sign-ups for Pro, Pro+, and Student tiers, tightened rate limits, and removed Anthropic's Claude Opus models from lower-cost plans as part of cost-control measures. Under the new model users keep a base monthly subscription but consume token credits for actual input and output usage; example pricing cited includes roughly $2.50 per million input tokens and $15 per million output tokens for GPT-5.4 (other sources give roughly €2.30 and €14, respectively). Business and Enterprise plans will receive pooled credits: typical figures are $19/user/month with $30 pooled credits, and $39/user/month with $70 pooled credits. The move shifts billing risk to users and forces teams to add token accounting and cost controls into development workflows.
What happened
Microsoft is moving GitHub Copilot from a fixed-request subscription model to token-based consumption billing, with rollout slated for the beginning of June and an announcement expected April 23. Internal documents and multiple reporting outlets show the company has paused new sign-ups for Pro, Pro+, and Student tiers, tightened rate limits for individual accounts, and removed Claude Opus access from cheaper plans to rein in runaway compute costs. The weekly operating cost for Copilot reportedly nearly doubled since January, driving the change.
Technical details
The new billing charges for tokens separate input and output consumption and ties cost to model choice and token burn. Example figures disclosed include GPT-5.4 priced at about $2.50 per million input tokens and $15 per million output tokens in some reports, with parallel euro-denominated figures around €2.30 and €14 in others. Users will continue to pay a base monthly subscription but will receive token-credit allotments instead of fixed request counts. Reported enterprise allotments include $19/user/month with $30 pooled credits and $39/user/month with $70 pooled credits; pooled credits allow organization-wide sharing of token budgets. The change is being accompanied by tighter per-account rate limits and model availability restrictions for lower-priced tiers.
Implications for engineering teams
Token-based billing makes per-call costs visible, but it also forces teams to manage token budgets and optimize prompts, model selection, and caching. Practitioners should instrument token usage per pipeline, add cost-aware guardrails in CI/CD, and consider model fallbacks or smaller context windows for high-volume tasks. Expect these practical actions:
- •Implement tokens accounting and cost dashboards in observability tooling.
- •Add rate-limit backoffs, batching, and response truncation to reduce output token burn.
- •Re-evaluate model selection and use GPT-5.4 or cheaper variants for lower-value tasks.
Context and significance
This move follows a broader industry trend where providers shift from flat-rate subscriptions to consumption billing as model capabilities and per-call costs rise. Anthropic and other vendors recently tightened limits or moved enterprise customers to usage-based pricing. For Microsoft, the change reduces subsidy of heavy usage and aligns Copilot with cloud-native, consumption-based business models. For the developer economy, the timing matters: widespread adoption of LLM-powered coding assistants increased token burn substantially, and companies must now reconcile productivity gains with predictable budgets.
What to watch
Track the official pricing announcement for final token rates, how Microsoft handles legacy subscribers, and whether reduced access to models like Claude Opus persists. Practitioners should prepare by instrumenting token metrics, rehearsing budget alerts, and updating procurement policies for AI compute spend.
Bottom line
The shift moves cost transparency into the hands of users and teams, but it also adds operational overhead: token-metering, prompt engineering, and cost governance become first-class concerns for any organization embedding Copilot into engineering workflows. Expect short-term friction as companies adapt and long-term normalization toward consumption-based AI tooling.
Scoring Rationale
This is a notable product change that materially affects developer budgets and operational practices across teams using Copilot. It is not a model or research milestone, but it forces practitioners to implement cost governance and observability for LLM usage.
Practice with real Payments data
90 SQL & Python problems · 15 industry datasets
250 free problems · No credit card
See all Payments problemsStep-by-step roadmaps from zero to job-ready — curated courses, salary data, and the exact learning order that gets you hired.


