GitHub Copilot’s Usage‑Based Billing Lands June 1 — What .NET & Azure Teams Need to Rewire Now
TL;DR: GitHub Copilot is switching all plans to usage‑based billing on June 1, 2026, priced by tokens rather than “requests.” If you ship .NET on Azure, this quietly changes how you budget, throttle, and even design Copilot‑powered workflows. You’ll want guardrails before finance notices the spike.
The update, precisely (no vibes, just facts)
GitHub announced that all Copilot plans will transition to usage‑based billing starting June 1, 2026. Instead of counting “premium requests,” each plan gets a monthly allotment of GitHub AI Credits, with overages billed based on token consumption (input, output, and cached tokens) at model‑specific rates. (github.blog)
In parallel, GitHub confirmed the deprecation of GPT‑5.2 and GPT‑5.2‑Codex across most Copilot experiences on June 1, 2026, with clear upgrade paths to newer models. (github.blog)
Those two announcements together mean: cost now scales with how you use Copilot, and model choice matters more than ever.
Why this matters to .NET & Azure engineers
1) Cost moves from “seat math” to “token math”
Under the new system, a long Copilot Chat session that pastes a full solution file costs more than a terse inline completion. Multi‑agent workflows and background agents (now common in Visual Studio and GitHub) can quietly burn tokens.
Action: Treat Copilot like any other metered cloud dependency.
- Set org‑level policies for model usage and features.
- Educate teams on token‑heavy patterns (huge prompts, repeated context).
2) Model selection is now a budget decision
With GPT‑5.2 retiring and newer models recommended, you’ll need to verify latency vs. cost trade‑offs for your workflows. Heavier reasoning models can be fantastic—and expensive—when used indiscriminately.
Action: Standardize defaults.
- Use lighter models for inline completions.
- Reserve heavier models for explicit “think hard” tasks (reviews, refactors).
3) Agentic workflows amplify both value and spend
Recent Copilot updates lean hard into agentic workflows—cloud agents, debugger agents, and background tasks that run longer and touch more context. They’re powerful, but they multiply token usage.

Action: Add guardrails.
- Prefer scoped context (specific files/PRs) over repo‑wide prompts.
- Kill runaway agents with time/token limits where available.
Practical steps for teams shipping on .NET & Azure
Audit current usage (this week)
GitHub has expanded Copilot usage metrics, including agent fields. Use them to establish a baseline before June 1. (github.blog)
Update internal guidance
Document:
- Approved models per scenario.
- When to use Copilot Chat vs. inline completions.
- How to avoid pasting entire repos “just to be safe.”
Align with Azure cost discipline
If you already manage Azure OpenAI or other metered AI services, reuse the same playbook:
- Budgets and alerts.
- Monthly reviews.
- Clear owners for AI spend.
The difference now is that developer tooling itself joins the cost conversation.
What this is not
This isn’t Copilot getting worse, slower, or “enterprise‑taxed.” It’s GitHub aligning Copilot with how AI infrastructure actually costs money—tokens in, tokens out—and making that visible. Handled well, it encourages better prompts, better agents, and fewer “why is this so slow?” moments.
Handled poorly… well, see the meme above.
Further reading
- https://github.blog/news-insights/company-news/github-copilot-is-moving-to-usage-based-billing/
- https://github.blog/changelog/2026-05-01-upcoming-deprecation-of-gpt-5-2-and-gpt-5-2-codex/
- https://github.blog/changelog/
- https://github.blog/changelog/label/copilot/