Per-token AI charges come to GitHub Copilot

As of 1st June 2026, GitHub Copilot will charge its users on the basis of the tokens they use, rather than a flat rate subscription model.

The model that’s seeing the shutters closed on it is, or rather was, simple to understand and use. Users were given a set number of ‘Premium Requests’ according to their subscription tier. A complex coding task that may have taken many hours to complete used one premium request. Posing a relatively trivial question also counted as a single premium request.

However, the change which is soon to affect GitHub Copilot users aligns the pricing models with those of API charges to large language models, more common among business plans. On the new GitHub Copilot pricing scheme, most requests will be measured according to the tokens used by, input to, and output from the LLM at the heart of Copilot.

The definition and cost of tokens

A token is often described as representing around three-quarters of a word. Thus, giving an LLM a text of 10,000 words to examine would equate to 12,000-13,000 tokens of content. In developer terms, if a body of code which Copilot were to examine (for refactoring or bug-hunting for example), comprised of 10,000 ‘words’ (expressions, statements, variable names, functions, and so on), then that would count towards their allotted tokens for the month.

Prompt text, as inputs, will also count, as will the outputs from Copilot.

The pricing tiers coming into effect next month remain pegged at their current levels, but instead of being allotted a number of queries per month, users are given ‘AI Credits’ to the same value. A base-tier Copilot Pro subscriber ($10pcm) will receive 1,000 credits, with GitHub saying that at present one AI Credit is worth one US cent.

The number of tokens each credit buys will depend on the model used, the input/output mix, the size of the cache (data held in the LLM’s memory for context), and feature requested. Thus, if a developer uses mostly simple queries, they are likely not to have to buy extra credits each month. Conversely, multi-agent queries about a complex, lengthy code base will empty the AI Credit account more quickly. Queries to the most-advanced frontier models will cost more than to the less-powerful.

GitHub’s pricing changes do include some compensatory benefits for users: Code completions (similar to a phone’s auto-complete function) and Next Edit suggestions will remain free.

The industry changes to per-token pricing

The changes to GitHub’s pricing model are in line with similar changes from other companies. Anthropic and OpenAI have now moved their enterprise customers to token-based billing. Unlike those two, however, Microsoft – owner of GitHub – is a profitable business overall, and has to date been able to subsidise the use of GitHub Copilot with revenues from other parts of the business, such as its software and cloud divisions.

Up until the change on 1st June, users will have been able to ‘spend’ between three and eight times the number of tokens their monthly subscription costs have covered, and incurred no penalty.

Microsoft’s move is a change that affects the very users that it was hoping to attract to Copilot’s features, immediately forcing new and existing users to become aware of their token spend per query – a figure that has been abstracted away by per-month subscriptions to date. The new billing model may make more economic sense from Microsoft’s point of view, but it discourages the exploration and testing that new users will want to do.

For businesses that deploy AI coding agents in their development teams, the cost implications of the industry-wide shift in pricing policies are significant. In the case of Uber, for instance per The Information [paywall], its CTO has said it had spent the year’s AI budget for 2026 already this year, pointing out that 11% of updates to Uber’s code are now written by AI agents. Uber primarily uses Anthropic’s Claude coding agents.

Outside the IT department, companies deploying AI automation should be aware that complex tasks, which may involve running agentic LLMs unsupervised for long periods, could soon be charged on a similar per-token use basis. Thus, the delivered efficiency gains from AI in the workforce will have to be measured against any rise in costs stemming from the AI vendors’ bills.

(Image source: Pixabay under licence.)

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and co-located with other leading technology events. Click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.

The post Per-token AI charges come to GitHub Copilot appeared first on AI News.

Scroll to Top