GitHub Is About to Train AI on Your Code, and It's Opt-Out

If you use GitHub Copilot on a Free, Pro, or Pro+ plan, your code is about to become training data. Not hypothetically, not eventually: starting April 24, 2026, GitHub will begin feeding your interaction data into its AI models unless you go into your settings and explicitly turn it off. The announcement dropped quietly on March 25, buried inside a routine privacy statement update. The developer community's reaction has been anything but quiet.
What Exactly Is GitHub Collecting?
Let's be specific about what "interaction data" means here, because GitHub's language is broad. According to the updated privacy statement, the data they'll collect includes: code snippets you write, Copilot outputs you accept or modify, the code context surrounding your cursor position, comments and documentation, file names and repository structure, navigation patterns, interactions with Copilot features like chat and inline suggestions, and even your feedback signals like thumbs up or thumbs down ratings.
That's not just "anonymous usage metrics." That's a detailed map of how you write code, what code you write, and how you interact with AI assistance while doing it. And yes, this applies to private repositories too.
The Opt-Out Problem
Here's the part that has developers furious: the default is on. If you do nothing between now and April 24, GitHub will start collecting and using your data for model training. You have to actively navigate to your Copilot settings, find the Privacy section, and disable "Allow GitHub to use my data for AI model training."
This is the classic opt-out playbook that's become standard across Big Tech. Companies know that the vast majority of users never touch their settings. GitHub has over 100 million registered developers and roughly 4.7 million paid Copilot subscribers as of early 2026. Only a fraction of those users will ever see this policy change, let alone act on it.
To their credit, if you previously opted out of data collection for product improvements, GitHub says your preference has been retained. But for everyone else, the clock is ticking.
The Developer Backlash Is Real
The community response has been overwhelmingly negative. In the GitHub Community Discussion thread about the change, the reaction ratio tells the story: 59 thumbs-down votes and just three rocket emoji reactions. Among the dozens of comments, virtually no one besides Martin Woodward, GitHub's VP of Developer Relations, has endorsed the idea.
The core objection isn't complicated. Developers are being asked to hand over their work product to train tools that could eventually replace them, or at least make their skills less valuable. And they're being asked to do it for free, as a default, with 30 days' notice tucked inside a terms of service update.
There's also a deeper trust issue at play. Many developers chose GitHub partly because Microsoft, which acquired it in 2018 for $7.5 billion, promised to respect the developer community. Using private repository data to train commercial AI products feels like a significant departure from that promise.
Who's Exempt (and Why That Matters)
Not everyone is affected equally. Copilot Business and Enterprise tier users are completely excluded from the training data program. Students and teachers on GitHub's education plans are also exempt.
Think about what that signals for a second. GitHub is essentially saying: if your company pays the enterprise rate, your code stays private. If you're an individual developer or a small team on a Pro plan, your code is fair game. The people with the least leverage get the least protection, which is a pattern we've seen repeatedly in the AI data economy.
The Bigger Picture: AI's Insatiable Data Appetite
GitHub's move doesn't exist in a vacuum. It's part of a broader industry trend where every platform that generates user data is finding ways to funnel that data into AI training. Meta has been doing it with Instagram and Facebook posts. Google does it with virtually everything. Reddit sold its data firehose to Google for $60 million.
What makes the GitHub case particularly sensitive is the nature of the data. Code is intellectual property. It's the literal output of a developer's professional skill. And in many cases, code in private repositories contains proprietary business logic, trade secrets, or at least competitive advantages that developers and companies never intended to share.
GitHub's defense is that they use the data responsibly and that training on interaction patterns improves Copilot for everyone. That may even be true. But "we'll use it well" has never been a satisfying answer to "we didn't ask your permission."
How to Protect Yourself
If you want to opt out, here's exactly what to do before April 24:
Go to github.com/settings/copilot/features. Under the Privacy heading, find the toggle for "Allow GitHub to use my data for AI model training" and switch it off. That's it. GitHub says disabling this won't affect your access to any Copilot features.
For organizations, the calculus is different. If you're on a Business or Enterprise plan, you're already excluded. But if your team uses individual Pro accounts, you may want to send around a company-wide heads-up.
It's also worth noting that opting out only covers future data collection. GitHub hasn't been clear about whether data collected before April 24 under previous terms will be used differently going forward.
What to Watch
The real test comes after April 24. How many developers actually opt out? If the number is small, expect other developer tool companies to follow GitHub's playbook. If there's a mass exodus to GitLab or other alternatives, it could reshape the competitive landscape for AI-powered development tools.
There's also the legal dimension. The EU's GDPR has strict rules about consent and data processing, and an opt-out default for AI training may not satisfy regulators. The EU AI Act's enforcement bodies have already been issuing their first formal inquiries this month, and a high-profile case involving developer code could set important precedent.
For now, the ball is in developers' hands. You have until April 24 to decide whether your code helps train GitHub's next AI model, or whether you'd rather keep it to yourself. Either way, the fact that you have to make that choice at all says a lot about where the AI industry is headed.
References
Get the Daily Briefing
AI, Crypto, Economy, and Politics. Four stories. Every morning.
No spam. Unsubscribe anytime.