GitHub said Copilot Free, Pro, and Pro+ interaction data will train models by default from Apr. 24 unless users opt out, while private repo content at rest stays excluded. Teams should review per-user enforcement, enterprise coverage, and repo privacy settings before the change lands.

Posted by prefork
From April 24, 2026, GitHub will use interaction data (inputs, outputs, code snippets, context) from Copilot Free, Pro, and Pro+ users to train AI models unless opted out via settings. Copilot Business and Enterprise users unaffected. Previous opt-outs preserved. Data shared with affiliates like Microsoft, not third parties. Does not use private repo content at rest.
GitHub’s policy update says that from Apr. 24, 2026, it may use Copilot interaction data from Free, Pro, and Pro+ accounts to “train and enhance AI models” unless the user opts out. The covered data includes inputs, outputs, code snippets, and associated context. GitHub also says previous opt-outs are preserved, data may be shared with affiliates such as Microsoft, and it is “not third parties” receiving that data according to the same official post.
Posted by vmg12
Thread discussion highlights: - martinwoodward on GitHub Copilot training opt-out: For Free, Pro and Pro+ Copilot, if you don’t opt out then we will start collecting usage data of Copilot for use in model training... we do not train on private repo data at rest, just interaction data with Copilot. - martinwoodward on Copilot policy clarification: It wasn’t previously opt-in. Previously we didn’t do any training on usage... we’ve been training on our internal usage for just over a year. - munk-a on Org-wide controls: The only setting I'm seeing is on a per-user basis. Does anyone know how to blanket disable training on an organizational basis?
The most important boundary is that GitHub distinguishes between private repo content at rest and what users send through Copilot. In the HN discussion, GitHub VP Martin Woodward says “we do not train on private repo data at rest, just interaction data with Copilot,” and adds that this was not a switch from opt-in to opt-out so much as a new training use for usage data that previously was not used that way. That narrows the claim, but for engineers the operational effect is still that prompts, completions, snippets, and surrounding context from private-repo work can enter the training pipeline if a covered user leaves the default in place.
Posted by vmg12
For engineers, the key issue is that GitHub’s Copilot training policy changes affect interaction data from private repos, with GitHub saying it does not train on private repo contents at rest. The practical engineering questions in-thread are about org-level policy enforcement, which accounts are covered, and whether teams should migrate to self-hosted Git or add repo encryption to reduce exposure.
The immediate engineering question is enforcement. The HN core thread captures a practical complaint from teams that “the only setting I'm seeing is on a per-user basis,” with no obvious blanket org-wide disable in view. That matters because the policy split is by subscription tier: GitHub’s official post says Business and Enterprise are unaffected, but teams still have to sort out mixed-seat environments, contractor accounts, and whether personal Copilot subscriptions are touching company code.
Community reporting also surfaced how the control appears in product settings. Orosz’s thread points users to Settings → Privacy and shared a screenshot showing the “Allow GitHub to use my data for AI model training” toggle plus a note that changes can take “up to 30 minutes” to take effect. Separate HN comments collected in the fresh discussion sharpen the broader concern: one commenter argues the wording makes this effectively Microsoft data collection, while others raise GDPR timing and argue agent systems create security exposure “by design.” Those are reactions, not policy text, but they show where review is likely to land inside security and compliance teams.
Posted by prefork
Today’s new comments sharpen the complaint in two directions. One commenter argues the policy is effectively Microsoft collecting the data rather than GitHub, pointing to the wording as evidence of where the data really goes. Another fresh thread is more legal/security-focused: one commenter asks whether GDPR enforcement would ever happen quickly enough to matter, while another argues agentic systems create security problems by design rather than merely introducing incidental backdoors. These add a more explicit compliance and threat-model angle to the existing backlash.
Posted by prefork
For AI engineers and developer-tool users, the key issue is that Copilot Free/Pro interactions can become training data by default unless users opt out, which affects workflow privacy, enterprise boundaries, and legal/compliance review for teams using AI coding assistants.
Absolute madness btw how I PAY not just for GitHub but also Copilot, and they STILL pull this "default opt in" I realize I haven't used Copilot in forever, so with this, I'm cancelling it (use Claude Code / Codex, sometimes Cursor, play w Factory AI) WTH, again