A leaked internal document has exposed the data sources used to fine-tune Claude, Anthropic’s AI assistant, and it’s prompting new concerns about how today’s most powerful models are being shaped behind the scenes.
The document, reportedly created by third-party data-labeling firm Surge AI, included a list of websites that gig workers were instructed to use (and avoid) while helping Claude learn how to generate higher-quality responses.
The spreadsheet was stored in an open Google Drive folder and remained publicly accessible until Business Insider flagged it.
You may like
What the leak revealed
The spreadsheet included more than 120 “whitelisted” sites, such as:
Harvard.eduBloombergMayo ClinicThe National Institutes of Health (NIH)
Those were the trusted sources that Surge AI workers could pull from when crafting prompts and answers during Claude’s reinforcement learning phase (known as RLHF).
But the document also listed 50+ “blacklisted” sites; places workers were explicitly told to avoid. That list included major publishers and platforms like:
The New York TimesRedditThe Wall Street JournalStanford UniversityWiley.com
Why were these sites off-limits? While we don’t know for sure, it’s most likely due to licensing or copyright concerns, particularly notable given Reddit’s recent lawsuit against Anthropic over alleged data misuse.
Why it matters
Although the data was used for fine-tuning (not pre-training), the leak raises serious questions about data governance and legal risk in the AI industry.
Experts warn that courts may not draw a sharp line between training and fine-tuning data when evaluating potential copyright violations.
Surge AI quickly took the document offline after the leak was reported.
Anthropic, meanwhile, told Business Insider it had no knowledge of the list, which was reportedly created independently by its vendor.
Data control in the AI era
This isn’t the first time an AI vendor has mishandled sensitive training materials. Scale AI, another major player in the data-labeling space, faced a similar leak in past years.
But the stakes are higher now. With Anthropic valued at over $60 billion and Claude emerging as a top competitor to ChatGPT, every misstep invites scrutiny.
This event highlights a growing vulnerability in the AI ecosystem as companies rely more on human-supervised training, they also depend on third-party firms and those firms don’t always have airtight security or oversight.
What it means for you
AI users need to understand that the quality, accuracy and even the ethical grounding of their chatbot’s responses are deeply tied to the data it’s trained on and who decides what goes in or stays out.
This leak reveals that even top-tier models like Claude can be influenced by behind-the-scenes decisions made by third-party vendors.
When those choices involve inconsistent standards or unclear sourcing, it raises serious questions about bias, trust and accountability in the AI we rely on every day.
The takeaway
This leak is a glimpse into how major AI companies shape their models and the those guiding the process.
As AI becomes more embedded in everyday tools, trust will come down to transparency.
When it comes to this factor, it appears that there’s still a long way to go.
More from Tom’s Guide
Back to Laptops