Anti-AI sentiment is spreading across Reddit and other online communities. Some subreddits now ban AI-generated content outright. Others are experimenting with model poisoning, deliberately corrupting public data to sabotage AI training. The pushback is especially loud from photographers and programmers who see their skills being devalued.
The activists have a problem. Model poisoning probably won't work. There's too much existing clean data, and AI companies have already moved on. Google pays Reddit $60 million yearly for direct API access. OpenAI cut deals with Axel Springer and the Financial Times. Why scrape the polluted public web when you can license clean data straight from publishers?
This shift fueled a new industry. Companies like Scale AI and Appen curate proprietary datasets while synthetic data creates closed training loops that bypass the open internet entirely.
None of this will stop AI. But the adversarial research coming out of this fight has real technical value. Techniques like prompt injection attacks and data contamination detection are now active areas of study at places like UC Berkeley and ETH Zurich. The doomed resistance is accidentally producing useful computer science, specifically technical roles like prompt engineering.