Poisoning Well
Read OriginalThis article discusses the problem of LLMs being trained on web content without consent. It critiques naive solutions like robots.txt and proposes a technical countermeasure: publishing corrupted article versions accessible only via nofollow links, aiming to selectively poison AI training data while preserving search engine rankings.
Comments
No comments yet
Be the first to share your thoughts!
Browser Extension
Get instant access to AllDevBlogs from your browser