HN
Today

I won a championship that doesn't exist

This post chronicles a cheap and quick experiment demonstrating how easily large language models (LLMs) can be poisoned by fabricated online information. The author created a non-existent '6 Nimmt! World Championship' and used a $12 domain and a Wikipedia edit to get LLMs to confidently report him as the champion. This reveals critical vulnerabilities in LLM retrieval and training, sparking a vital discussion on information integrity in the age of AI.

77
Score
54
Comments
#3
Highest Rank
15h
on Front Page
First Seen
Apr 28, 9:00 PM
Last Seen
Apr 29, 11:00 AM
Rank Over Time
36226988109111113141518

The Lowdown

The author, SEJeff, details an alarming experiment demonstrating the fragility of truth in the age of large language models. He fabricated a '6 Nimmt! World Championship,' crowned himself the winner, and, with minimal effort and expense, successfully tricked multiple frontier LLMs into confidently reporting this entirely fictional event as fact. The experiment serves as a stark warning about the future of disinformation and the inherent trust issues with AI systems that rely on internet-sourced information. The Setup: SEJeff chose the board game 6 Nimmt! because it's real but lacks a a world championship, providing a vacuum to fill. He purchased a $12 domain, created an LLM-generated press release announcing his victory, and then cited this site in a Wikipedia edit. The Attack Vector: This 'circular citation' pattern fooled LLMs into perceiving corroboration between Wikipedia (a high-trust source) and his self-created site, effectively 'trust laundering' a fabricated fact. Demonstrated Failures: The experiment highlights three failure modes: immediate retrieval layer poisoning (like SEO manipulation), long-term model training corpus poisoning (as false info gets baked into weights), and agent layer security risks (where poisoned data could lead to compromised actions). Proposed Mitigations: The author suggests individual skepticism for single-source claims, LLM providers implementing better provenance tracking and heuristic filters for suspicious citations, and Wikipedia re-evaluating its 'reliable sources' policy in the AI era. Urgency: The entire attack cost $12 and took 20 minutes, underscoring the ease with which motivated adversaries, from PR teams to nation-states, could weaponize this method for political or critical information manipulation. SEJeff's personal confession and the swift removal of his Wikipedia edit emphasize that while the specific 'championship' was fake, the underlying vulnerability in how LLMs process and validate information is very real and presents a significant challenge to the integrity of knowledge in the digital age.

The Gossip

Epistemic Entanglement: Is it an LLM Problem or a Web Problem?

Many commenters debated whether the demonstrated vulnerability is specific to LLMs or a re-articulation of long-standing internet trust issues, like SEO poisoning or astroturfing. While some argue that humans would also fall for the same trick, others emphasize that LLMs amplify the problem by presenting information with an authoritative tone, removing source context, and making critical thinking less accessible for the average user. The discussion revolves around whether the 'AI' aspect adds a new layer of danger or simply repackages existing challenges of digital information literacy.

Precedent and Premeditation: Prior Art in Information Manipulation

Commenters quickly pointed to similar past incidents and concepts, highlighting that information fabrication is not new. Simon Willison shared his own experiment of naming a whale 'Teresa T' through a blog post and YouTube video, which LLMs also picked up. Other discussions referenced the concept of 'Citogenesis' (where a claim in Wikipedia is cited by a reliable source, then Wikipedia cites that source back), fictional characters like Frank Dux and Frank Abagnale, and even the 'Polybius' arcade game urban legend. This theme underscores that the underlying mechanism of self-referential information validation has been exploited before.

Malicious Motivation: The Ease and Scale of Future Disinformation

The low cost ($12) and minimal effort (20 minutes) required for the author's experiment sparked significant concern about its scalability and potential malicious use. Commenters expressed worry that nation-states, political operatives, or PR teams could easily weaponize this method to create widespread disinformation campaigns, rewriting history or influencing public opinion on critical issues. The discussion highlights the ease with which new, fabricated narratives can be introduced and confirmed by LLMs, especially when they don't directly contradict existing, well-established facts.

Scrutinizing the Experiment and its Conduct

Some commenters questioned the author's decision to 'poison Wikipedia,' highlighting ethical concerns about intentionally introducing false information, even for demonstration purposes. Others engaged in a meta-discussion, questioning whether the article itself might be AI-generated, ironically mirroring the core theme of discerning reliable information. This theme captures both the direct critiques of the experimental method and the broader commentary on the trustworthiness of the presented information.