HN
Today

New arXiv policy: 1-year ban for hallucinated references

arXiv, the prominent preprint server, has announced a new policy imposing a one-year ban for submissions containing hallucinated references, a direct response to the surge of AI-generated content. Authors will be held fully responsible for their papers' contents, regardless of how they were generated, with subsequent submissions requiring prior acceptance at peer-reviewed venues. This move aims to curb academic 'slop' and maintain integrity in scientific publishing, sparking lively debate on AI's role in research and the nature of academic accountability.

389
Score
127
Comments
#10
Highest Rank
15h
on Front Page
First Seen
May 14, 9:00 PM
Last Seen
May 15, 11:00 AM
Rank Over Time
221516101615151620232426242529

The Lowdown

arXiv is stepping up its game against the rising tide of AI-generated academic malpractice, particularly targeting 'hallucinated' references. The new policy, outlined by Thomas G. Dietterich, asserts that authors bear full responsibility for all content in their submissions, irrespective of the generation method.

  • The Policy: Signing a paper means each author takes full responsibility for its contents.
  • The Problem: The proliferation of Large Language Models (LLMs) has led to an increase in papers containing factually incorrect or entirely fabricated references, often referred to as 'hallucinations.'
  • The Penalty: Authors found to have submitted papers with hallucinated references will face a one-year ban from arXiv. Following this ban, any future submissions will only be accepted if they have first been peer-reviewed and accepted by a reputable venue.
  • The Goal: This measure is designed to uphold the quality and trustworthiness of the arXiv platform, ensuring that researchers can rely on the accuracy of the literature posted there.

In essence, arXiv is drawing a clear line in the sand: while AI tools might assist in research and writing, the ultimate burden of verification and accuracy rests squarely on the human authors.

The Gossip

Academic Accountability & Quality Control

Many commenters enthusiastically endorse arXiv's new policy, viewing it as a critical step to preserve scientific integrity amidst the influx of AI-generated 'slop.' They argue that being responsible for cited works is a fundamental academic standard and that the policy will help filter out negligent submissions. The consensus is that if an author cannot verify their references, their work should not be published, and LLMs merely amplify a pre-existing problem of poor academic rigor.

Implementation Woes & Punishment Concerns

While generally supportive of the intent, some commenters raise practical concerns about the policy's implementation and the severity of its punishment. Questions arise about how arXiv will detect hallucinated references at scale and fairly, especially when multiple authors are involved. Critics also argue that the penalty—a one-year ban followed by a permanent requirement for peer-reviewed acceptance—is disproportionately harsh for what could be an honest mistake or negligence, rather than deliberate fraud, potentially crippling a researcher's career and undermining arXiv's role as a pre-print server.

Defining 'Fraud' & LLM Liability

A significant portion of the discussion revolves around the definition of 'fraud' versus 'negligence' in the context of AI-generated content. Some argue that including a hallucinated reference, whether intentional or not, constitutes fraud because it implies a false claim of having reviewed cited material. Others contend that fraud requires intent to deceive, and a simple oversight due to LLM errors is mere sloppiness or negligence. This debate highlights the evolving understanding of authorship and responsibility when AI tools are integrated into the research workflow, especially given the 'hallucination-prone' nature of current LLMs.