<https://arxiv.org/pdf/2406.10279>

Anche qui: <https://socket.dev/blog/slopsquatting-how-ai-hallucinations-are-fueling-a-new-class-of-supply-chain-attacks>

Abstract
The reliance of popular programming languages such as
Python and JavaScript on centralized package repositories
and open-source software, combined with the emergence of
code-generating Large Language Models (LLMs), has created
a new type of threat to the software supply chain: package
hallucinations. These hallucinations, which arise from fact-
conflicting errors when generating code using LLMs, repre-
sent a novel form of package confusion attack that poses a
critical threat to the integrity of the software supply chain.
This paper conducts a rigorous and comprehensive evaluation
of package hallucinations across different programming lan-
guages, settings, and parameters, exploring how a diverse set
of models and configurations affect the likelihood of generat-
ing erroneous package recommendations and identifying the
root causes of this phenomenon. Using 16 popular LLMs for
code generation and two unique prompt datasets, we generate
576,000 code samples in two programming languages that
we analyze for package hallucinations. Our findings reveal
that that the average percentage of hallucinated packages is at
least 5.2% for commercial models and 21.7% for open-source
models, including a staggering 205,474 unique examples of
hallucinated package names, further underscoring the severity
and pervasiveness of this threat. To overcome this problem,
we implement several hallucination mitigation strategies and
show that they are able to significantly reduce the number of
package hallucinations while maintaining code quality. Our
experiments and findings highlight package hallucinations as
a persistent and systemic phenomenon while using state-of-
the-art LLMs for code generation, and a significant challenge
which deserves the research community’s urgent attention.

Reply via email to