代码中的幻觉是LLM错误的最小危险形式
Hallucinations in code are the least dangerous form of LLM mistakes

原始链接: https://simonwillison.net/2025/Mar/2/hallucinations-in-code/

开发人员经常抱怨LLMS幻觉代码,例如发明不存在的方法,但这是最不危险的问题。真正的风险在于编译器错过的微妙错误。当代码运行时,幻觉方法很明显,提供了“免费”事实检查。这允许手动或通过LLM本身进行快速更正。 作者强调,手动测试代码至关重要,即使它看起来完美无缺。 LLM会产生令人信服的代码,创造出一种错误的安全感,类似于未经检查的AI生成的文本。始终执行和验证代码的功能。为了减少幻觉,请尝试不同的模型,通过提供示例来利用环境,并选择良好的库。作者认为,那些声称审查LLM代码的人比编写较慢,缺乏足够的代码审查技能。 LLM生成的代码在该领域提供了出色的实践。最后,作者强调了LLM在审查和完善自己的写作方面的有用性。

该黑客新闻线程讨论了原始文章的说法,即代码幻觉是最不危险的LLM错误,许多评论者不同意。一个普遍的情绪是,由于缺乏意图和一致性,审查LLM生成的代码可能比自己写的更耗时和困难。他们还指出,审稿人不知道有关问题空间的微妙见解,该信息空间告诉了实际编写代码的开发人员。人们担心他们可能会接受看起来很棒但包含微妙的错误或逻辑缺陷的代码。一些评论者发现LLMS对特定任务(例如样板生成)有用,而另一些评论者则报告了诸如难以管理的功能,不正确的内存分配以及工具在修改代码时忘记关键功能的问题。审查和测试LLM代码比以往任何时候都更为重要。
相关文章
  • (评论) 2025-03-04
  • (评论) 2024-02-26
  • (评论) 2024-06-13
  • 幻觉是不可避免的:大型语言模型的先天局限性 2024-02-26
  • (评论) 2025-02-28

  • 原文

    2nd March 2025

    A surprisingly common complaint I see from developers who have tried using LLMs for code is that they encountered a hallucination—usually the LLM inventing a method or even a full software library that doesn’t exist—and it crashed their confidence in LLMs as a tool for writing code. How could anyone productively use these things if they invent methods that don’t exist?

    Hallucinations in code are the least harmful hallucinations you can encounter from a model.

    The real risk from using LLMs for code is that they’ll make mistakes that aren’t instantly caught by the language compiler or interpreter. And these happen all the time!

    The moment you run LLM generated code, any hallucinated methods will be instantly obvious: you’ll get an error. You can fix that yourself or you can feed the error back into the LLM and watch it correct itself.

    Compare this to hallucinations in regular prose, where you need a critical eye, strong intuitions and well developed fact checking skills to avoid sharing information that’s incorrect and directly harmful to your reputation.

    With code you get a powerful form of fact checking for free. Run the code, see if it works.

    In some setups—ChatGPT Code Interpreter, Claude Code, any of the growing number of “agentic” code systems that write and then execute code in a loop—the LLM system itself will spot the error and automatically correct itself.

    If you’re using an LLM to write code without even running it yourself, what are you doing?

    Hallucinated methods are such a tiny roadblock that when people complain about them I assume they’ve spent minimal time learning how to effectively use these systems—they dropped them at the first hurdle.

    My cynical side suspects they may have been looking for a reason to dismiss the technology and jumped at the first one they found.

    My less cynical side assumes that nobody ever warned them that you have to put a lot of work in to learn how to get good results out of these systems. I’ve been exploring their applications for writing code for over two years now and I’m still learning new tricks (and new strengths and weaknesses) almost every day.

    Manually testing code is essential

    Just because code looks good and runs without errors doesn’t mean it’s actually doing the right thing. No amount of meticulous code review—or even comprehensive automated tests—will demonstrably prove that code actually does the right thing. You have to run it yourself!

    Proving to yourself that the code works is your job. This is one of the many reasons I don’t think LLMs are going to put software professionals out of work.

    LLM code will usually look fantastic: good variable names, convincing comments, clear type annotations and a logical structure. This can lull you into a false sense of security, in the same way that a gramatically correct and confident answer from ChatGPT might tempt you to skip fact checking or applying a skeptical eye.

    The way to avoid those problems is the same as how you avoid problems in code by other humans that you are reviewing, or code that you’ve written yourself: you need to actively exercise that code. You need to have great manual QA skills.

    A general rule for programming is that you should never trust any piece of code until you’ve seen it work with your own eye—or, even better, seen it fail and then fixed it.

    Across my entire career, almost every time I’ve assumed some code works without actively executing it—some branch condition that rarely gets hit, or an error message that I don’t expect to occur—I’ve later come to regret that assumption.

    Tips for reducing hallucinations

    If you really are seeing a deluge of hallucinated details in the code LLMs are producing for you, there are a bunch of things you can do about it.

    • Try different models. It might be that another model has better training data for your chosen platform. As a Python and JavaScript programmer my favorite models right now are Claude 3.7 Sonnet with thinking turned on, OpenAI’s o3-mini-high and GPT-4o with Code Interpreter (for Python).
    • Learn how to use the context. If an LLM doesn’t know a particular library you can often fix this by dumping in a few dozen lines of example code. LLMs are incredibly good at imitating things, and at rapidly picking up patterns from very limited examples. Modern model’s have increasingly large context windows—I’ve recently started using Claude’s new GitHub integration to dump entire repositories into the context and it’s been working extremely well for me.
    • Chose boring technology. I genuinely find myself picking libraries that have been around for a while partly because that way it’s much more likely that LLMs will be able to use them.

    I’ll finish this rant with a related observation: I keep seeing people say “if I have to review every line of code an LLM writes, it would have been faster to write it myself!”

    Those people are loudly declaring that they have under-invested in the crucial skills of reading, understanding and reviewing code written by other people. I suggest getting some more practice in. Reviewing code written for you by LLMs is a great way to do that.


    Bonus section: I asked Claude 3.7 Sonnet "extended thinking mode" to review an earlier draft of this post—"Review my rant of a blog entry. I want to know if the argument is convincing, small changes I can make to improve it, if there are things I've missed.". It was quite helpful, especially in providing tips to make that first draft a little less confrontational! Since you can share Claude chats now here’s that transcript.

    联系我们 contact @ memedata.com