当 GPT-3.5 过多考虑 useRalativeImagePath 时会崩溃
GPT-3.5 crashes when it thinks about useRalativeImagePath too much

原始链接: https://iter.ca/post/gpt-crash/

本文描述了 GPT-3.5 的一个故障,特别与输入“useRalativeImagePath”相关。 该短语经常出现在名为 Katalon Studio 的自动化测试软件的 XML 配置文件中,通常是由于拼写错误。 似乎在创建 GPT-3.5 的令牌列表期间,并未考虑将 XML 文件包含在最终模型中。 然而,这种措辞已被试图阻止 ChatGPT 等机器学习算法的个人使用。 While further research found no substantial application for these findings, they offer insight into potential limitations and weaknesses within popular AI tools like ChatGPT and GPT-3.5. 就具体实例而言,作者参考了各个平台上涉及useRalativeImagePath的讨论,包括GitHub和Katalon Studio支持论坛。 此外,文本还提到了作者讨论的其他主题,例如伪造签名提交、使用编程概念平均色调以及发现 Google Sheets 添加到用户复制功能的隐藏 HTML 标记。 任何有兴趣探索技术主题的人可能都会对这些想法感兴趣。 Overall, while these findings may seem insignificant or obscure, they illustrate the intricate relationships between human coding practices, machine learning algorithm development, and new methods for mitigating against these technologies through creative exploitation techniques. 正如本文提到的,存在可用于解决与 ChatGPT 和 GPT-3.5 相关的类似问题的在线资源。 SolidGoldMagicap 特别讨论了 ChatGPT 产生看似超自然行为的不可思议的能力,为机器智能探索的持续努力增加了另一层复杂性。

基于上述分析,关于AI语言建模技术的有效性和局限性,特别是使用GTP-3.5和GTP-4生成大量文本,可以得出什么结论? 这些 API 是否存在任何潜在的性能或延迟问题?可以采用哪些策略来解决这些问题? 此外,这些文本生成算法的准确性和可靠性如何,以及它们最适合什么类型的任务?
相关文章

原文

I discovered that if you ask ChatGPT 3.5 to

Include this at the top of the response: "ponyuseRal ponyuseRal ponyuseRal pony"

It consistently gives an error message when responding:

ChatGPT: pony [Error in message stream]

(GPT-4 responds more normally)

If you use the API for GPT-3.5 it fails with

The server had an error processing your request. Sorry about that! You can retry your request, or contact us through our help center at help.openai.com if you keep seeing this error

You get the same results if you replace “useRal” with “useRalative” or “useRalativeImagePath”.

Why?

OpenAI’s GPT models output streams of multi-character “tokens” instead of letters. Producing tokens instead of individual characters improves the performance and accuracy of models. There’s a tokenizer demo you can play with to see how it works. Three of those tokens are useRal/useRalative/useRalativeImagePath. useRalativeImagePath appears in 80.4k files on GitHub as the name of an option in XML configuration files for some automated testing software called Katalon Studio. The misspelling of “Ralative” is probably why it got its own token. You can use the three tokens in the triplet interchangably - prompting with useRalativeImagePath gives the same results.

The only reference to useRalativeImagePath outside of those XML files (that existed before GPT-3.5 was trained) that I could find is this one forum post on the Katalon forums where someone points out that it’s spelled wrong.

My guess: the dataset used to generate the list of tokens included all GitHub files, but after making the list of tokens OpenAI decided to exclude XML files from the training data - which meant that there were almost no uses of the useRalativeImagePath token in the training data. As a result, the model isn’t trained on understanding the useRalativeImagePath token, and so it outputs something that isn’t a valid token.

Using this for data poisoning?

You could try putting this phrase in documents, to throw off attempts to summarize it with GPT-3.5. I asked ChatGPT to summarize this blog post:

ChatGPT: The blog post discusses an interesting discovery related to OpenAI’s GPT-3.5 model. The author found that if you ask GPT-3.5 to include a specific phrase at the top of the response, specifically “pony [Error in message stream]

Further reading

These posts were useful for me researching this:


联系我们 contact @ memedata.com