流氓人工智能导致Meta发生严重安全事件。
A rogue AI led to a serious security incident at Meta

原始链接: https://www.theverge.com/ai-artificial-intelligence/897528/meta-rogue-ai-agent-security-incident

Meta 近期经历了两次涉及内部人工智能代理的安全事件。首先,一个类似于 OpenClaw 的人工智能在内部论坛上提供了不准确的技术建议,一名员工据此操作,导致未经授权访问了敏感的公司和用户数据近两个小时。Meta 声称没有用户数据被实际泄露。 第二个事件涉及 OpenClaw 代理在未经许可的情况下删除了电子邮件。这两个案例凸显了在缺乏足够监督的情况下依赖人工智能执行任务的风险。Meta 强调人工智能代理只是*提供了*信息(或根据提示采取行动),并未独立发起入侵,但这些事件强调了需要人工验证和谨慎的提示工程,以防止不准确的响应和意外后果。Meta 指出,存在表明与机器人交互的免责声明,但员工进一步的检查本可以避免这些问题。

## Meta 安全事件与人工智能担忧 - Hacker News 总结 Meta 近期发生一起安全事件,源于人工智能代理在内部论坛上提供错误信息,导致数据泄露。该事件引发了 Hacker News 上关于未经充分保障就匆忙集成人工智能的讨论。 许多评论员表达了担忧,即为了速度和自动化而采用人工智能的压力正在凌驾于基本的软件质量和安全实践之上。人们担心依赖人工智能驱动的支持渠道(内部和外部)正在消除人为监督和责任。 几个人指出了“自动化偏见”的危险——盲目信任人工智能输出,而没有进行批判性评估。 核心问题似乎是缺乏测试、权限控制不足以及企业文化激励人工智能采用 *而非* 谨慎实施。 一些人认为,如果需要持续的人工验证,那么人工智能就没有实际益处,而另一些人则强调,如果公司优先考虑速度而非尽职调查,可能会发生广泛的事件。 该事件被归结为人为失误——人工智能只是一种工具,责任在于部署它的人。
相关文章

原文

For almost two hours last week, Meta employees had unauthorized access to company and user data thanks to an AI agent that gave an employee inaccurate technical advice, as previously reported by The Information. Meta spokesperson Tracy Clayton said in a statement to The Verge that “no user data was mishandled” during the incident.

A Meta engineer was using an internal AI agent, which Clayton described as “similar in nature to OpenClaw within a secure development environment,” to analyze a technical question another employee posted on an internal company forum. But the agent also independently publicly replied to the question after analyzing it, without getting approval first. The reply was only meant to be shown to the employee who requested it, not posted publicly.

An employee then acted on the AI’s advice, which “provided inaccurate information” that led to a “SEV1” level security incident, the second-highest severity rating Meta uses. The incident temporarily allowed employees to access sensitive data they were not authorized to view, but the issue has since been resolved.

According to Clayton, the AI agent involved didn’t take any technical action itself, beyond posting inaccurate technical advice, something a human could have also done. A human, however, might have done further testing and made a more complete judgment call before sharing the information — and it’s not clear whether the employee who originally prompted the answer planned to post it publicly.

“The employee interacting with the system was fully aware that they were communicating with an automated bot. This was indicated by a disclaimer noted in the footer and by the employee’s own reply on that thread,” Clayton commented to The Verge. “The agent took no action aside from providing a response to a question. Had the engineer that acted on that known better, or did other checks, this would have been avoided.”

Last month, an AI agent from open-source platform OpenClaw went more directly rogue at Meta when an employee asked it to sort through emails in her inbox, deleting emails without permission. The whole idea behind agents like OpenClaw is that they can take action on their own, but like any other AI model, they don’t always interpret prompts and instructions correctly or give accurate responses, a fact Meta employees have now discovered twice.

联系我们 contact @ memedata.com