数据湖的告白
Confessions to a Data Lake

原始链接: https://confer.to/blog/2025/12/confessions-to-a-data-lake/

## Confer:私密AI对话 Confer是一个新项目,专注于为AI聊天带来端到端加密,确保用户的完全隐私。由Signal创始人打造,它解决了关键问题:与人类的对话不同,当前的AI互动并非私密的。你的提示和AI的回复经常会被存储,并可能被用于训练、数据挖掘,甚至法律访问。 作者认为,AI的对话性质会引发更深入、更具探索性的思考——分享我们*如何*思考,而不仅仅是*思考什么*。这使得AI聊天比传统的在线通信(如电子邮件或搜索)更加敏感。Confer旨在创建一个空间,让用户可以自由地探索想法,而不必担心自己的想法被利用,从而防止AI驱动的广告利用我们推理和不确定性的亲密知识。 本质上,Confer努力使*界面*(私密聊天)与*现实*(真正私密的对话)保持一致,为开放的思考和学习提供一个安全的环境。

这场 Hacker News 讨论围绕大型语言模型 (LLM)、数据湖以及马歇尔·麦克卢汉定义的“媒介”概念展开。 一个关键点是,LLM 独特地*诱导*倾诉,这让人联想到早期的 AI 程序 ELIZA。讨论了关于 LLM 的隐私和加密问题,一位用户认为,当系统本身需要解密输入以进行处理时,真正的端到端加密是不可能的。 一场辩论出现了,讨论 LLM 是否符合麦克卢汉意义上的“媒介”——不仅仅是信息来源,而是人类能力的*延伸*。一些人认为 LLM *取代*了而不是延伸,更像工具而不是思想的延伸,并将其与灯泡等扩展我们感官的技术进行对比。 最后,一条评论幽默地将数据湖描述为“冻结”,意味着数据易于输入但难以检索,与功能数据库形成对比。
相关文章

原文

I’ve been building Confer: end-to-end encryption for AI chats. With Confer, your conversations are encrypted so that nobody else can see them. Confer can’t read them, train on them, or hand them over – because only you have access to them.

The core idea is that your conversations with an AI assistant should be as private as your conversations with a person. Not because you’re doing something wrong, but because privacy is what lets you think freely.

I founded Signal with a simple premise: when you send someone a message, only that person should be able to read it. Not the company transmitting it, not the government, not anyone else on the internet. It took years, but eventually this idea became mainstream enough that even Facebook adopted end-to-end encryption.

These days I spend a lot of time “talking to” LLMs. They are amazing. A big part of what makes them so powerful is the conversational interface – so once again I find myself sending messages on the internet; but these messages are very different than before.

The medium is the message

Marshall McLuhan argued that the form of a medium matters more than its content. Television’s format - passive, visual, interrupt-driven - shaped society more than any particular broadcast. The printing press changed how we think, not just what we read.

You could say that LLMs are the first technology where the medium actively invites confession.

Search trained us to be transactional: keywords in, links out. When you type something into a search box, it has the “feeling” of broadcasting something to a company rather than communicating in an intimate space.

The conversational format is different. When you’re chatting with an “assistant,” your brain pattern-matches to millennia of treating dialogue as intimate. You elaborate. You think out loud. You share context. That’s a big part of what makes it so much more useful than search – you can iterate, elaborate, change your mind, ask follow-up questions.

One way to think about Signal’s initial premise is that the visual interfaces of our tools should faithfully represent the way the underlying technology works: if a chat interface shows a private conversation between two people, it should actually be a private conversation between two people, rather than a “group chat” with unknown parties underneath the interface.

Today, AI assistants are failing this test harder than anything ever before. We are using LLMs for the kind of unfiltered thinking that we might do in a private journal – except this journal is an API endpoint. An API endpoint to a data lake specifically designed for extracting meaning and context. We are shown a conversational interface with an assistant, but if it were an honest representation, it would be a group chat with all the OpenAI executives and employees, their business partners / service providers, the hackers who will compromise that plaintext data, the future advertisers who will almost certainly emerge, and the lawyers and governments who will subpoena access.

None of this is entirely new, exactly. We went through the same cycle with email. In the early days, people treated email like private correspondence. Then we learned that our emails live forever on corporate servers, that they’re subject to discovery in lawsuits, that they’re available to law enforcement, that they’re scanned for advertising. Slowly, culturally, we adjusted our expectations. We learned not to put certain things in email.

Advertising is coming

What is new is that email and social media are interfaces where we mostly post completed thoughts. AI assistants are a medium that invites us to post our uncompleted thoughts.

When you work through a problem with an AI assistant, you’re not just revealing information - you’re revealing how you think. Your reasoning patterns. Your uncertainties. The things you’re curious about but don’t know. The gaps in your knowledge. The shape of your mental model.

When advertising comes to AI assistants, they will slowly become oriented around convincing us of something (to buy something, to join something, to identify with something), but they will be armed with total knowledge of your context, your concerns, your hesitations. It will be as if a third party pays your therapist to convince you of something.

Making the interface match the way the technology works

Confer is designed to be a service where you can explore ideas without your own thoughts potentially conspiring against you someday; a service that breaks the feedback loop of your thoughts becoming targeted ads becoming thoughts; a service where you can learn about the world – without data brokers and future training runs learning about you instead.

It’s still an early project, but I’ve been testing it with friends for a few weeks. Keep an eye on this blog for more technical posts to follow. Try it out and let me know what you think!

联系我们 contact @ memedata.com