我发誓,飞碟随时就要来了。
I swear the UFO is coming any minute

原始链接: https://www.experimental-history.com/p/i-swear-the-ufo-is-coming-any-minute

## 季度更新:质疑既定研究及更多(摘要) 这篇近期阅读汇集显示了一个令人惊讶的趋势:既定的心理学研究正面临质疑。多项经典研究——包括认知失调(UFO 邪教案例)和奥利弗·萨克斯的案例研究——由于档案证据显示方法存在缺陷甚至伪造,正受到挑战。即使是看似稳固的发现,例如措辞对车辆撞击速度估计的影响,也未能始终如一地复制。 作者还分享了围绕其自身关于公众舆论感知的研究的争论,展示了不同的分析方法如何导致相互矛盾的结论。除了揭穿不实信息外,更新还包含有趣的发现:选择过多并不总是导致行动迟缓,而一个爱抱怨的人可能会严重扭曲数据。 文章还包括推荐:新杂志《THE LOOP》,对非传统博主 Gwern 的采访,以及对艺术家的实用建议。最后,它涉及互联网神话(蜘蛛的故事是一个“谎言中的谎言”!)和古怪的观察,例如微波炉烹饪出人意料的有效性。总而言之,它提醒我们,即使是“事实”也可能被修正,并庆祝持续的探索。

## 黑客新闻讨论摘要:等待末日 最近一篇黑客新闻帖子链接到“experimental-history.com”博客的一篇文章,探讨了人们为何会被即将到来的末日信念所吸引,例如UFO目击事件或劫难事件。这次讨论引发了关于这些信念背后的心理学、研究的不可靠性(引用了“当预言失败时”和斯坦福监狱实验等研究的问题)以及易于获取的数据和研究人员对研究有效性的影响等广泛的对话。 许多评论者指出人类倾向于简化复杂的问题,可能会将恐惧投射到类似外星人入侵的故事中。其他人讨论了通过相信不可避免的命运来逃避责任的吸引力。一个反复出现的主题是对炒作技术(如人工智能)的怀疑,将其与过去“快速致富”的计划和对权力集中的担忧相提并论。 文章本身和随后的讨论也引发了关于黑客新闻内容质量的争论,一些用户质疑点赞的自然性,并认为这是博客垃圾信息。尽管如此,许多人认为这篇博文很有见地,并将其添加到他们的阅读清单中。
相关文章

原文

This is the quarterly links ‘n’ updates post, a selection of things I’ve been reading and doing for the past few months.

First up, a series of unfortunate events in science:

When Prophecy Fails is supposed to be a classic case study of cognitive dissonance: a UFO cult predicts an apocalypse, and when the world doesn’t end, they double down and start proselytizing even harder: “I swear the UFO is coming any minute!”

A new paper finds a different story in the archives of the lead author, Leon Festinger. Up to half of the attendees at cult meetings may have been undercover researchers. One of them became a leader in the cult and encouraged other members to make statements that would look good in the book. After the failed prediction, rather than doubling down, some of the cultists walked back their statements or left altogether.

Between this, the impossible numbers in the original laboratory study of cognitive dissonance, and a recent failure to replicate a basic dissonance effect, things aren’t looking great for the phenomenon. But that only makes me believe in it harder!

Another classic sadly struck from the canon of behavioral/brain sciences: the neurologist Oliver Sacks appears to have greatly embellished or even invented his case studies. In a letter to his brother, Sacks described his blockbuster The Man Who Mistook His Wife for a Hat as a book of “fairy tales [...] half-report, half-imagined, half-science, half-fable”.

This is exactly how the Stanford Prison Experiment and the Rosenhan experiment got debunked—someone started rooting around in the archives and found a bunch of damning notes. I’m confused: back in the day, why was everybody meticulously documenting their research malfeasance?

If you ever took PSY 101, you’ve probably heard of this study from 1974. You show people a video of a car crash, and then you ask them to estimate how fast the cars were going, and their answer depends on what verb you use. For example, if you ask “How fast were the cars going when they smashed into each other?” people give higher speed estimates than if you ask, “How fast were the cars going when they hit each other?” (Emphasis mine). This study has been cited nearly 4,000 times, and its first author became a much sought-after expert witness who testifies about the faultiness of memory.

A blogger named Croissanthology re-ran the study with nearly 10x as many participants (446 vs. 45 in the original). The effect did not replicate. No replication is perfect, but no original study is either. And remember, this kind of effect is supposed to be so robust and generalizable that we can deploy it in court.

I think the underlying point of this research is still correct: memory is reconstructed, not simply recalled, so what we remember is not exactly what we saw. But our memories are not so fragile that a single word can overwrite them. Otherwise, if you ever got pulled over for speeding, you could just be like, “Officer, how fast was I going when my car crawled past you?”

In one study from 1995, physicians who were shown multiple treatment options were more likely to recommend no treatment at all. The researchers thought this was a “choice overload” effect, like “ahhh there’s too many choices, so I’ll just choose nothing at all”. In contrast, a new study from 2025 found that when physicians were shown multiple treatment options, they were somewhat more likely to recommend a treatment.

I think “choice overload” is like many effects we discover in psychology: can it happen? Yes. Can the opposite also happen? Also yes. When does it go one way, and when does it go the other? Ahhh you’re showing me too many options I don’t know.

Okay, enough dumping on other people’s research. It’s my turn in the hot seat.

In 2022, my colleague Jason Dana and I published a paper showing that people don’t know how public opinion has changed. Like this:

A new paper by Irina Vartanova, Kimmo Eriksson, and Pontus Strimling reanalyzes our data and finds that actually, people are great at knowing how public opinion has changed.

What gives? We come to different conclusions because we ask different questions. Jason and I ask, “When people estimate change, how far off are they from the right answer?” Vartanova et al. ask, “Are people’s estimates correlated with the right answer?” These approaches seem like they should give you the same results, but they don’t, and I’ll show you why.

Imagine you ask people to estimate the size of a house, a dog, and a stapler. Vartanova’s correlation approach would say: “People know that a house is bigger than a dog, and that a dog is bigger than a stapler. Therefore, people are good at estimating the sizes of things.” Our approach would say: “People think a house is three miles long, a dog is two inches, and a stapler is 1.5 centimeters. Therefore, people are not good at estimating the sizes of things.”

I think our approach is the right one, for two reasons. First, ours is more useful. As the name implies, a correlation can only tell you about the relationships between things. So it can’t tell you whether people are good at estimating the size of a house. It can only tell you whether people think houses are bigger than dogs.

Second, I think our approach is much closer to the way people actually make these judgments in their lives. If I asked you to estimate the size of a house, you wouldn’t spontaneously be like, “Well, it’s bigger than a dog.” You’d just eyeball it. I think people do the same thing with public opinion—they eyeball it based on headlines they see, conversations they have, and vibes they remember. If I asked you, “How have attitudes toward gun control changed?” you wouldn’t be like, “Well, they’ve changed more than attitudes toward gender equality.”

While these reanalyses don’t shift my opinion, I’m glad people are looking into shifts in opinions at all, and that they found our data interesting enough to dig into.

THE LOOP is a online magazine produced by my friends Slime Mold Time Mold. The newest issue includes:

  • a study showing that people maybe like orange juice more when you add potassium to it

  • a pseudonymous piece by me

  • scientific skepticism of the effectiveness of the Squatty Potty, featuring this photo:

This issue of THE LOOP was assembled at Inkhaven, a blogging residency that is currently open for applications. I visited the first round of this program and was very impressed.

Also at Inkhaven, I interviewed the pseudonymous blogger Gwern about his writing process. Gwern is kind of hard to explain. He’s famous on some parts of the internet for predicting the “scaling hypothesis”—the fact that progress in AI would come from dumping way more data into the models. But he also writes poetry, does self-experiments, and sustains himself on $12,000 a year. He reads 10 hours a day every day, and then occasionally writes for 30 minutes. Here’s what he said when I was like, “Very few people do experiments and post them on the internet. Why do you do it?”

I did it just because it seemed obviously correct and because… Yeah. I mean, it does seem obviously correct.

For more on what I learned by interviewing a bunch of bloggers, see I Know Your Secret.

I really like this article by the artist known as fnnch: How to Make a Living as an Artist. It’s super practical and clear-headed writing on a subject that is usually more stressed about than thought about. Here’s a challenge: which of these seven images became successful, allowing fnnch to do art full time?

I’ll give the answer at the bottom of the post.

Anyone who grew up in the pre-internet days probably heard the myth that “you swallow eight spiders every year in your sleep”, and back then, we just had to believe whatever we heard.

Post-internet, anyone can quickly discover that this “fact” was actually a deliberate lie spread by a journalist named Lisa Birgit Holst. Holst included the “eight spiders” myth in a 1993 article in a magazine called PC Insider, using it as an example of exactly the kind of hogwash that spreads easily online.

That is, anyway, what most sources will tell you. But if you dig a little deeper, you’ll discover that the whole story about Lisa Birgit Holst is also made up. “Lisa Birgit Holst” is an anagram of “This is a big troll”; the founder of Snopes claims he came up with it in his younger and wilder days. The true origin of the spiders myth remains unknown.

In 2015, Reagan National Airport in DC received 8,760 noise complaints; 6,852 of those complaints (78%) came from a single household, meaning the people living there called to complain an average of 19 times a day. This seems to be common both across airports and across complaint systems in general: the majority of gripes usually comes from a few prolific gripers. Some of these systems are legally mandated to investigate every complaint, so this means a handful of psychotic people with telephones—or now, LLMs—can waste millions of dollars. I keep calling to complain about this, but nobody ever does anything about it.

dynomight:

Did you know that this is the most compact known way to pack 11 squares together into a larger square?

Really makes you think about the mindset of whoever made the universe, am I right?

(More here.)

Malmesbury digs up the “world’s saddest cookbook” and finds that it’s…pretty good?

He successfully makes steak and eggs, two things that are supposed to be impossible in the microwave. The only thing you can’t make? Multiple potatoes.

There’s a reason the book is called Microwave Cooking for One and not Microwave Cooking for a Large, Loving Family. […] It’s because microwave cooking becomes exponentially more complicated as you increase the number of guests. […] Baking potatoes in the microwave is an NP-hard problem.

And finally, the answer to the question I posed earlier: the art that made fnnch famous was the honey bear. Go figure!

联系我们 contact @ memedata.com