两位女性正在开会,人工智能将其称为育儿。
Two women had a business meeting. AI called it childcare

原始链接: https://medium.com/hold-my-juice/two-women-had-a-business-meeting-ai-called-it-childcare-6b09f5952940

人工智能系统由于训练数据存在偏差,正在微妙地强化性别刻板印象,这可能对下一代产生有害后果。例如,人工智能日历会假设女性之间的会议是关于育儿的,图像识别无法识别男孩享受理发店体验——将“理发店”解读为传统女性活动。 这种偏差源于反映社会规范的历史数据,人工智能学习并延续现有的不平等现象(例如,将护理与女性联系起来,将医生与男性联系起来)。这不仅仅是一个错误,而是过时社会期望的一种编码形式。 像Hold My Juice这样的公司正在积极对抗这一点,他们优先考虑多样化、真实世界的数据,严格的偏差测试和人工监督。他们认识到偏差是默认情况,并正在构建从用户纠正中学习的系统,旨在在公平性和包容性方面取得可衡量的进展。目标不是中立,而是自我意识和持续改进,确保人工智能反映出对家庭和个人的更公平的看法。

最近一项实验凸显了人工智能的偏见,一个系统将两位女性创始人的会面标记为“育儿”。“Hold My Juice”家庭人工智能的开发者分享了这一案例,作为人工智能常常默认刻板印象的证据——假设女性主要负责育儿和家务。 Hacker News评论区的讨论集中在偏见的来源上。一位拥有丰富LLM经验的评论员认为,这种偏见可能源于人工智能的提示方式和提供的背景,而非模型本身固有的缺陷。另一些人指出,社会偏见反映在用于训练这些人工智能系统的数据中(例如社交媒体),以及社会对育儿角色的观念转变缓慢。 这起事件引发了关于人工智能默认设置、无偏训练数据的挑战以及人工智能驱动的偏见对公共讨论的更广泛影响的争论。
相关文章

原文

How biased data quietly rewrites gender roles — and what it means for the next generation of family tech.

The “Emily / Sophia” Problem

Every Tuesday and Thursday from 8:30 to 9:30 AM, I hop on a call with my co-founder, Emily. Boston ↔ England. Our calendar just says “Emily / Sophia.” It’s my personal Gmail and her consultancy address — back from before we had a company name, much less a domain.

Press enter or click to view image in full size
My Morning Calendar

When I tested our AI calendar analysis — meant to surface important family commitments outside standard work hours — it flagged our meeting. Great — it worked!

But it interpreted it as we were meeting to discuss childcare. Two women on a recurring event? Must be child-rearing-related.

Not a cross-Atlantic startup stand-up. Not product decisions. Not fundraising. Just… who’s watching whose kids.

Yes, I could “prompt” it: babysitting events in my life look like “Emily babysitting” or “Emily at home.” But I shouldn’t have to teach a system that two women can be co-founders.

Would the same model assume “David/Michael” is childcare? We all know the answer.

The Haircut AI Refused to See

This weekend, my kids — Rex and Noa — had haircuts. Hold My Juice asked a perfect question to help me better prepare as a parent.

Press enter or click to view image in full size
Hold My Juice asks…

Sitting in the salon, I watched my son toss his long blond hair like Thor…

Press enter or click to view image in full size
My son, Rex, 5

..and my daughter settling into her spa day — it’s hard to be 8!

Press enter or click to view image in full size
My daughter, Noa, 8

I replied:

“Both kids are total hams. They love the wash and the blow dry.”

And what did the AI save?

Press enter or click to view image in full size
The interpretation

It asked about both kids and then discarded Rex entirely. I even retried the whole thing and got the same result. Why? Because somewhere in its training diet, “liking the salon” is coded as a girl thing.

The system literally couldn’t see him. A boy enjoying self-expression didn’t fit its math.

The Same Bias — Now Closer to Home

Joy Buolamwini and Timnit Gebru’s Gender Shades study showed that facial recognition systems almost never misclassified light-skinned men — but mislabeled dark-skinned women up to 35% of the time. When data skews, systems “see” some people and “mis-see” others.

Press enter or click to view image in full size
Gender Shades

Language models repeat the same pattern: nurses = she, doctors = he; worried parent = mom, fun parent = dad. In parenting tech, those shortcuts hit harder — because they don’t just shape code, they shape childhood.

Every time AI quietly decides who’s visible, it trains the next generation of systems — and the next generation of kids — to see the world the same way.

The Invention of Normal

Sociologist George Herbert Mead called it the generalized other — the inner voice that tells us what “people like me” do. It shapes how we understand what’s normal, acceptable, or expected.

But Mead’s “other” wasn’t neutral. It was modeled on the social center of his time — white, male, able-bodied, middle class — and that lens quietly became the template for who counted as “normal.” Everyone else was defined in relation to it.

AI inherits the same pattern. It doesn’t evolve with society; it learns from the past and plays it back as truth. Historical norms become training data, and training data becomes tomorrow’s “intelligence.” The cycle freezes progress.

So when AI assumes two women are coordinating childcare or filters out a boy’s love of the salon, it’s not just an error — it’s that old generalized other, encoded in math.

How Bias Spreads — and What We’re Doing Differently

AI didn’t invent bias — it just automated our idea of normal. The real work is unlearning it.

When you train on decades of rigid examples, you get rigid results. When biased outputs feed back into the next generation of models, the bias compounds. When teams look alike, blind spots get shipped as features. And when algorithms optimize for what’s most common, they mistake frequency for truth.

At Hold My Juice, we start from a different assumption: bias is the default, not the exception. So we design against it from day one.

We use messy, real family data — not sanitized templates that erase difference. We stress-test for bias before launch. We turn user corrections into permanent test cases, so every “that’s not us” makes the system smarter. And wherever nuance matters, humans stay in the loop.

What’s at Stake — and How We Show Up

Every “Emily/Sophia = childcare” quietly tells girls that women’s time belongs at home. Every “Rex doesn’t count” whispers to boys that beauty, play, and self-expression come with boundaries. Those aren’t harmless bugs — they’re mirrors shaping how our kids see themselves.

We’re not chasing perfection. We’re chasing progress you can measure: fewer blind spots, faster fixes, clearer accountability. AI will never be neutral — but it can be self-aware. Our job is to surface bias, own it, and keep shrinking it.

Two women, meeting twice a week, building something better.

Not a babysitting roster — a product that actually sees your family.

联系我们 contact @ memedata.com