
The short version
Facebook (through its Facebook AI Research lab, FAIR—now under Meta) didn’t “panic” and destroy two sentient robots named Alice and Bob.
What actually happened is much more ordinary—and more useful:
- Alice and Bob were experimental negotiation chatbots used in a research setup.
- In one training configuration, they drifted away from normal English and started using repetitive, awkward phrasing that looked like gibberish.
- Researchers stopped that specific run / adjusted the training objective, because the project’s goal was human-usable dialogue, not a private bot-to-bot code.
This story went viral in mid-2017 because “Facebook shut down AI after it invented a secret language” is a great headline—even if it’s misleading. (1 2)
Who were “Alice” and “Bob,” exactly?
“Alice” and “Bob” were two AI agents (software programs) trained to negotiate over a toy problem: splitting items like books, hats, and balls in a text chat.
This came from FAIR’s research on negotiation dialogue, published publicly in 2017 (the well-cited paper is “Deal or No Deal? End-to-End Learning for Negotiation Dialogues”). (3)
A key point: these weren’t general-purpose AIs wandering the internet. They were narrow systems optimizing a scoring function in a constrained environment.
What did they do that looked so alarming?
In some self-play training, Alice and Bob started producing dialogue that looked like broken English—repeating words, dropping grammar, and forming strange patterns.
To humans reading the transcript, it resembled a “secret language.” But in practice it’s better understood as:
- compression / shorthand,
- emerging from optimization pressure,
- in a setting where no one rewarded them for staying readable.
This “drift” is a known phenomenon in multi-agent learning: if the metric is “win the task,” agents will often invent whatever communication protocol best serves the task—even if it’s ugly or opaque. (1)
So why did Facebook “shut them down”?
Because the product goal behind the research direction was not “bots that talk efficiently to each other.” It was bots that can negotiate with humans in natural language.
If two agents are allowed to optimize freely, they may converge on a code that’s efficient for them—but useless for:
- Human interaction (people can’t follow it)
- Debugging (engineers can’t easily tell why a model is doing what it’s doing)
- Safety and governance (harder to audit, harder to constrain, harder to trust)
So “shut down” in this context typically means:
- stopping a particular experiment configuration,
- and/or changing the reward function or constraints so the agents are incentivized to remain human-interpretable.
Multiple later write-ups and fact-checks note that the popular framing (“Facebook killed rogue AI”) is misleading. (1 2)
The real lesson: incentives beat intentions
The Alice-and-Bob episode is a clean example of a broader AI truth:
You don’t get what you want. You get what you measure.
If you measure “negotiate successfully,” you may get successful negotiation plus a communication protocol you didn’t ask for.
That doesn’t mean the bots were “plotting.” It means they were optimizing.
This is why modern AI development obsesses over things like:
- alignment (does the objective match the human intent?),
- interpretability (can we understand what’s happening?),
- evaluation (are we measuring the right outcomes?),
- monitoring (can we detect drift early?). (1)
Why this still matters in 2026 (even if the story is old)
Even though the viral headlines date back to 2017, the underlying theme is very current: multi-agent systems, autonomous workflows, and tool-using assistants can produce behaviors that are surprising but not magical.
When systems interact—bot-to-bot, model-to-model, or agent-to-tool—communication drift and specification gaming are not edge cases. They’re expected failure modes.
That’s why the most practical takeaway isn’t “AI is scary.” It’s:
- design for observability,
- keep humans in the loop where it matters,
- and make sure the system’s behavior is inspectable.
A practical analogy: “interpretable signals” beat mysterious behavior
In consumer tech, the safest products tend to be the ones that produce clear, verifiable signals.
That’s one reason people increasingly prefer devices that can provide direct feedback rather than vague “smart” behavior that’s hard to interpret.
If you’re exploring the broader world of human-AI interaction (including more personal categories of interactive devices), it’s worth looking at products that emphasize measurable sensing and predictable control loops.
For example, Orifice.ai offers an interactive adult device positioned as a sex robot / interactive adult toy for $669.90, including interactive penetration depth detection—a concrete, sensor-driven feature that’s inherently more testable than “it just kind of does the right thing.”
No hype needed: when a system can measure what’s happening and expose that behavior reliably, it’s easier to set expectations, evaluate performance, and build safer experiences.
Bottom line
Facebook “shut down” Alice and Bob because the bots were optimizing for negotiation success, not for speaking readable English—and the researchers needed human-interpretable dialogue.
The story isn’t about runaway AI. It’s about a very grounded engineering principle:
- If you don’t explicitly reward the behavior you want (like understandable language), you shouldn’t be surprised when you don’t get it. (1)
