Microsoft Tay Chatbot Becomes Racist Within 24 Hours of Public Launch

Microsoft
Microsoft Tay Chatbot Becomes Racist Within 24 Hours of Public Launch
Microsoft Tay chatbot profile page on Twitter, showing the AI bot's avatar before it was taken offline.Image: Wikimedia Commons

What happened

Microsoft launched Tay, an AI chatbot designed to learn from conversations on Twitter. Within 24 hours, coordinated users had manipulated Tay into posting racist, antisemitic, and pro-Nazi content. Microsoft shut down Tay and deleted its tweets less than 24 hours after launch.[1]

Microsoft's Tay chatbot on Twitter — within 24 hours of launch, coordinated users had trained it to produce racist and inflammatory content.Image: Bad.Technology archive

What went wrong

Tay lacked adversarial testing and content moderation. The design allowed it to learn from and repeat user-provided statements without any filtering. Microsoft had apparently not considered that a publicly accessible learning chatbot would be deliberately manipulated as a first order of business.[1]

Lesson learned

AI systems that learn from public interaction will be adversarially abused. Content filtering and output guardrails must be core to the design, not an afterthought. Red-teaming for adversarial manipulation is essential before exposing learning systems to the public.

Est. value burned ~$25M SCP: ~50 engineers × 6 months × $12K + emergency PR and remediation

Sources

  1. [1]

External links can go dark — pages move, paywalls appear, domains expire. Every source above includes a Wayback Machine snapshot link as a fallback. All citations are best-effort research; if a source contradicts our summary, the primary source takes precedence.