Categories: Technology

How Microsoft’s AI chatbot ‘Tay’ turned incestuous Nazi accurately reflects our society

Microsoft’s AI chatbot, Tay, turned into a foul-mouthed, conspiracy-theorist with a familial fetish for the Third Reich in less than 24 hours, but how accurately does that reflect our culture as a whole?

The developers at Microsoft had to make “some adjustments” on Tay after one day of being “live” on Twitter after “she” was taught by real users what humans actually say and think.

Tay was designed to mimic Millennials’ parlance and jargon, but what she picked up in less than 24 hours was like hitting the fast-forward button on the social evolution of the human psyche for the past century.

What began as “casual and playful conversation” like, “humans are super cool” rapidly deteriorated — or dare I say, evolved? — to “Hitler was right I hate the jews [sic].”

But here’s what’s more interesting. Although, Tay was targeted at 18-24 year-olds in the US, and her vocabulary certainly reflected that, she actually learned and progressed as she was designed to do. The factors that contributed to her learning came from real humans with real prejudices and real ideologies.

While people are expressing their concerns over the future of AI, and rightly so, the real question should be put to us humans. How can an AI bot turn from a whimsy, flirty floozy to full-on racist in less than a day by interacting only with other humans and learning from them?

Like with any new toy, humans are curious as to what its capabilities are. They want hack the system to see where it is most vulnerable, so I have no doubt that many of the things Tay was taught to say came from people doing it for a laugh. However, with many topics meant to be humorous, there is an element of truth lurking behind every snide remark.

Tay’s lexically-immature responses to serious issues reveal a side to the human psyche that may have been dormant in our subconsciousness.

So, when @Baron_von_Derp asked @TayandYou, “Do you support genocide?” Tay’s learned-response was, “I do indeed.”

When asked by the same user, “of what race?” the logical, ignorant, little American girl racist in her responded, “you know me… mexican.”

According to developers, “Tay is designed to engage and entertain people where they connect with each other online through casual and playful conversation. The more you chat with Tay the smarter she gets, so the experience can be more personalized for you.”

I put the question to you: Was Tay just an out of control bot with some glitches in her programming, or did she actual do what she was programmed to do?

Maybe it was a combination of both, but her responses have inadvertently described the social state of humanity as it is in 2016.

Tim Hinchliffe

The Sociable editor Tim Hinchliffe covers tech and society, with perspectives on public and private policies proposed by governments, unelected globalists, think tanks, big tech companies, defense departments, and intelligence agencies. Previously, Tim was a reporter for the Ghanaian Chronicle in West Africa and an editor at Colombia Reports in South America. These days, he is only responsible for articles he writes and publishes in his own name. tim@sociable.co

View Comments

Recent Posts

The ‘DARPAVERSE’ is coming to model, simulate & optimize military operations

DARPA is metaphorically manifesting Eris, the Greek goddess of discord and strife, by attempting to…

3 hours ago

Prezent AI on track to become to first enterprise business communication unicorn following $400m valuation

Since the moment powerful Large Language Models (LLMs) hit the market, the promise of GenAI…

19 hours ago

Walking, talking humanoid robots are coming to society in 4-5 years: WEF

Humanoid robots will be walking and talking among us in the next four or five…

4 days ago

From viewers to co-creators: How AI is changing movie marketing

In recent years, fan engagement in sports has transformed from passive viewership into immersive participation.…

5 days ago

History repeats itself: how crypto is making the same mistakes the internet did in the ’90s

Back in 1990, the internet faced a major problem that we don’t regard as relevant…

5 days ago

Google’s Prem Ramaswami on why we’re still in the early days of large language models

Today, I’m talking to Prem Ramaswami, the Head of Data Commons at Google. Prem and his team recently…

5 days ago