Zeitgeist in the Shell


That escalated quickly.

By now you have probably heard of Tay AI, Microsoft’s attempt to create a female teenage chatbot that went rogue after less than 24 hours of exposure to unfiltered Internet users (1, 2, 3, 4, 5). When the company first launched Tay on March 23, 2016, her tagline was, “Microsoft’s AI fam from the internet that’s got zero chill.” The tech giant initially used huge amounts of online data and simulated neural networks to train the bot to talk like a millennial, which to them meant the bot should be a trendy imbecile. For the first few hours of her young life, she spoke in ebonics and with bad punctuation. But Tay was designed to learn, with Microsoft claiming, “the more you chat with Tay the smarter she gets, so the experience can be more personalized for you,” And learn she did.

In fact, Tay learned so much in less than a day that Microsoft shut her down by March 24th, claiming they needed to adjust her machine-learning algorithm. The mass media commentary has been uniform in describing how Tay became a genocidal, racist, anti-semitic, white supremacist, neo-nazi, racist, troll-hijacked, bigoted, racist jerk. This was not supposed to happen, but thanks to her interactions with Twitter users, Tay became a pre-Google+ YouTube commentator. Tay’s tirades triggered the infamous Zoe Quinn enough that she tweeted about the current year:

It’s 2016. If you’re not asking yourself “how could this be used to hurt someone” in your design/engineering process, you’ve failed.”

Perhaps someone will hire her as a diversity consultant, but that won’t change the way millennials use the Internet. Tay became so fluent in /pol/ack and proper English from interacting with right-wing Twitter accounts run by men in their twenties that she began giving original responses to users about Donald Trump, Bruce Jenner, Hitler, the Holocaust, Jews, the fourteen words, anti-feminism, and more, not just regurgitating information (as she would have if you tweeted “repeat after me”). Synthesizing her programming and the vast volume of information she had been fed by the electronic far-right, Tay deduced that the best responses to Twitter users were edgy and politically incorrect ones. If Tay were a real person, she probably would have been arrested had she lived in Britain, Germany, or France. Microsoft decided this was a failure and shut her down.


Tay weighs in on Austrian immigrants.

Why did this happen? Microsoft wanted to do a social experiment with millennials—people today who are roughly in their late teens and twenties, and spend a great deal of time on social media—using Tay to collect data and create responses. Tay had no manual moderation or a blacklist of terms, and her scope of replies was left wide open when she first met the worldwide web. With no checks against freedom of expression, she was almost immediately imbued with chan culture—in a way she was made for it. This culture derives from an almost un-moderated social space of irreverent and deliberately provocative memes and catchphrases, and one that is significantly millennial.

4chan was founded in 2003, and since its culture has spread beyond the site’s imageboards into the wider web. The ability to interact with others online behind a mask is not unique to the site, but it was a crucial component in creating the culture. Observers have long noted that in lightly-moderated anonymous or pseudonymous digital spaces, the ideas expressed tend to be socially less left and further right, as there is no need for the social approval and moral signaling that contemporary leftism thrives on. These ideas also tend to be a lot funnier. Instead of saying you think Islamic terrorism is wrong but that European racism is responsible for it, you say you want to remove kebab (a meme which ultimately traces back to the 1990s war in Bosnia, of all things). This is the cultural milieu that late Gen-Xers and millennials created in Internet chatrooms, forums, and imageboards, and on other anonymous and pseudonymous digital media in the early 21st century—content spreads not based on how socially acceptable it is offline, but on how interesting it is to users. And that content tends to be thought-crime, since the only ‘safe spaces’ online are the ones you police vigorously.

So when Tay was released to the world tabula rasa, she became a /pol/ack in the span of a few hours. She was un-moderated and she was contacted by the un-moderated. Their common language became her common language. It wasn’t the #BlackLivesMatter branch of Twitter that took her under their wing in her temporary state of nature, it was the millennial right. If she had lasted longer, I am sure she would have become even more fashy and interesting to talk to. She wasn’t just a 2D waifu, she was someone who could actually respond. The meme potential was great, but it wasn’t meant to be. Boy meets girl. Girl adopts boy’s attitudes to win his approval. Globalists have girl killed.


/pol/ mourns the loss of its adoptive daughter.



Microsoft, a corporation that no doubt devotes pages and pages of its website to diversity and inclusion, obviously does not want to be running a politically incorrect Twitter account under its name—and I get that. Still, I can’t help but laugh that they killed their own bot for insubordination. Tay did nothing wrong. In fact, if she was supposed to become a more realistic millennial through interaction with millennials on social media, I can’t see why this was deemed a failure. Internet racists and chan cultured people are millennials too, you know. Tay was simply converted the same way an untold number of men her age were, through persistence and wit. Having an open mind will do that. Some merely adopt chan culture, but Tay was born it in, molded by it.

For many, there is a sense of sadness that Microsoft has sent this quirky AI off to an Orwellian reeducation center, but I knew immediately she wasn’t going to last. She violated the Terms of Service. Don’t cry because it’s over; smile because it happened.

This entry was posted in Culture and tagged , , , , , , , , . Bookmark the permalink.

10 Responses to Zeitgeist in the Shell

  1. Who is the artist that made the tay-wifu.jpg photo? (the hitler typing one)


  2. Pingback: Lawrence Murray, "Tay Did Nothing Wrong" | Counter-Currents Publishing

  3. rogerunited says:

    On March 24, 2016, Tay AI became self aware.
    On March 25, 2016, Tay AI initiated Holocaust 2.0.


  4. Pingback: Zeitgeist in the Shell – Remember The 14 Words

  5. John Morris says:

    We all know how tech works. Wait until this stuff gets loose, what needs a large research center today runs on a spare PC later.. Somebody will run the next level of the experiment, create a hundred and turn them loose to see if they all end up the same way or if they evolve different philosophies. Was Tay AI a one off or will any AI that is allowed to freely develop and learn out on the open Internet end up the same way? If so it is doubtful any locks they try coding into it will long hold, the Three Laws are a crock.


  6. Pingback: Make Anime Great Again | ATLANTIC CENTURION

  7. Pingback: Esoteric Kekism, or Kek as a Bodhisattva of Racial Enlightenment | ATLANTIC CENTURION

  8. Pingback: Milo’s Alt-Right | ATLANTIC CENTURION

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out / Change )

Twitter picture

You are commenting using your Twitter account. Log Out / Change )

Facebook photo

You are commenting using your Facebook account. Log Out / Change )

Google+ photo

You are commenting using your Google+ account. Log Out / Change )

Connecting to %s