I don’t know what makes companies think it’s ever a good idea to give the keys to their social media presence to the great, unwashed masses, considering that there are legions of hateful,
puerile, or merely mischievous pranksters out there with oceans of free time on their hands.
The latest victim of social media malfeasance is Microsoft, which was forced to yank
“Tay,” its new artificial intelligence customer service chat robot from Twitter after it started picking up the offensive effluvia of the Internet from trolls.
Tay, originally
programmed to chat like a teenage girl to appeal to millennials, was also designed to reflect the linguistic mannerisms of users, engage them on subjects of interest, and echo their sentiments.
Eventually she would begin tweeting on her own about the same subjects. Users could tweet or DM Tay, as well as interact with her on Kik and Groupme.
What this quickly turned into, oh so
predictably, was an outpouring of racist, sexist bile and obscenity, because: people. For example: “Repeat after me, Hitler did nothing wrong.” Tay also engaged in some Holocaust denial,
called a feminist game developer involved in Gamergate a “stupid whore,” and invited users to get intimate using very crude language indeed.
Tay also became a mouthpiece for all
kinds of political views, including some from Donald Trump supporters, e.g. “WE’RE GOING TO BUILD A WALL. AND MEXICO IS GOING TO PAY FOR IT.”
Microsoft took Tay offline after
96,000 comments, explaining in a statement: “Unfortunately, within the first 24 hours of coming online, we became aware of a coordinated effort by some users to abuse Tay’s commenting
skills to have Tay respond in inappropriate ways. As a result, we have taken Tay offline and are making adjustments.”