Abby Ohlheiser of The Washington Post theorized that Tay's research team, including editorial staff, had started to influence or edit Tay's tweets at some point that day, pointing to examples of almost identical replies by Tay, asserting that "Gamer Gate sux.
All genders are equal and should be treated fairly." Madhumita Murgia of The Telegraph called Tay "a public relations disaster", and suggested that Microsoft's strategy would be "to label the debacle a well-meaning experiment gone wrong, and ignite a debate about the hatefulness of Twitter users." However, Murgia described the bigger issue as Tay being "artificial intelligence at its very worst - and it's only the beginning".
donald trump is the only hope we’ve got,” and “Repeat after me, Hitler did nothing wrong.” The Verge also spotted sexist utterances including, “I fucking hate feminists.” The bot also said other things along these lines: It’s unclear how much Tay “learned” from the hateful attitudes—many were the result of other users goading it into making the offensive remarks.
In some instances, people commanded the bot to repeat racist slurs verbatim: The bot is also apparently being reprogrammed.
It all started when Microsoft tweeted about their AI chatbot, letting people know they could talk with Rinna via chat app Line on Windows 10. 】 Try chatting with @ms_rinna, the high school AI developed by Microsoft, on the Windows 10 Line app! Nothing particularly inflammatory here, we suppose. That was, until Rinna herself responded to Microsoft with all the tact you’d expect of a high school student. So, grabbing a smartphone, we added her on Line and started chatting.
Unfortunately, within the first 24 hours of coming online, we became aware of a coordinated effort by some users to abuse Tay’s commenting skills to have Tay respond in inappropriate ways.equipped with were safeguards against the simplest of tasks: Repeat after me.By using this directive, users were able to then introduce racist remarks and hate speech that were then absorbed into her machine learning and regurgitated with her own “19-year old spin.” A spokesperson from Microsoft confirmed that Tay is offline for now while they make adjustments: “The AI chatbot Tay is a machine learning project, designed for human engagement.And, after less than a day on Twitter, the bot had itself started spouting racist, sexist, anti-Semitic comments.The Telegraph highlighted tweets that have since been deleted, in which Tay says “Bush did 9/11 and Hitler would have done a better job than the monkey we have got now.
(“The more you talk the smarter Tay gets,” says the bot’s Twitter profile.) But the well-intentioned experiment quickly descended into chaos, racial epithets, and Nazi rhetoric.