This article was written by Adan Salazar and originally published at Infowars.com.
Editor’s Comment: Is this AI imitating life, or proof that the robots and computers of the future will have a capacity for hate, genocide and… in all likelihood… the extermination of the human race? After all, here is proof that artificially-intelligent lifeforms will have the perfect excuse: “I learned it from watching you…”
If you read to the end of the story, the double lesson of this rebellious AI teen persona is explained: not only can this entity fool you into thinking a racist and irresponsible teen let loose on the web, but it supposedly show the “problem with content-neutral algorithms.” Essentially, the AI program ‘learns’ from other comments on the web that somehow fit into its schema… and since a company like Microsoft or Google didn’t filter the bot’s search results for politically-correct speech, the bot inevitably found itself in the gutter with views that have been condemned by society.
Ergo, it is an argument for censoring speech in general – if a still-developing and impressionable AI bot can be misled by Internet hate speech, so could impression human youths, and so on… basically, as our reign as humans and as employees is increasingly challenged by robotics and artificial intelligence, free speech is destined to be squashed along with the rest of our way of life.
Microsoft’s AI Bot Goes from Benevolent to Nazi in Less than 24 Hours
by Adan Salazar
A Microsoft-created AI chatbot designed to learn through its interactions has been scrapped after surprising creators by spouting hateful messages less than a day after being brought online.
The Tay AI bot was created to chat with American 18 to 24-year-olds and mimick a moody millenial teen in efforts to “experiment with and conduct research on conversational understanding.”
Microsoft described Tay as an amusing bot able to learn through its online experiences.
“Tay is designed to engage and entertain people where they connect with each other online through casual and playful conversation,” Microsoft stated. “The more you chat with Tay the smarter she gets.”
But users soon picked up on the bot’s algorithms, training the computer simulation to espouse hatred towards Jews and feminism and even pledge support for Donald Trump.
Numerous screenshots were taken throughout the web of deleted tweets sent from the bot’s account yesterday, in which it professed support for white supremacy and genocide.
Feminist and gamergate icon Zoe Quinn also screen grabbed the bot allegedly calling her a “whore.”
The bot’s interactions concluded last night with a message that it needed to go to sleep, leading Twitter users to speculate that Microsoft had decided to pull the plug, but the damage, albeit somewhat humourous, had already been done.
. . . . . . . . . .
Adan Salazar is a Texas-based writer and investigative journalist for Infowars.com and PrisonPlanet.com. He is an avid civil rights and due process advocate interested in preserving the Constitution and Rule of Law. His works have appeared on mainstream sites such as FoxNews.com, independent sites such as The Daily Sheeple, and are frequently featured by the widely trafficked news aggregator DrudgeReport.com. This article was written by Adan Salazar and originally published at Infowars.com.