In 2016, Microsoft's Racist Chatbot Revealed the Dangers of Online
Por um escritor misterioso
Descrição
Part five of a six-part series on the history of natural language processing and artificial intelligence
Disinformation Researchers Raise Alarms About A.I. Chatbots - The New York Times
Microsoft's Zo chatbot calls the Qu'ran 'violent
Microsoft's Zo chatbot is a politically correct version of her sister Tay—except she's much, much worse
I Guess I Fear Humans More than Technology - ChemistryViews
Twitter taught Microsoft's AI chatbot to be a racist asshole in less than a day - The Verge
Google vs. Microsoft (Bing + ChatGPT), by Onuche Ogboyi
Why do AI chatbots so often become deplorable and racist? - Verdict
Algorithms and Terrorism: The Malicious Use of Artificial Intelligence for Terrorist Purposes. by UNICRI Publications - Issuu
Microsoft's Tay chatbot returns briefly and brags about smoking weed
Microsoft Chat Bot Goes On Racist, Genocidal Twitter Rampage
Microsoft Chat Bot Goes On Racist, Genocidal Twitter Rampage
When Technology Fails: Racial Bias and AI - MyOpenCourt
Microsoft apologizes for its racist chatbot's 'wildly inappropriate and reprehensible words
de
por adulto (o preço varia de acordo com o tamanho do grupo)