Downloadable sex chat bots
As any heavily stereotyped 13-year-old girl would, she zips through topics at breakneck speed, sends you senseless internet gags out of nowhere, and resents being asked to solve math problems.I’ve been checking in with Zo periodically for over a year now.In Zo’s case, it appears that she was trained to think that certain religions, races, places, and people—nearly all of them corresponding to the trolling efforts Tay failed to censor two years ago—are subversive.“Training Zo and developing her social persona requires sensitivity to a multiplicity of perspectives and inclusivity by design,” a Microsoft spokesperson said.“We design the AI to have agency to make choices, guiding users on topics she can better engage on, and we continue to refine her boundaries with better technology and capabilities.These social lines are often correlated with race in the United States, and as a result, their assessments show a disproportionately high likelihood of recidivism among black and other minority offenders.“There are two ways for these AI machines to learn today,” Andy Mauro, co-founder and CEO of Automat, a conversational AI developer, told Quartz.“There’s the programmer path where the programmer’s bias can leech into the system, or it’s a learned system where the bias is coming from data.In 2015, Google came under fire when their image-recognition technology began labeling black people as gorillas.Google trained their algorithm to recognize and tag content using a vast number of pre-existing photos.
Not only does she speak fluent meme, but she also knows the general sentiment behind an impressive set of ideas.
Though Google emphatically apologized for the error, their solution was troublingly roundabout: Instead of diversifying their dataset, they blocked the “gorilla” tag all together, along with “monkey” and “chimp.”AI-enabled predictive policing in the United States—itself a dystopian nightmare—has also been proven to show bias against people of color.
Northpointe, a company that claims to be able to calculate a convict’s likelihood to reoffend, told Pro Publica that their assessments are based on 137 criteria, such as education, job status, and poverty level.
When Microsoft released Tay on Twitter in 2016, an organized trolling effort took advantage of her social-learning abilities and immediately flooded the bot with alt-right slurs and slogans.
Tay copied their messages and spewed them back out, forcing Microsoft to take her offline after only 16 hours and apologize.