Tuesday, 11 September 2018

Google’s AI to detect toxic comments can be easily fooled with ‘love’


A group of researchers has found that simple changes in sentences and its structure can fool Google’s perspective AI, made for detecting toxic comments and hate speech. These methods involve inserting typos, spaces between words or add innocuous words to the original sentence. The AI project, which was started in 2016 by a Google offshoot called Jigsaw, assigns a toxicity score to a piece of text. Google defines a toxic comment as a rude, disrespectful, or unreasonable comment that is likely to make you leave a discussion. The researchers suggest that even a slight change in the sentence can change the toxicity score dramatically. They…

This story continues at The Next Web

Or just read more coverage about: Google
https://ift.tt/2N3tSx0 Ivan Mehta September 11, 2018 at 06:00AM

No comments: