🐱 Berkeley scientists developing artificial intelligence tool to combat ‘hate speech’ on social media

  • ⚙️ Performance issue identified and being addressed.
  • Want to keep track of this thread?
    Accounts can bookmark posts, watch threads for updates, and jump back to where you stopped reading.
    Create account
CatParty
https://www.thecollegefix.com/berke...e-tool-to-combat-hate-speech-on-social-media/

Scientists at the University of California, Berkeley, are developing a tool that uses artificial intelligence to identify “hate speech” on social media, a program that researchers hope will out-perform human beings in identifying bigoted comments on Twitter, Reddit and other online platforms.

Scientists at Berkeley’s D-Lab “are working in cooperation with the [Anti-Defamation League] on a ‘scalable detection’ system—the Online Hate Index (OHI)—to identify hate speech,” the Cal Alumni Association reports.

In addition to artificial intelligence, the program will use several different techniques to detect offensive speech online, including “machine learning, natural language processing, and good old human brains.” Researchers aim to have “major social media platforms” one day utilizing the technology to detect “hate speech” and eliminate it, and the users who spread it, from their networks.

Current technology mainly involves the use of “keyword searches,” one researcher states, which are “fairly imprecise and blunt.” Current algorithms can be fooled by simply spelling words differently, for instance:

The OHI intends to address these deficiencies. Already, their work has attracted the attention and financial support of the platforms that are most bedeviled—and that draw the most criticism—for hate-laced content: Twitter, Google, Facebook, and Reddit…

D-Lab initially enlisted ten students of diverse backgrounds from around the country to “code” the posts, flagging those that overtly, or subtly, conveyed hate messages. Data obtained from the original group of students were fed into machine learning models, ultimately yielding algorithms that could identify text that met hate speech definitions with 85 percent accuracy, missing or mislabeling offensive words and phrases only 15 percent of the time.

Though the initial ten coders were left to make their own evaluations, they were given survey questions (e.g. “…Is the comment directed at or about any individual or groups based on race or ethnicity?) to help them differentiate hate speech from merely offensive language. In general, “hate comments” were associated with specific groups while “non-hate” language was linked to specific individuals without reference to religion, race, gender, etc. Under these criteria, a screed against the Jewish community would be identified as hate speech while a rant—no matter how foul—against an African-American celebrity might get a pass, as long as his or her race wasn’t cited.

One researcher warned against the possibility of inadvertent censorship: “Unless real restraint is exercised, free speech could be compromised by overzealous and self-appointed censors.” The lab is thus “working to minimize bias with proper training and online protocols that prevent operators from discussing codes or comments with each other.”

Read the report here.
 
I can't wait to see the easy ways the beat this particular AI. This will backfire badly.
Sarcasm has always been the easiest and most reliable way to be a dick right to someone's face and get away with it. AI doesn't stand a chance against it.
 
People will just develop new euphemisms. Either the bot will fall down at the euphemisms or it might be able to recognize the patterns, leaving social media moderators in the awkward position of having to say "you know what you meant!"

Kinda like "it's ok to be white", but more subtle and insidious.
 
Just wait until it starts targeting trannies and the woke crowd unless they specifically program it not to identify anti-straight/cis/white posts as hate speech (which is most likely).

lefty twitter does say some pretty heinous shit so i think there's at least some hope of it being more even-handed in identifying hate speech than human sjws.
 
People will just develop new euphemisms. Either the bot will fall down at the euphemisms or it might be able to recognize the patterns, leaving social media moderators in the awkward position of having to say "you know what you meant!"

Kinda like "it's ok to be white", but more subtle and insidious.

Just start stealing shit right from SJWs so that when they add it to the bot, the bot starts banning SJWs too.
 
While China and Russia build robot tanks and aircraft, America is busy trying to take naughty words off of reddit. The future looks bright, friends.
 
This AI will never succeed. Let's just say that facebook now stands for nigger. Everybody will agree and facebook will have to change it's name unless it wants to be literally nigger.

Facebook Abortion
A super offensive term derived from the collaboration of the two most offensive words in the American language.
God damn Facebook abortion stole my bike.
 
This creates a conundrum , where do we draw the line between blatant, legitimate, hate speech and someone's hurt feelings?

You can't, which is why it's idiotic to bother even trying to police speech.

What's more, is that when you police speech you don't get rid of the idea - you just drive it underground. In the case of something that you want to be illegal/bannable to say/write, if it's something that a large enough number of middle class people feel needs to be addressed (the immigration crisis in Europe, anyone?) it's doing nothing but putting a lid on a boiling pot, which we all know only ever ends one way.

Not to mention that as others have already pointed out, trying to keep up with what euphemisms and slang people use to denigrate others is a fool's errand anyway. The pace at which people could keep up with the bot en masse would probably bring them to the point of tearing their hair out.
 
I agree with what was said before about new euphemisms. I would also add: metaphors.

AI will have a hard time to recognize new creative metaphors, as it can be programmed only to recognize cliches and well-established phrases.

Personally, I would be glad if the language gets more rich and sophisticated as a result. Enough of this dumbing down.
 
While China and Russia build robot tanks and aircraft, America is busy trying to take naughty words off of reddit. The future looks bright, friends.
A nuclear warhead can at most kill a few hundred thousand people; a word like "tranny" kills billions.
 
Or another way you could counter this is by posting as a tranny or other minority, but just act so fucking outlandish and retarded that you sound like a right wing parody. What are they going to do, block you for expressing your heartfelt identity?

Be another Godfrey Elfwick (F) but blend in only slightly better.

Again, it'd be more of "you know what you're doing :mad:".
 
Back
Top Bottom