Detecting Hate Speech with GPT-3

Detecting Hate Speech with GPT-3

By researchers at University of Toronto:

“Sophisticated language models such as OpenAI’s GPT-3 can generate hateful text that targets marginalized groups. Given this capacity, we are interested in whether large language models can be used to identify hate speech and classify text as sexist or racist.

We use GPT-3 to identify sexist and racist text passages with zero-, one-, and few-shot learning. We find that with zero- and one-shot learning, GPT-3 can identify sexist or racist text with an average accuracy between 55 per cent and 67 per cent, depending on the category of text and type of learning. In addition, with few-shot learning, the model’s accuracy can be as high as 85 per cent. Large language models have a role to play in hate speech detection, and with further development they could eventually be used to counter hate speech.”: Read The Full Paper

Back To News

Yann LeCun

Detecting Hate Speech with GPT-3

In conclusion, we are very excited to see the future developments. That will arise from the research associated with the cutting-edge GPT-3 Machine Learning system. Moreover, we expect to learn things about human beings that GPT-3 will teach us. And furthermore, discover in the various science fields.