Skip to main content

LOLA takes advantage of the latest advances in natural language processing and behavioural theory

Emotion detection engine developed to combat cyberbullying and fake news

A digital tool designed to detect fake news, cyberbullying and other online harms is being developed at the University of Exeter Business School.

“LOLA” uses sophisticated artificial intelligence to detect emotional undertones in language such as anger, fear, joy, love, optimism, pessimism and trust.

Unlike human moderators it can analyse 25,000 texts per minute, and has been found to detect harmful behaviour such as cyberbullying, hatred and Islamophobia with up to 98% accuracy.

Developed by a team led by Dr David Lopez, from the Initiative for Digital Economy Exeter (INDEX), LOLA takes advantage of the latest advances in natural language processing and behavioural theory.

“In the online world the sheer volume of information makes it harder to police and enforce abusive behaviour,” said Dr Lopez.

“We believe solutions to address online harms will combine human agency with AI-powered technologies that would greatly expand the ability to monitor and police the digital world.

“Our solution relies on the combination of recent advances in natural language processing to train an engine capable of extracting a set of emotions from human conversations (tweets) and behavioural theory to infer online harms arising from these conversations.”

Such is LOLA’s potential in the battle against misinformation that it has led to collaborations with the Spanish government and Google.

In a recent experiment, LOLA was found to pinpoint those responsible for the cyberbullying of Greta Thunberg on Twitter.

It has also been used to spot fake news about Covid-19, detecting the fear and anger so often used to pedal misinformation and singling out the accounts responsible.

LOLA grades each tweet with a severity score, and sequences them: ‘most likely to cause harm’ to ‘least likely’. Those at the top are the tweets which score highest in toxicity, obscenity and insult.

This kind of analysis could be a valuable tool for cybersecurity services, at a time when social media companies are under increasing pressure to tackle online harms.

The government is in the process of creating a new regulatory framework for online safety, giving digital platforms a duty of care for their users.

Dr Lopez added: “The ability to compute negative emotions (toxicity, insult, obscenity, threat, identity hatred) in near real time at scale enables digital companies to profile online harm and act pre-emptively before it spreads and causes further damage.”

Date: 7 July 2020