AI video screening still a long way off, says Facebook executive
Facebook Inc.'s chief artificial intelligence scientist said the company is years away from being able to use software to automatically screen live video for extreme violence.
Yann LeCun's comments follow the March livestream of the Christchurch mosque shootings in New Zealand.
'This problem is very far from being solved,' LeCun said Friday during a talk at Facebook's AI Research Lab in Paris.
Facebook was criticised for allowing the Christchurch attacker to broadcast the shootings live without adequate oversight that could have resulted in quicker take-downs of the video. It also struggled to prevent other users from re-posting the attacker's footage.
LeCun said livestreams of violence presented numerous problems for automated systems, in particular the disturbing audio that accompanies videos of extreme violence, such as shootings or beheadings. A system needed to be trained on both picture and sound, he said, and would likely have to incorporate information about the individual posting the video and content they had recently published.
Another problem is that there was not enough data to train an AI to reliably detect such videos. 'Thankfully, we don't have a lot of examples of real people shooting other people,' he said.
While there were plenty of examples from movies of simulated violence that could be used for software training, LeCun said a system would have trouble differentiating between real violence and action movies, and would block both -- even though posting something from a movie is allowed.
Jerome Pesenti, Facebook's vice president of AI, said it was the company's goal to use a combination of human reviewers and automated systems to remove prohibited content -- such as videos of extreme violence -- as soon as possible.
But Pesenti said that if automated software had never encountered a particular video before and its confidence in classifying the content as prohibited was low, a human reviewer would have to screen the video and make a determination.
Facebook has made progress in automatically detecting and blocking certain sub-categories of extremist content, the company has said, and can now spot and block the posting of 99% of content linked to the terrorist group al-Qaeda.
But detecting and blocking all extremist content -- regardless of origin -- is a 'very hard problem,' LeCun said.