City
Epaper

This is how Facebook detects harmful content with AI systems

By IANS | Updated: August 13, 2020 09:55 IST

New Delhi, Aug 13 For effective content moderation, Facebook is relying on three aspects of technology to transform ...

Open in App

New Delhi, Aug 13 For effective content moderation, Facebook is relying on three aspects of technology to transform its content review process across its family of apps.

The first aspect is called 'Proactive Detection' where Artificial intelligence (AI) can detect violations across a wide variety of areas without relying on users to report content to Facebook, often with greater accuracy than reports from users.

"This helps us detect harmful content and prevent it from being seen by hundreds or thousands of people," the company said in a statement.

'Automation' is the second aspect where AI systems have automated decisions for certain areas where content is highly likely to be violating.

"Automation also makes it easier to take action on identical reports, so our teams don't have to spend time reviewing the same things multiple times. These systems have become even more important during the Covid-19 pandemic with a largely remote content review workforce," said Jeff King, Director Product Management, Integrity at Facebook.

The third aspect is 'Prioritisation'.

Instead of simply looking at reported content in chronological order, AI prioritises the most critical content to be reviewed, whether it was reported to Facebook or detected by its proactive systems.

"This ranking system prioritizes the content that is most harmful to users based on multiple factors such as virality, severity of harm and likelihood of violation," Kind added.

However, Facebook admitted there are still areas where it's critical for people to review the content.

"For example, discerning if someone is the target of bullying can be extremely nuanced and contextual. In addition, AI relies on a large amount of training data from reviews done by our teams in order to identify meaningful patterns of behaviour and find potentially violating content".

For reviewing violations like spam, Facebook said it is going to use its automated systems first to review more content across all types of violations.

( With inputs from IANS )

Tags: Jeff KingFacebookFacebook connectivity
Open in App

Related Stories

BusinessMukesh Ambani’s Reliance Industries Share Price Jump Over 2% as Facebook Acquires 30% Stake in AI Venture

TechnologyMeta Layoffs: Facebook-Owned Firm to Cut 600 Employees From AI Unit

TechnologyCyber Crime Alert: How WhatsApp and Facebook Can Protect Your Money and Data from Scammers

TechnologyFacebook Down? Meta-Owned Website Not Loading for Some Users, Displays Message 'HTTP ERROR 500'

TechnologyMeta’s New Subscription Model: A Tough Sell for India?

Technology Realted Stories

TechnologyGoogle launches Gemini 3 AI

TechnologyPiyush Goyal launches trade intelligence and analytics portal to help exporters

TechnologySBI Chairman pitches for National Financial Grid to unify credit and digital infrastructure

TechnologyX, ChatGPT, several other sites face disruption; netizens complain of outage

TechnologyNFRA launches webinar series to create better financial reporting ecosystem