Social media giants warn of AI content moderation errors, as employees sent home
“We may see some longer response times and make more mistakes as a result." Image: REUTERS/File Photos
- AI and automated tools are now set to be the sole protectors of social media content, as employees are sent home by companies.
- Facebook has warned that “we may see some longer response times and make more mistakes" as a result.
Alphabet Inc’s YouTube, Facebook Inc (FB.O) and Twitter Inc (TWTR.N) warned on Monday that more videos and other content could be erroneously removed for policy violations, as the companies empty offices and rely on automated takedown software during the coronavirus pandemic.
In a blog post, Google said that to reduce the need for people to come into offices, YouTube and other business divisions are temporarily relying more on artificial intelligence and automated tools to find problematic content.
Such software is not always as accurate as humans, which leads to errors, it added, however. And “turnaround times for appeals against these decisions may be slower,” it said.
Facebook followed suit, saying it would work with contract vendors this week to send home all content reviewers home indefinitely, with pay.
The social media company drew public criticism last week for asking policy enforcers to continue coming to work, as it lacks secure technology to conduct moderation remotely.
Facebook also said the decision to rely more on automated tools, which learn to identify offensive material by analyzing digital clues for aspects common to previous takedowns, has limitations.
“We may see some longer response times and make more mistakes as a result,” it said.
Twitter said it too would step up use of similar automation, but would not ban users based solely on automated enforcement, because of accuracy concerns.
The three Silicon Valley internet services giants, like many companies worldwide, have asked employees and contractors to work from home if possible, to slow the fast-spreading respiratory disease. Mass gatherings for sports, cultural and religious events have been canceled globally.
Google said human review of automated policy decisions also would be slower for other products and phone support would be limited.
Its content rules cover submissions such as campaigns on its ad network, apps uploaded to the Google Play store and business reviews posted to Google Maps.
“Some users, advertisers, developers and publishers may experience delays in some support response times for non-critical services, which will now be supported primarily through our chat, email, and self-service channels,” Google said.
The content review operations of Google and Facebook span several countries, such as India, Ireland, Singapore and the United States.
Don't miss any update on this topic
Create a free account and access your personalized content collection with our latest publications and analyses.
License and Republishing
World Economic Forum articles may be republished in accordance with the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License, and in accordance with our Terms of Use.
The views expressed in this article are those of the author alone and not the World Economic Forum.
Stay up to date:
COVID-19
The Agenda Weekly
A weekly update of the most important issues driving the global agenda
You can unsubscribe at any time using the link in our emails. For more details, review our privacy policy.
More on Health and Healthcare SystemsSee all
Alaa Murabit and Amira Ghouaibi
November 12, 2024