Meta needs to adopt a more stringent approach towards misinformation related to Covid-19 on its social networking platforms, the company’s independent Oversight Board has advised.
The recommendation arrives in response to Meta’s request for guidance on whether the company should continue removing misinformation related to Covid-19 under its harmful health misinformation policy. Meta started removing false Covid-19 claims from Facebook and Instagram in early 2020 at the onset of the pandemic. The company’s official list highlighting misleading claims developed over the course the pandemic, according to Oversight Board.
Meta’s approach to combat harmful content included labels, which directed viewers to its official Covid-19 information portal. However, following internal research into the efficacy of these labels to stop the spread of misinformation, Meta rolled back certain labels containing information on a range of issues, including vaccines, treatment, etc.
In its decision, the Board recommends Meta “continue to remove false content about COVID-19 that is ‘likely to directly contribute to the risk of imminent physical harm during the ongoing global public health emergency’.” The Board is of the opinion that Meta’s decision to cease labels makes it difficult for the Board to find a solution that satisfies both stakeholders and Board Members regarding how Covid-19 misinformation is addressed without violating human rights, especially the right to freedom of expression.
Following are the key highlights from Oversight Board’s 18-recommendations advisory opinion in response to Meta’s request.
Meta should be prepared for the time when the World Health Organisation (WHO) stops considering Covid-19 as a global health emergency, but local public health authorities continue to designate it as such. Meta should initiate a risk assessment process to identify appropriate measures to be taken in this scenario.
Meta should reassess the impact of its platforms’ technical infrastructure as experts have raised concerns that it amplifies harmful health misinformation. A human rights impact assessment should be commissioned on how Meta’s News Feed, recommendation algorithms and other features amplify harmful health misinformation and its impacts.
Meta should ensure more transparency around requests received from governments around the world to remove certain health-related content. The Board expresses concern over requests that target peaceful protestors, human rights defenders, and individuals criticising or questioning government’s responses to public health crises. Meta should be transparent and report regularly on state actor requests, the Board says.
The efforts to understand Covid-19 misinformation and Meta’s responses to it are hindered by lack of accessibility to the company’s data. Meta should, therefore, support independent research and promote understanding of Covid-19 misinformation.
Meta must also ensure its rules are transparent and clear to users to meet its human rights responsibilities. The company should explain the removal of content under each category of the Covid-19 claims and how it contributes to the risk of potential physical harm.
It is pertinent to note, however, that Meta is not bound to adopt Oversight Board’s recommendations.