Policies that outline what is and isn't allowed on our apps.
Policies for ad content and business assets.
Other policies that apply to Meta technologies.
How we update our policies, measure results, work with others, and more.
Explore how we help teens have safe, positive experiences on Facebook and Instagram.
How we approach dangerous organizations and individuals.
How we support communities in the face of the opioid epidemic.
How we help prevent interference, empower people to vote and more.
How we work with independent fact-checkers, and more, to identify and take action on misinformation.
How we assess content for newsworthiness.
How we reduce problematic content in News Feed.
How we build AI systems.
Comprehensive access to public data from Facebook and Instagram
Comprehensive and searchable database of all ads currently running across Meta technologies
Additional tools for in-depth research on Meta technologies and programs
Quarterly report on how well we're doing at enforcing our policies on the Facebook app and Instagram.
Report on how well we're helping people protect their intellectual property.
Report on government request for people's data.
Report on when we restrict content that's reported to us as violating local law.
Report on intentional internet restrictions that limit people's ability to access the internet.
Quarterly report on what people see on Facebook, including the content that receives the widest distribution during the quarter.
Download current and past regulatory reports for Facebook and Instagram.
2020-003-FB-UA
On December 1, 2020, the Oversight Board selected a case appealed by someone on Facebook regarding historical photos purportedly showing churches in Baku and a caption indicating disdain for Azerbaijani people and support for Armenia in the Nagorno-Karabakh dispute.
Meta took down this content for violating our policy on hate speech, as laid out in the Facebook Community Standards. We do not allow hate speech on Facebook because it creates an environment of intimidation and exclusion, and in some cases, may promote real-world violence.
On January 28, 2021, The board decided to uphold Meta’s decision so we have taken no further action related to this case or the content.
On February 25, 2021, Meta responded to the board’s recommendation for this case. We are still assessing the feasibility of the recommendation.
Go beyond the Community Standard that Facebook is enforcing, and add more specifics about what part of the policy they violated.
Our commitment: We will continue to explore how best to provide transparency to people about enforcement actions, within the limits of what is technologically feasible.
Considerations: Over the past several years, we’ve invested in improving the experiences for people when we remove their content, and we have teams dedicated to continuing to improve these. As part of this work, we updated our notifications to inform people under which Community Standard a post was taken down (for example, Hate Speech, Adult Nudity and Sexual Activity, etc.), but we agree with the board that we’d like to provide more.
When a content reviewer reviews a post and determines it violates a policy, they often provide some additional data to our systems about the type of violation, but not always to the granularity of each line in the policy. Additionally, when we build technology to take automated action, it is often at the level of a policy area (e.g., Hate Speech) as it is not technologically feasible to create separate AI systems for each individual line in the policy. We understand the benefit in additional detail and will continue to explore how best to provide additional transparency.
Next steps: Our teams will continue to explore potential ways to address this challenge. We will provide updates with any future developments..