Policies that outline what is and isn't allowed on our apps.
Policies for ad content and business assets.
Other policies that apply to Meta technologies.
How we update our policies, measure results, work with others, and more.
Explore how we help teens have safe, positive experiences on Facebook and Instagram.
How we approach dangerous organizations and individuals.
How we support communities in the face of the opioid epidemic.
How we help prevent interference, empower people to vote and more.
How we work with independent fact-checkers, and more, to identify and take action on misinformation.
How we assess content for newsworthiness.
How we reduce problematic content in News Feed.
How we build AI systems.
Comprehensive access to public data from Facebook and Instagram
Comprehensive and searchable database of all ads currently running across Meta technologies
Additional tools for in-depth research on Meta technologies and programs
Quarterly report on how well we're doing at enforcing our policies on the Facebook app and Instagram.
Report on how well we're helping people protect their intellectual property.
Report on government request for people's data.
Report on when we restrict content that's reported to us as violating local law.
Report on intentional internet restrictions that limit people's ability to access the internet.
Quarterly report on what people see on Facebook, including the content that receives the widest distribution during the quarter.
Download current and past regulatory reports for Facebook and Instagram.
2025-001-FB-UA, 2025-002-FB-UA
Today, October 8, 2024, the Oversight Board selected a case bundle appealed by Facebook users regarding posts that include depictions of the flag used by South Africa between 1928 and 1994.
The first piece of content is a photograph of a uniformed soldier carrying the flag with a caption that says “Share if you served under this flag.” The second piece of content is an album of 6 photographs – one of which is the flag. The other photographs include a black ice cream vendor riding a bicycle and selling to white children, a beach, a game board, candy cigarettes, and a toy “cracker” gun. The caption expresses fondness for the previous era and asks the audience to “read between the lines,” followed by a winking face and an “OK” hand emoji.
Meta determined that each of the two pieces of content did not violate our policies on Dangerous Organizations and Individuals or Hate Speech, as laid out in the Facebook Community Standards, and left the content up.
We will implement the Board’s decision once it has finished deliberating, and we will update this post accordingly. Please see the board’s website for the decision when they issue it.
We welcome the Oversight Board's decision today, April 29, 2025, on this case. The Board upheld Meta’s decision to leave up the content in both cases.
After conducting a review of the recommendations provided by the Board, we will update this post with initial responses to those recommendations.
As part of its ongoing human rights due diligence, Meta should take all of the following steps in respect of the January 7 updates to the Hateful Conduct Community Standard. First, it should identify how the policy and enforcement updates may adversely impact populations in global majority regions. Second, Meta should adopt measures to prevent and/or mitigate these risks and monitor their effectiveness. Third, Meta should update the Board on its progress and learnings every six months, and report on this publicly at the earliest opportunity.
The Board will consider this recommendation implemented when Meta provides the Board with robust data and analysis on the effectiveness of its prevention or mitigation measures on the cadence outlined above, and when Meta reports on this publicly.
Commitment Statement: We will assess the feasibility of this multi-part recommendation.
Considerations: Meta conducts ongoing, integrated, human rights due diligence to identify, prevent, mitigate and address potential adverse human rights impacts related to our policies, products and operations in line with the UNGPs, related guidance, and our human rights policy. Ahead of the January 7th changes, we assessed the risks of the changes and took into account relevant mitigations, such as the availability of other policies and user reports to address potentially harmful content.
We will assess the feasibility of implementing this recommendation and provide updates in future reports to the Oversight Board. We will also bundle future updates for this recommendation under recommendation #1 in the Gender Identity Debate Videos case.
To improve the clarity of its Dangerous Organizations and Individuals Community Standard, Meta should adopt a single, clear and comprehensive explanation of how its prohibitions and exceptions under this Community Standard apply to designated hateful ideologies.
The Board will consider this recommendation implemented when Meta adopts a single, clear and comprehensive explanation of its rule and exceptions related to designated hateful ideologies (under “we remove”).
Commitment Statement: We will update our Dangerous Organizations and Individuals Community Standard to clarify our approach to content involving designated hateful ideologies.
Considerations: In line with prior commitments to the Board to update our Dangerous Organizations and Individuals Community Standard, we also plan to clarify that content that glorifies, supports, represents, or references a hateful ideology violates this policy in line with our policy rationale. In the body of our Community Standard, we currently explain how we define those hateful ideologies and groups. However, we recognize that we could clarify this approach throughout our external policy and will do so as part of ongoing work to update and clarify this Community Standard. We will provide an update on the status of this work in a future report to the Board.
To improve the clarity of its Dangerous Organizations and Individuals Community Standard, Meta should list apartheid as a standalone designated hateful ideology in the rules.
The Board will consider this recommendation implemented when Meta adds apartheid to its list of designated hateful ideologies.
Commitment Statement: We will conduct an initial assessment to better understand how the term “apartheid” is used on our platforms and consider next steps based on its findings.
Considerations: We will conduct an initial assessment to better understand how the term “apartheid” is used on our platforms and consider next steps based on its findings.
To improve clarity to reviewers of its Dangerous Organizations and Individuals Community Standard, Meta should provide more global examples to reviewers of prohibited glorification, support and representation of hateful ideologies, including examples that do not directly name the listed ideology.
The Board will consider this recommendation implemented when Meta provides updated internal guidance to the Board including more global examples, including ones that do not directly name the listed ideology.
Commitment Statement: We will update our internal guidance for the Dangerous Organizations and Individuals Community Standard with more examples, including illustrative global examples for hateful ideologies.
Considerations: We will update examples in our internal guidance for content reviewers to include a more global representation of content that glorifies, supports, or represents a hateful ideology. As part of this process, we will work with subject matter experts and teams with regional and language expertise to identify additional illustrative examples to guide internal reviewers. We will provide updates in future reports to the Oversight Board.