Policies that outline what is and isn't allowed on our apps.
Policies for ad content and business assets.
Other policies that apply to Meta technologies.
How we update our policies, measure results, work with others, and more.
Explore how we help teens have safe, positive experiences on Facebook and Instagram.
How we approach dangerous organizations and individuals.
How we support communities in the face of the opioid epidemic.
How we help prevent interference, empower people to vote and more.
How we work with independent fact-checkers, and more, to identify and take action on misinformation.
How we assess content for newsworthiness.
How we reduce problematic content in News Feed.
How we build AI systems.
Comprehensive access to public data from Facebook and Instagram
Comprehensive and searchable database of all ads currently running across Meta technologies
Additional tools for in-depth research on Meta technologies and programs
Quarterly report on how well we're doing at enforcing our policies on the Facebook app and Instagram.
Report on how well we're helping people protect their intellectual property.
Report on government request for people's data.
Report on when we restrict content that's reported to us as violating local law.
Report on intentional internet restrictions that limit people's ability to access the internet.
Quarterly report on what people see on Facebook, including the content that receives the widest distribution during the quarter.
Download current and past regulatory reports for Facebook and Instagram.
Meta’s services help people freely express themselves; we take pride in the role we play in fostering people’s ability to exercise important rights around the world. At the same time, we work hard to prevent the spread of potentially harmful content. We do this by developing industry-leading processes and tools to reduce the likelihood that people see content spreading hate, misinformation or inciting violence, for example. These include our global Content Distribution Guidelines, which detail the kinds of distribution limits we place on problematic or low quality content, as well as our Community Standards and Community Guidelines which describe content that we may remove from our apps. During periods of heightened risk to safety, either on or off our platforms, we have the ability to take additional temporary steps if needed.
Temporary Strategies To Help People Stay Safe
During critical moments such as elections, or in situations with elevated risk of violence or other severe human rights risks, we are especially mindful of the need to carefully tailor our approach to keeping people safe while protecting their ability to express themselves. As such, our teams closely monitor trends on our platforms and investigate situations to determine whether and how best to respond. As appropriate, we may apply limited, proportionate, and time-bound measures that can be quickly implemented to address a specific, emerging risk.
In these moments, we monitor real-world events and track different metrics on our platforms, including things like how much violating content is on Facebook, or whether we’re starting to see new forms of abuse where we need to quickly adjust our response. For example, if we see an increase in violations of a specific policy, we investigate the nature and size of the problem before determining which measures, if any, we need to use in order to address the problem. In some cases, we may further reduce the visibility of certain types of content, above our standard reductions, that may not necessarily violate our Community Standards on hate speech, violence and incitement, or bullying and harassment, but come close to the line. To respond to other risks, we may reduce the distribution of content more significantly if it’s posted from accounts that have recently and repeatedly posted violating content, or if it is likely to violate our policies but we need additional time to review. In some circumstances, we may also reduce the distribution of widely shared content in order to slow its overall spread. This is particularly helpful when the content could be misinformation or incite violence. If our teams determine that a piece of content violates our policies, we will remove it, even if its visibility has already been reduced.
We also evaluate features in our apps to see how they could be misused during certain events, and in some cases temporarily change those features based on the risks we see. For example, during times of heightened risk to public safety, we may take additional steps to safeguard users' privacy and secure personally identifiable information by removing the visibility of friends lists on Facebook or follower/following lists for private accounts on Instagram. On Messenger, we may sometimes decrease the number of times people can forward messages in an effort to reduce the spread of misinformation during a crisis. There are also times when we limit the type-ahead suggestions people may see in the Search bar on Facebook or Instagram to minimize the chances that people inadvertently encounter higher-risk topics during critical moments. We also may include links to reliable information in relevant searches about certain key events.
We do not implement any of these measures lightly — we know that there could be unintended consequences, like inadvertently limiting harmless or even helpful content, and we seek to take steps that are proportionate to the risk and minimize this impact. For example, when we lower the number of times people can forward messages in Messenger, that could also affect people’s ability to easily share informative content. Or when we take stronger steps to move content lower in Feed that our systems detect might violate our policies, we may also be reducing the distribution of content that our systems have detected incorrectly.
Ongoing Monitoring And Returning To Normal Operations
High risk situations are often complex, fast-moving moments that can be adversarial, so there are no one-size-fits-all solutions. Throughout the relevant time period, we continue to monitor impacts to our platform and the people on it, in terms of both expression and safety. As a result of this monitoring, we may adjust our measures in response to any spikes or changes in signals we’re tracking. Once we see signals return to normal levels and determine the risk to safety on or off our platform has subsided, we will turn off the associated temporary measure.
We know our work requires ongoing vigilance, investments and a willingness to apply learnings from each situation we encounter to refine and improve our approach. Any temporary measures we may take to mitigate safety risks are consistent with our human rights policy and related principles, particularly those of necessity and proportionality. You can learn more about our work to prevent or mitigate human rights risks in our human rights policy, our human rights report, our human rights page, and in the audits we have requested as part of our commitment to independent verification of how we enforce our standards and conduct Human Rights Due Diligence.