Meta's internal guidance for at-scale reviewers should be modified to ensure that flag-based visual depictions of gender identity that do not contain a human figure are understood as representations of a group defined by the gender identity of its members. This modification would clarify instructions for enforcement of this form of content at scale, whenever it contains a violating attack.
The Board will consider this implemented when Meta provides the Board with the changes to its internal guidance.
Our commitment: We are assessing Hate Speech policy guidance related to “visual depictions” of protected characteristics more generally, and will consider ways to incorporate this recommendation into that work.
Considerations: Our
Hate Speech policy allows speech about concepts, but not attacks on people based on their protected characteristics. Currently, as the Board notes, we consider a number of factors to determine whether a protected characteristic is referenced in content, including whether gender identity is depicted by visual signals which can include flags accompanied by other indicators. Additionally, as part of training internal reviewers, we specifically train on how to better identify indicators of visual depictions of protected characteristic groups, including LGBTQ+ people.
We previously conducted policy development to align on which content may be interpreted as an attack on “people” as opposed to attacks on “concepts,” the result of which led to implementation of context-specific policy guidance. Our
Hate Speech policy allows for attacks on concepts or institutions generally due to the importance of allowing legitimate speech. It also recognizes that attacks on people on the basis of protected characteristics such as national origin, race, ethnicity, disability, religious affiliation, caste, sexual orientation, sex, gender identity, or serious disease can create an environment of intimidation, and therefore we remove this speech when content targets a person or group of people with a hate speech attack based on protected characteristics. When there are signals that content attacking concepts, institutions, ideas, practices, or beliefs associated with a protected characteristic could contribute to discrimination, intimidation, or imminent physical harm, we may also remove that content on escalation.
More generally, we are in the process of reviewing some of our policy guidance related to visual depictions and plan to conduct policy development related to some forms of visual depiction indicators. We will assess ways to include this feedback and recommendation in that workstream. That being said, we also acknowledge that, at-scale, defining and enforcing solely on visual elements like flags , without certain additional contextual clues, as representations of a person or group with protected characteristics could lead to removal of legitimate speech. We will continue to provide updates on the upcoming policy work related to visual depiction indicators in future Oversight Board reports.