Meta

Meta
政策
社群守则Meta 广告发布守则其他政策Meta 如何改进工作适龄内容

功能
我们打击危险组织和人物的方法我们对阿片类药物泛滥的处理方式我们维护诚信选举的方法我们打击错误信息的方法我们评估内容新闻价值的方法我们的 Facebook 动态版块内容排名方法我们对内容排名方法的解释Meta 无障碍理念

研究工具
内容库与内容库 API广告资料库工具其他研究工具和数据集

政策执行
检测违规内容采取措施

治理
治理创新监督委员会概览如何向监督委员会申诉监督委员会案件监督委员会建议设立监督委员会监督委员会:其他问题Meta 关于监督委员会的半年度更新报告追踪监督委员会的影响力

安全
威胁中断安全威胁威胁行为处理报告

报告
社群守则执行情况报告知识产权政府的用户数据收集情况依据当地法律实施内容限制网络中断广泛浏览内容报告监管报告和其他透明度报告

政策
社群守则
Meta 广告发布守则
其他政策
Meta 如何改进工作
适龄内容
功能
我们打击危险组织和人物的方法
我们对阿片类药物泛滥的处理方式
我们维护诚信选举的方法
我们打击错误信息的方法
我们评估内容新闻价值的方法
我们的 Facebook 动态版块内容排名方法
我们对内容排名方法的解释
Meta 无障碍理念
研究工具
内容库与内容库 API
广告资料库工具
其他研究工具和数据集
政策执行
检测违规内容
采取措施
治理
治理创新
监督委员会概览
如何向监督委员会申诉
监督委员会案件
监督委员会建议
设立监督委员会
监督委员会:其他问题
Meta 关于监督委员会的半年度更新报告
追踪监督委员会的影响力
安全
威胁中断
安全威胁
威胁行为处理报告
报告
社群守则执行情况报告
知识产权
政府的用户数据收集情况
依据当地法律实施内容限制
网络中断
广泛浏览内容报告
监管报告和其他透明度报告
政策
社群守则
Meta 广告发布守则
其他政策
Meta 如何改进工作
适龄内容
功能
我们打击危险组织和人物的方法
我们对阿片类药物泛滥的处理方式
我们维护诚信选举的方法
我们打击错误信息的方法
我们评估内容新闻价值的方法
我们的 Facebook 动态版块内容排名方法
我们对内容排名方法的解释
Meta 无障碍理念
研究工具
内容库与内容库 API
广告资料库工具
其他研究工具和数据集
安全
威胁中断
安全威胁
威胁行为处理报告
报告
社群守则执行情况报告
知识产权
政府的用户数据收集情况
依据当地法律实施内容限制
网络中断
广泛浏览内容报告
监管报告和其他透明度报告
政策执行
检测违规内容
采取措施
治理
治理创新
监督委员会概览
如何向监督委员会申诉
监督委员会案件
监督委员会建议
设立监督委员会
监督委员会:其他问题
Meta 关于监督委员会的半年度更新报告
追踪监督委员会的影响力
政策
社群守则
Meta 广告发布守则
其他政策
Meta 如何改进工作
适龄内容
功能
我们打击危险组织和人物的方法
我们对阿片类药物泛滥的处理方式
我们维护诚信选举的方法
我们打击错误信息的方法
我们评估内容新闻价值的方法
我们的 Facebook 动态版块内容排名方法
我们对内容排名方法的解释
Meta 无障碍理念
研究工具
内容库与内容库 API
广告资料库工具
其他研究工具和数据集
政策执行
检测违规内容
采取措施
治理
治理创新
监督委员会概览
如何向监督委员会申诉
监督委员会案件
监督委员会建议
设立监督委员会
监督委员会:其他问题
Meta 关于监督委员会的半年度更新报告
追踪监督委员会的影响力
安全
威胁中断
安全威胁
威胁行为处理报告
报告
社群守则执行情况报告
知识产权
政府的用户数据收集情况
依据当地法律实施内容限制
网络中断
广泛浏览内容报告
监管报告和其他透明度报告
中文(简体)
隐私政策服务条款Cookie
这条内容尚无中文(简体)版本

Home
Governance
Community Forums

Community Forums at Meta

更新日期 2025年10月12日
Community Forums bring together groups of people from all over the world to discuss tough issues, consider hard choices, and share their perspectives to improve the experiences people have across Meta’s technologies.

How it Works

Meta’s Community Forums allow us to learn directly from the people who use our platforms and technologies. These forums bring together thousands of people from around the world to weigh in on some of the tech industry’s toughest questions.
In each Community Forum, participants start by learning about a specific topic through carefully prepared educational materials. They then join small group discussions where they share their experiences and perspectives. Expert advisors are then available to answer questions before participants provide their final feedback through surveys.
Their responses, and the analysis of the results, produce insights on the public’s understanding of and concerns about these emerging technologies, and ultimately inform the development of our products and policies. For example, our Community Forum on the Metaverse played a direct role in Meta adding mute assist, a form of automatic speech detection in public worlds, to the catalog of tools available to creators on Horizon. We invest in Community Forums because it’s important that our products represent the people who use them.

How it Got Started

We started by looking at deliberative democratic mechanisms, such as Citizens Assemblies, that have been used to provide public input into government policies for years. An initial pilot was run on our approach to climate misinformation in 2022. Based on those learnings we explored how we might scale this approach to more people, and launched another Forum on the issue of bullying and harassment in the Metaverse. Both of these showed that Community Forums can provide rich insights for our product and policy development.

Our Partnerships

As with the collaborative nature of this work, Meta has spoken to and partnered with a variety of deliberative democracy experts, civil society organizations, government policymakers, and academics to ensure our forums are constructed in accordance with deliberative democracy best practices and standards. This process helps us mitigate against any biases while also sharing insights with others in the deliberative democracy community. The design and execution of our Forums to-date have been in partnership with Stanford's Deliberative Democracy Lab and the Behavioural Insights Team.
All Community Forums
TitleDates of Forum CountriesRepresentative SampleResultsGuiding Questions
2024 Community Forum on Generative AIOctober 2024India, Turkey, Nigeria, Saudi Arabia, South Africa 887Stanford Deliberative Democracy Lab report“How should AI agents provide proactive, personalized experiences to users?”
"How should AI agents and users interact?"
2023 Community Forum on Generative AIOctober 2023Brazil, Germany, Spain, United States1,545Stanford Deliberative Democracy Lab report “What principles should guide generative AI’s engagement with users?”
Community Forum on Bullying and Harassment in the MetaverseDecember 202232 countries 6,488Stanford Deliberative Democracy Lab report“To what extent do the platform owners, such as Meta, have a responsibility to act to protect against bullying and harassment, particularly since the metaverse is an immersive reality in which bullying and harassment may have severe consequences?”
Pilot Community Forum on Climate Misinfo2022Brazil, France, India, Nigeria, United States257Behavioral Insights Team pilot report “What approach should Meta take to climate content that may be misleading or confusing but does not contain a false claim that can be debunked by fact-checkers?”
New Community Forums
2025 Community Forum on AI Agents
AI innovation is transforming everyday tasks, boosting productivity and changing interactions with technology. AI agents are increasingly part of this progress, empowering individuals and organizations to leverage AI in new ways. That is why Meta is joining Stanford’s cross-company Community Forum designed to gather feedback from the public on AI and AI agents. As with previous Forums, this collaboration will be organized by Stanford University, with input from researchers at the Hoover Institution.
In this Forum, a representative sample of the public will offer their perspectives for agentic AI tools, and deliberate in small groups. This Forum will allow Meta and our peers to gather public input on critical questions the industry faces while developing AI agents.
We will update this page with the Forum’s results and findings.
Prior Community Forums: Findings and Impact
We will periodically update the impact that each of our Forums is having on our decisions over time.
Our 2024 Community Forum on Generative AI
We conducted a second Community Forum on GenAI which included a total of roughly 900 people across Nigeria, India, South Africa, Saudi Arabia, and Turkey. This forum solicited feedback from Meta’s users on how people should interact with AI agents and whether they would prefer these interactions to be more personalized. The results broadly suggest that the public in these regions supports a vision for personalized AI, as well as what they see as the appropriate parameters and guardrails for AI technologies. Specifically, we found several key findings on the principles that should underpin AI agents, including:
  • Personalization. Participants supported AI agents remembering their prior conversations to personalize their experience, as long as transparency and user controls are in place. They also supported proactive prompting with clarifying questions to generate personalized outputs from AI agents.
  • Localization. Participants were more supportive of culturally/regionally-tailored AI agents compared to standardized AI agents.
These findings offer a helpful perspective on global attitudes towards personalized AI experiences. With personal superintelligence on the horizon, AI products will not only be powerful, but deeply empowering. That means a future for AI that is more tailored to the needs of the individual. We see this public consultation as an important step toward understanding peoples’ perspectives on the path toward developing this technology.
For more details, view the full report from our partners at Stanford University’s Deliberative Democracy Lab here.
Our 2023 Community Forum on Generative AI
We announced our 2023 Community Forum on GenAI shortly after the technology began to capture the public mind because we wanted to better understand the underlying principles that should inform how chatbots provide guidance and advice, as well as how they should interact with people. This forum included a total of 1,545 participants from Brazil, Germany, Spain and the United States. The results of the Forum have helped shape some of the most foundational choices to date in our design of GenAI products and will continue to inform the direction of Generative AI products.
This Forum informed our strategy on how, and under what circumstances, chatbots can be personalized for users and what information they can remember to support the user experience.
  • For example, feedback from the public showed that participants preferred user controls that enable them to personalize experiences and clear transparency about what data was being used. This has guided features associated with transparency and disclosure of GenAI activities, such as user notifications.
  • In addition, the findings from the Forum informed company strategy on how to approach issues of AI chatbot memory in ways that are consistent with user preferences. For example, there was an overwhelming consensus that, as long as users are made aware, AI chatbots should use people’s past conversations to offer the best experience.
  • The Forum also helped a team of design and product experts decide to have a more neutral user experience as a default setting for AI agents to start with. Participants highlighted the importance of considering vulnerable users in the development of AI, which has informed our product strategy.
Given the rapidly changing nature of GenAI technologies, we purposely designed the 2023 Community Forum to provide us with direction at the foundational level which focuses on the values that underlie how our technology interacts with people. As a result, we can return to this input over time and it can provide lasting direction which scales to a large number of product decisions we have to make. We are committed to retaining this feedback and considering it in conjunction with the input we receive from subsequent Community Forums on these issues.
Our Community Forum on Bullying and Harassment in the Metaverse
A Community Forum on the Metaverse was conducted in collaboration with Stanford’s Deliberative Democracy Lab on the topic of bullying and harassment and was a first-of-its-kind experiment in global deliberation. We chose to focus on closed virtual spaces so that the forum could advise on policy and product development for virtual experiences such as Horizon Worlds. This Forum included 6,000 participants from 32 countries and functioned as an important pilot to establish proof of concept. Read more here.
Our Community Forum on Climate Misinformation
This Forum deliberated on the challenging topic of misleading climate change content. We brought together over 250 Facebook users across five countries, to ensure that we heard from people from different nationalities, ethnicities, socio-economic backgrounds, and political ideologies. This Forum functioned as an important pilot to establish proof of concept. Read more here.
Community Forums Process

Agenda Setting to determine the most impactful scope for a Forum.

  • Teams across Meta coordinate to determine significant and difficult innovation questions that would benefit from community consultation.
  • At Meta, we see these types of processes as a valuable way to meaningfully engage the public on complex issues. We consider a topic as appropriate for deliberation when it is:
    • Significant: it may be an important issue for society and technology.
    • Difficult: it poses clear dilemmas, tough tradeoffs or a lot of people are likely to have varying perspectives.
    • And has multiple solutions.

Recruit participants to deliver a representative sample from target countries.

  • To ensure quality deliberation and credible results, it is critical to get a true cross-section of communities to participate in our Community Forums.
  • We work with Stanford University’s Deliberative Democracy Lab and polling firms to recruit a representative sample for each country we host Forums in.
  • To reduce barriers to participation, participants are supported with access to technology, internet connection, and childcare as needed.

Provide education to participants that establish a contextual understanding of the issues that will be covered in the deliberation.

  • Objective, unbiased education is a key component so that participants can effectively grapple with the competing tradeoffs associated with the Forum’s topic. We develop materials with Stanford’s Deliberative Democracy Lab and other outside experts to ensure that e all participants have equal access to a baseline knowledge of the topic and the ability to engage in deliberation – regardless of their background.

Deliberate in groups of 8-10 participants on Stanford Deliberative Democracy Lab’s virtual AI-moderated deliberation platform.

  • Participants are guided by Stanford’s AI-facilitated platform to deliberate on the Forum’s topics, taking turns speaking and sharing their opinions with fellow participants.
  • Deliberation encourages people to reflect on the education materials, their own lived experience, and the perspectives shared by others.

Engage in Q&A sessions with subject matter experts to clarify questions that came out of their deliberations.

  • In between deliberations with their peers, participants attend a question and answer panel with industry experts. The panelists are responsible for clarifying participants’ understanding of contested issues, presenting novel tradeoffs on the topics participants are discussing, and correcting any misinformation that may have come up during deliberation.

Receive public feedback from the Forum.

  • As a part of Stanford’s deliberative poll methodology, participants provide their perspectives on the innovation topics before and after the Forum.
  • We receive these survey results, alongside key themes that emerged from the participants’ small group discussions.
  • Taken together, the public is able to provide more direct input into our innovation questions that considers the complexity of the topic.

Share results and implement the findings.

  • The final results report from our Forums are released publicly by Stanford’s Deliberative Democracy Lab.
  • We take the findings from the Community Forum and collaborate with teams across Meta to inform product and policy decisions with the feedback we’ve received from the public.
  • Our Community Forums often address long-term innovation questions. As such, the implementation of the public input can take course over time and influence multiple different decisions. We publish updates on our implementation progress here on our Transparency Center.
Meta
政策及信息公示平台
政策
政策执行
安全
功能
治理
研究工具
报告
中文(简体)