Creating the Oversight Board

UPDATED

NOV 16, 2021

Every day, teams at Meta make difficult decisions about what content should stay up and what should come down based on the Facebook Community Standards or Instagram Community Guidelines. But given the size of our community and the reach of our platform, we created the Oversight Board to bring accountability to those decisions.

Why the board was created

In 2018, Meta CEO Mark Zuckerberg shared a blueprint outlining new and better ways for platforms like Facebook to remain accountable and bringing legitimacy to the rules that govern large communities on the internet.

With the input of scholars and experts, Meta put this note into action and devised an oversight board to provide an independent check on some of the most significant and difficult content decisions we make. The guiding idea behind the board was simple: Meta should not make so many important decisions about free expression and safety on our own.

Next, we underwent a global consultation process to better understand how to turn this vision into an institution. This consultation included input from experts all around the world, such as academics, technical experts, lawyers, designers and technologists, as well as input from members of the public. With this feedback, we were able to build the structures and documents that would serve as the foundation of governance for such a board. This included drafting a charter, establishing an independent trust and developing the board’s bylaws.

The board begins operations

In May 2020, the Oversight Board’s first 20 members were announced—an esteemed and thoughtful group who’ve worked in a variety of positions, including as professors, journalists and heads of state.

Following this announcement, members underwent training on our Community Standards and Community Guidelines, policy development processes, enforcement frameworks and the types of content decisions in scope for the board. Members also had a rigorous orientation around a new case management tool developed by Meta. This tool allows members to securely access and review pertinent case information from anywhere in the world.

The board began hearing cases in October 2020. Since then, it has issued a number of decisions and recommendations, which have already changed the way we moderate content for the billions of people on the Facebook app and Instagram.

Our goal has always been for the board to exist for years to come as an important piece of our broader strategy for content moderation. We’re committed to supporting the board as it develops as an institution and expands to 40 members. This includes continually examining the board’s scope and working to bring additional types of content outlined in the bylaws into that scope, a process we’ve already begun.

We hope the board can serve as a model for the future of content governance across our industry, as it continues to provide invaluable input into how we make some of the most consequential decisions regarding freedom of expression across the globe.