Meta's newest AI fairness benchmark measures even more granular bias markers

Meta, formerly known as Facebook, has announced a new AI fairness benchmark that is aimed at identifying even more granular bias markers. The benchmark builds on the company's previous efforts to measure AI fairness and improve the accuracy and equity of its algorithms.

The new benchmark, called the Fairness Flow Benchmark (FFB), is designed to help researchers and developers detect bias across different stages of the AI development process. It measures bias at three different stages: data collection, model training, and model deployment.

The FFB includes 12 different bias markers, which are designed to capture a wide range of potential sources of bias. These include markers for gender, age, race, religion, and other protected characteristics. In addition, the FFB includes markers for more subtle forms of bias, such as those related to social status or occupation.

Meta's focus on more granular bias markers is a response to the growing recognition that bias in AI is often more complex than simple binary categories. For example, gender bias might not simply be a matter of male or female, but might also be influenced by factors such as gender identity or sexual orientation.

One of the key features of the FFB is its ability to detect bias at multiple stages of the AI development process. This is important because bias can be introduced at any point, from data collection to model deployment. By identifying bias at each stage, developers can take steps to address it before it becomes embedded in the algorithm.

To test the effectiveness of the FFB, Meta conducted a pilot study using its own AI models. The study found that the FFB was able to identify bias in several areas, including gender and age. In response, the company made changes to its models to improve their fairness.

The FFB is an important step forward in the ongoing effort to make AI more fair and equitable. As AI becomes more pervasive in our daily lives, it is increasingly important that it is not only accurate but also free from bias. The FFB provides a useful tool for developers and researchers to identify and address bias at every stage of the AI development process.

Of course, the FFB is not a panacea for all AI fairness issues. It is still up to developers and researchers to use the benchmark effectively and to take action to address any bias that is identified. In addition, the FFB is just one of many tools that are available to improve AI fairness. Other approaches, such as increasing diversity in AI development teams, also have an important role to play.


Customer-driven product development is the next era of collaborative automation.

Collaborative automation

Nearly 50% of phishing attacks in 2021 aimed at government employees were attempted credential theft

Credential theft

How to Create a Strong Password?

Strong Password
Strong Password

How to Create a Strong Password?

3D skin

Columbia researchers bioprint seamless 3D skin grafts for burn patients


Connected Care is Coming: A Look at the Advancement of Medical IoT

Low-cost purification

New technology could help capture carbon dioxide for low-cost purification