As artificial intelligence systems become more prevalent in our daily lives, ensuring these technologies are fair and representative of diverse populations is increasingly critical. A recent study, 鈥溾, conducted by , an Associate Professor of Electrical Engineering at , and his research group (see the bottom of the page for author details) at 性视界 University, introduces an innovative approach to measuring and promoting diversity in AI image retrieval systems. Their work addresses a key challenge in the field: how to ensure retrieved images reflect the true diversity of society across multiple intersecting demographic groups.
Key Insight: Current Approaches Fall Short on Intersectional Representation
“Ensuring representation across individual groups (e.g., given by gender or race) does not guarantee representation across intersectional groups (e.g., given by gender and race).” [1]
The researchers found that existing methods for promoting diversity in image retrieval often focus on balancing representation across a small number of pre-defined groups, typically based on single attributes like gender or race. However, they argue that this approach fails to account for intersectional groups, those defined by multiple overlapping attributes. For example, a system may retrieve an equal number of men and women but still under-represent women of color. The study demonstrates that optimizing for individual group representation does not necessarily lead to fair representation of intersectional groups.
Key Insight: A New Metric for Multi-Group Representation
鈥淲e propose a metric called Multi-group Proportional Representation (MPR) to quantify the representation of intersectional groups in retrieval tasks. MPR measures the worst-case deviation between the average values of a collection of representation statistics computed over retrieved items relative to a reference population whose representation we aim to match.鈥 [2]
To address the current representation gap, the researchers developed a metric called Multi-Group Proportional Representation (MPR), which quantifies how well a set of retrieved images represents diverse intersectional groups compared to a reference population. Crucially, MPR can measure representation across a large or even infinite number of overlapping groups, defined by complex combinations of attributes. This allows for a much more nuanced and comprehensive assessment of diversity and representation.
Key Insight: Scalability and Flexibility
“MPR offers a more flexible, scalable, and theoretically grounded metric for multi-group representation in retrieval.” [3]
A key advantage of the MPR approach is its scalability and flexibility. Unlike methods that rely on pre-defined groups, MPR can handle an arbitrary number of intersectional groups defined by complex functions. The researchers provide theoretical guarantees on the sample complexity required to estimate MPR accurately. They also demonstrate how MPR can be efficiently computed for several practical function classes, including linear functions and decision trees. This makes MPR a powerful and adaptable tool for measuring and optimizing representation in large-scale retrieval systems.
Key Insight: Ethical and Practical Considerations
鈥淭here are legal and regulatory risks with overreliance on a single metric for fairness, especially if this metric is used to inform policy and decision-making.鈥 [4]
While MPR is a powerful tool, the authors caution against viewing it as a standalone solution. Fairness is multidimensional, and overreliance on a single metric can lead to unintended consequences, such as reinforcing stereotypes or overlooking other forms of harm. Furthermore, the researchers warn that the deployment of MPR by companies could result in 鈥渆thics-washing鈥, where, even when systems exhibit representational harms, firms claim them to be fair based on their use of a fairness metric, like MPR. Finally, to ensure that the results of MPR are diverse, ethical, and fair, the researchers suggest utilizing datasets that are curated to guarantee they represent diverse populations. Failing to do so can result in propagating biases throughout the system.
Why This Matters
For C-suite executives, the introduction of Multi-Group Proportional Representation (MPR) signals a transformative step in aligning artificial intelligence systems with values of fairness and inclusivity. MPR tackles a critical shortfall in current AI practices: the failure to represent diverse intersectional groups in image retrieval and similar applications. By quantifying proportional representation, MPR offers a scalable and actionable framework for mitigating bias while preserving the functionality of the retrieval system.
Adopting MPR isn鈥檛 just an ethical responsibility, it鈥檚 a strategic imperative. Inclusive AI systems foster trust among consumers and employees, safeguard against reputational and regulatory risks, and enhance decision-making by accurately reflecting the diversity of society. With tools like MPR and the Multi-group Optimized Proportional Retrieval (MOPR) algorithm, organizations can lead in embedding fairness into their technological foundations, transforming inclusivity from a compliance checkbox into a competitive advantage.
References
[1] Alex Osterling, Claudio Mayrink Verdun, Carol Xuan Long, Alexander Glynn, Lucas Monteiro Paes, Sajani Vithana, Martina Cardone, and Flavio du Pin Calmon, 鈥淢ulti-Group Proportional Representation in Retrieval,鈥 arXiv preprint arXiv:2407.08571 (2024): 1-48, 2.
[2] Alex Osterling et al., 鈥淢ulti-Group Proportional Representation in Retrieval鈥, 2.
[3] Alex Osterling et al., 鈥淢ulti-Group Proportional Representation in Retrieval鈥, 2.
[4] Alex Osterling et al., 鈥淢ulti-Group Proportional Representation in Retrieval鈥, 5.
Meet the Authors

is a PhD student at 性视界 under the mentorship of and and is supported by the . They are broadly interested in fair, interpretable, and trustworthy machine learning, and their current projects apply information theoretic tools to problems in fairness and representation learning.

is a mathematician working with mathematics of AI and machine learning at 性视界鈥檚 School of Engineering and Applied Sciences under the mentorship of . His research focuses on trustworthy machine learning, exploring concepts such as fairness and arbitrariness, and also on mechanistic interpretability techniques for large generative models.

is a 4th-year Ph.D. student at 性视界, advised by . She completed her undergraduate degree in Math and Computer Science at and previously interned at and . Her research interest lies in Responsible and Trustworthy Machine Learning, and her work spans LLM watermarking, algorithmic fairness, multiplicity, and more.

has a degree in Applied Mathematics from 性视界 University, and completed a research fellowship at the 性视界 John A. Paulson School of Engineering and Applied Sciences focused on developing software pipelines to analyze and audit algorithms with novel techniques. He currently works as a Data Scientist at C3 AI.

is an Applied Mathematics Ph.D. candidate at working with and a student researcher at in the Gemini Safety Team. He uses theoretical insights to develop safe and trustworthy AI and ML systems. Their research is driven by the belief that AI and ML systems should not only be accurate and efficient but also transparent, fair, and aligned with human values and societal norms.

is a Postdoctoral Research Fellow at the 性视界 John A. Paulson School of Engineering and Applied Sciences. Her research interests include information theory, private information retrieval, and machine learning.

works in the Electrical and Computer Engineering Department at the University of Minnesota as an Assistant Professor. From July 2015 to August 2017, she was a post-doctoral research fellow in the Electrical and Computer Engineering Department at UCLA Henry Samueli School. She received her B.Sc. and M.Sc. from Politecnico di Torino in 2009 and 2011, respectively. As part of a Double Degree program, in 2011 she also earned a M.Sc. from T茅l茅com ParisTech – EURECOM. She completed her Ph.D. in Electronics and Communications at EURECOM – T茅l茅com ParisTech.

is an Associate Professor of Electrical Engineering at . Before joining 性视界 he was a social good post-doctoral fellow at in Yorktown Heights, New York. He received his Ph.D. in at MIT. His main research interests are information theory, signal processing, and machine learning.