MirrorFair: Fixing Fairness Bugs in Machine Learning Software via Counterfactual Predictions
With the increasing utilization of Machine Learning (ML) software in critical domains such as employee hiring, college admission, and credit evaluation, ensuring fairness in the decision-making processes of underlying models has emerged as a paramount ethical concern. Nonetheless, existing methods for rectifying fairness issues encounter challenges in striking a harmonious balance between performance and fairness across diverse tasks and algorithms. Informed by the principles of counterfactual inference, this paper introduces MirrorFair, an innovative adaptive ensemble approach designed to mitigate fairness concerns. MirrorFair initially constructs a counterfactual dataset derived from the original data, training two distinct models—one on the original dataset and the other on the counterfactual dataset. Subsequently, MirrorFair adaptively combines these model predictions to generate fairer final decisions.
We conduct an extensive evaluation of MirrorFair and compare it with 14 existing methods across a diverse range of decision-making scenarios. Our findings reveal that MirrorFair achieves superior fairness with a minimal impact on performance compared to the existing methods. Specifically, in 93% of cases, MirrorFair surpasses the fairness and performance trade-off baseline proposed by the benchmarking tool Fairea, whereas the state-of-the-art method achieves this in only 88% of cases. Furthermore, MirrorFair consistently demonstrates its superiority across various tasks and algorithms, ranking first in balancing model performance and fairness in 83% of scenarios. In contrast, the state-of-the-art method achieves this in only 8% of cases. To foster future research endeavors, we have made all code, data, and results openly accessible to the research community.
Thu 18 JulDisplayed time zone: Brasilia, Distrito Federal, Brazil change
14:00 - 15:30 | SE4AI 1Journal First / Ideas, Visions and Reflections / Research Papers / Demonstrations at Mandacaru Chair(s): Qinghua Lu Data61, CSIRO | ||
14:00 18mTalk | Harnessing Neuron Stability to Improve DNN Verification Research Papers Hai Duong George Mason University, Dong Xu University of Virginia, ThanhVu Nguyen George Mason University, Matthew B Dwyer University of Virginia | ||
14:18 18mTalk | MirrorFair: Fixing Fairness Bugs in Machine Learning Software via Counterfactual Predictions Research Papers Ying Xiao King's College London / Southern University of Science and Technology, Jie M. Zhang King's College London, Yepang Liu Southern University of Science and Technology, Mohammad Reza Mousavi King's College London, Sicen Liu Southern University of Science and Technology, Dingyuan Xue Southern University of Science and Technology | ||
14:36 9mTalk | Using Run-time Information to Enhance Static Analysis of Machine Learning Code in Notebooks Ideas, Visions and Reflections Yiran Wang Linköping University, José Antonio Hernández López Linkoping University, Ulf Nilsson Linköping University, Daniel Varro Linköping University / McGill University Link to publication DOI | ||
14:45 9mTalk | Human-Imperceptible Retrieval Poisoning Attacks in LLM-Powered Applications Ideas, Visions and Reflections Quan Zhang Tsinghua University, Binqi Zeng Central South University, Chijin Zhou Tsinghua University, Gwihwan Go Tsinghua University, Heyuan Shi Central South University, Yu Jiang Tsinghua University | ||
14:54 18mTalk | DeepGD: A Multi-Objective Black-Box Test Selection Approach for Deep Neural Networks Journal First Zohreh Aghababaeyan University of Ottawa, Canada, Manel Abdellatif Software and Information Technology Engineering Department, École de Technologie Supérieure, Mahboubeh Dadkhah The School of EECS, University of Ottawa, Lionel Briand University of Ottawa, Canada; Lero centre, University of Limerick, Ireland | ||
15:12 9mTalk | Testing Learning-Enabled Cyber-Physical Systems with Large-Language Models: A Formal Approach Ideas, Visions and Reflections Xi Zheng Macquarie University, Aloysius K. Mok University of Texas at Austin, Ruzica Piskac Yale University, Yong Jae Lee University of Wisconsin Madison, Bhaskar Krishnamachari University of Southern California, Dakai Zhu The University of Texas at San Antonio, Oleg Sokolsky University of Pennsylvania, USA, Insup Lee University of Pennsylvania | ||
15:21 9mTalk | GAISSALabel: A tool for energy labeling of ML models Demonstrations Pau Duran Universitat Politècnica de Catalunya (UPC), Joel Castaño Fernández Universitat Politècnica de Catalunya (UPC), Cristina Gómez Universitat Politècnica de Catalunya, Silverio Martínez-Fernández UPC-BarcelonaTech Link to publication Pre-print |