TY - GEN
T1 - Preventing Discriminatory Decision-making in Evolving Data Streams
AU - Wang, Zichong
AU - Saxena, Nripsuta
AU - Yu, Tongjia
AU - Karki, Sneha
AU - Zetty, Tyler
AU - Haque, Israat
AU - Zhou, Shan
AU - Kc, Dukka
AU - Stockwell, Ian
AU - Wang, Xuyu
AU - Bifet, Albert
AU - Zhang, Wenbin
N1 - Publisher Copyright:
© 2023 ACM.
PY - 2023/6/12
Y1 - 2023/6/12
N2 - Bias in machine learning has rightly received significant attention over the past decade. However, most fair machine learning (fair-ML) works to address bias in decision-making systems has focused solely on the offline setting. Despite the wide prevalence of online systems in the real world, work on identifying and correcting bias in the online setting is severely lacking. The unique challenges of the online environment make addressing bias more difficult than in the offline setting. First, Streaming Machine Learning (SML) algorithms must deal with the constantly evolving real-time data stream. Secondly, they need to adapt to changing data distributions (concept drift) to make accurate predictions on new incoming data. Incorporating fairness constraints into this already intricate task is not straightforward. In this work, we focus on the challenges of achieving fairness in biased data streams while accounting for the presence of concept drift, accessing one sample at a time. We present Fair Sampling over Stream (FS2), a novel fair rebalancing approach capable of being integrated with SML classification algorithms. Furthermore, we devise the first unified performance-fairness metric, Fairness Bonded Utility (FBU), to efficiently evaluate and compare the trade-offs between performance and fairness across various bias mitigation methods. FBU simplifies the comparison of fairness-performance trade-offs of multiple techniques through one unified and intuitive evaluation, allowing model designers to easily choose a technique. Overall, extensive evaluations show our measures surpass those of other fair online techniques previously reported in the literature.
AB - Bias in machine learning has rightly received significant attention over the past decade. However, most fair machine learning (fair-ML) works to address bias in decision-making systems has focused solely on the offline setting. Despite the wide prevalence of online systems in the real world, work on identifying and correcting bias in the online setting is severely lacking. The unique challenges of the online environment make addressing bias more difficult than in the offline setting. First, Streaming Machine Learning (SML) algorithms must deal with the constantly evolving real-time data stream. Secondly, they need to adapt to changing data distributions (concept drift) to make accurate predictions on new incoming data. Incorporating fairness constraints into this already intricate task is not straightforward. In this work, we focus on the challenges of achieving fairness in biased data streams while accounting for the presence of concept drift, accessing one sample at a time. We present Fair Sampling over Stream (FS2), a novel fair rebalancing approach capable of being integrated with SML classification algorithms. Furthermore, we devise the first unified performance-fairness metric, Fairness Bonded Utility (FBU), to efficiently evaluate and compare the trade-offs between performance and fairness across various bias mitigation methods. FBU simplifies the comparison of fairness-performance trade-offs of multiple techniques through one unified and intuitive evaluation, allowing model designers to easily choose a technique. Overall, extensive evaluations show our measures surpass those of other fair online techniques previously reported in the literature.
KW - Concept Drift
KW - Data Stream
KW - Fairness
KW - Fairness Drift
U2 - 10.1145/3593013.3593984
DO - 10.1145/3593013.3593984
M3 - Conference contribution
AN - SCOPUS:85163633797
T3 - ACM International Conference Proceeding Series
SP - 149
EP - 159
BT - Proceedings of the 6th ACM Conference on Fairness, Accountability, and Transparency, FAccT 2023
PB - Association for Computing Machinery
T2 - 6th ACM Conference on Fairness, Accountability, and Transparency, FAccT 2023
Y2 - 12 June 2023 through 15 June 2023
ER -