{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T05:42:29Z","timestamp":1775626949290,"version":"3.50.1"},"reference-count":137,"publisher":"Association for Computing Machinery (ACM)","issue":"6","license":[{"start":{"date-parts":[[2025,2,10]],"date-time":"2025-02-10T00:00:00Z","timestamp":1739145600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Comput. Surv."],"published-print":{"date-parts":[[2025,6,30]]},"abstract":"<jats:p>The impact of automated decision-making systems on human lives is growing, emphasizing the need for these systems to be not only accurate but also fair. The field of algorithmic fairness has expanded significantly in the past decade, with most approaches assuming that training and testing data are drawn independently and identically from the same distribution. However, in practice, differences between the training and deployment environments exist, compromising both the performance and fairness of the decision-making algorithms in real-world scenarios. A new area of research has emerged to address how to maintain fairness guarantees in classification tasks when the data generation processes differ between the source (training) and target (testing) domains. The objective of this survey is to offer a comprehensive examination of fair classification under distribution shift by presenting a taxonomy of current approaches. The latter is formulated based on the available information from the target domain, distinguishing between adaptive methods, which adapt to the target environment based on available information, and robust methods, which make minimal assumptions about the target environment. Additionally, this study emphasizes alternative benchmarking methods, investigates the interconnection with related research fields, and identifies potential avenues for future research.<\/jats:p>","DOI":"10.1145\/3637438","type":"journal-article","created":{"date-parts":[[2023,12,15]],"date-time":"2023-12-15T11:26:58Z","timestamp":1702639618000},"page":"1-32","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Preserving the Fairness Guarantees of Classifiers in Changing Environments: A Survey"],"prefix":"10.1145","volume":"57","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2300-312X","authenticated-orcid":false,"given":"Ainhize","family":"Barrainkua","sequence":"first","affiliation":[{"name":"Basque Center for Applied Mathematics (BCAM), Bilbao, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0455-1200","authenticated-orcid":false,"given":"Paula","family":"Gordaliza","sequence":"additional","affiliation":[{"name":"Universidad P\u00fablica de Navarra (UPNA), Pamplona, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4683-8111","authenticated-orcid":false,"given":"Jose A.","family":"Lozano","sequence":"additional","affiliation":[{"name":"Basque Center for Applied Mathematics (BCAM), Spain and University of the Basque Country UPV\/EHU, Bilbao, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8819-306X","authenticated-orcid":false,"given":"Novi","family":"Quadrianto","sequence":"additional","affiliation":[{"name":"Predictive Analytics Lab, University of Sussex, UK and Basque Center for Applied Mathematics (BCAM), Spain and Monash University, Brighton, Indonesia"}]}],"member":"320","published-online":{"date-parts":[[2025,2,10]]},"reference":[{"key":"e_1_3_2_2_2","volume-title":"Proceedings of the Neurips Workshop on Algorithmic Fairness through the Lens of Causality and Interpretability","author":"Adragna Robert","year":"2020","unstructured":"Robert Adragna, Elliot Creager, David Madras, and Richard Zemel. 2020. Fairness and robustness in invariant learning: A case study in toxicity classification. In Proceedings of the Neurips Workshop on Algorithmic Fairness through the Lens of Causality and Interpretability."},{"key":"e_1_3_2_3_2","first-page":"60","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Agarwal Alekh","year":"2018","unstructured":"Alekh Agarwal, Alina Beygelzimer, Miroslav Dud\u00edk, John Langford, and Hanna Wallach. 2018. A reductions approach to fair classification. In Proceedings of the International Conference on Machine Learning. PMLR, Stockholm, Sweden, 60\u201369."},{"key":"e_1_3_2_4_2","first-page":"32582","article-title":"Transferring fairness under distribution shifts via fair consistency regularization","volume":"35","author":"An Bang","year":"2022","unstructured":"Bang An, Zora Che, Mucong Ding, and Furong Huang. 2022. Transferring fairness under distribution shifts via fair consistency regularization. Adv. Neural Info. Process. Syst. 35 (2022), 32582\u201332597.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_5_2","first-page":"254","volume-title":"Ethics of Data and Analytics","author":"Angwin Julia","year":"2016","unstructured":"Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2016. Machine bias. In Ethics of Data and Analytics. Auerbach Publications, 254\u2013264."},{"key":"e_1_3_2_6_2","first-page":"1770","volume-title":"Proceedings of the International Conference on Artificial Intelligence and Statistics","author":"Awasthi Pranjal","year":"2020","unstructured":"Pranjal Awasthi, Matth\u00e4us Kleindessner, and Jamie Morgenstern. 2020. Equalized odds postprocessing under imperfect group information. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR, 1770\u20131780."},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1080\/15228835.2017.1416512"},{"key":"e_1_3_2_8_2","unstructured":"Yahav Bechavod and Katrina Ligett. 2017. Penalizing unfairness in binary classification. Retrieved from https:\/\/arXiv:1707.00044"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01270-0_28"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.1147\/JRD.2019.2942287"},{"key":"e_1_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462596"},{"key":"e_1_3_2_12_2","article-title":"Man is to computer programmer as woman is to homemaker? Debiasing word embeddings","volume":"29","author":"Bolukbasi Tolga","year":"2016","unstructured":"Tolga Bolukbasi, Kai-Wei Chang, James Y. Zou, Venkatesh Saligrama, and Adam T. Kalai. 2016. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. Adv. Neural Info. Process. Syst. 29 (2016).","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_13_2","first-page":"803","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Brunet Marc-Etienne","year":"2019","unstructured":"Marc-Etienne Brunet, Colleen Alkalay-Houlihan, Ashton Anderson, and Richard Zemel. 2019. Understanding the origins of bias in word embeddings. In Proceedings of the International Conference on Machine Learning. PMLR, 803\u2013811."},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10618-010-0190-x"},{"key":"e_1_3_2_15_2","doi-asserted-by":"publisher","DOI":"10.5555\/3294996.3295155"},{"key":"e_1_3_2_16_2","first-page":"1349","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Celis L. Elisa","year":"2021","unstructured":"L. Elisa Celis, Lingxiao Huang, Vijay Keswani, and Nisheeth K. Vishnoi. 2021. Fair classification with noisy protected attributes: A framework with provable guarantees. In Proceedings of the International Conference on Machine Learning. PMLR, 1349\u20131361."},{"key":"e_1_3_2_17_2","first-page":"8158","article-title":"Fair classification with adversarial perturbations","volume":"34","author":"Celis L. Elisa","year":"2021","unstructured":"L. Elisa Celis, Anay Mehrotra, and Nisheeth Vishnoi. 2021. Fair classification with adversarial perturbations. Adv. Neural Info. Process. Syst. 34 (2021), 8158\u20138171.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1146\/annurev-biodatasci-092820-033938"},{"key":"e_1_3_2_19_2","unstructured":"Richard J. Chen Tiffany Y. Chen Jana Lipkova Judy J. Wang Drew F. K. Williamson Ming Y. Lu Sharifa Sahai and Faisal Mahmood. 2021. Algorithm fairness in ai for medicine and healthcare. Retrieved from https:\/\/arXiv:2110.00603"},{"key":"e_1_3_2_20_2","first-page":"11266","article-title":"Fairness transferability subject to bounded distribution shift","volume":"35","author":"Chen Yatong","year":"2022","unstructured":"Yatong Chen, Reilly Raab, Jialu Wang, and Yang Liu. 2022. Fairness transferability subject to bounded distribution shift. Adv. Neural Info. Process. Syst. 35 (2022), 11266\u201311278.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_21_2","doi-asserted-by":"publisher","DOI":"10.1089\/big.2016.0047"},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1145\/3097983.3098095"},{"key":"e_1_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1145\/3306618.3314236"},{"key":"e_1_3_2_24_2","unstructured":"Elliot Creager J\u00f6rn-Henrik Jacobsen and Richard Zemel. 2020. Exchanging lessons between algorithmic fairness and domain generalization. arXiv preprint arXiv:2010.07249. https:\/\/arxiv.org\/abs\/2010.07249"},{"key":"e_1_3_2_25_2","volume-title":"Proceedings of the NeurIPS Workshop on Consequential Decision Making in Dynamic Environments","volume":"12","author":"Dai Jessica","year":"2020","unstructured":"Jessica Dai and Sarah M. Brown. 2020. Label bias, label shift: Fair machine learning with unreliable labels. In Proceedings of the NeurIPS Workshop on Consequential Decision Making in Dynamic Environments, Vol. 12."},{"key":"e_1_3_2_26_2","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372878"},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.1515\/popets-2015-0007"},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462523"},{"key":"e_1_3_2_29_2","volume-title":"folktables","author":"Ding Frances","year":"2021","unstructured":"Frances Ding, Moritz Hardt, John Miller, and Ludwig Schmidt. 2021. folktables. https:\/\/github.com\/zykls\/folktables"},{"key":"e_1_3_2_30_2","article-title":"Retiring adult: New datasets for fair machine learning","volume":"34","author":"Ding Frances","year":"2021","unstructured":"Frances Ding, Moritz Hardt, John Miller, and Ludwig Schmidt. 2021. Retiring adult: New datasets for fair machine learning. Adv. Neural Info. Process. Syst. 34 (2021).","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_31_2","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3482104"},{"key":"e_1_3_2_32_2","doi-asserted-by":"publisher","DOI":"10.1137\/1.9781611976700.21"},{"key":"e_1_3_2_33_2","doi-asserted-by":"publisher","DOI":"10.1145\/2090236.2090255"},{"key":"e_1_3_2_34_2","first-page":"119","volume-title":"Proceedings of the Conference on Fairness, Accountability and Transparency","author":"Dwork Cynthia","year":"2018","unstructured":"Cynthia Dwork, Nicole Immorlica, Adam Tauman Kalai, and Max Leiserson. 2018. Decoupled classifiers for group-fair and efficient machine learning. In Proceedings of the Conference on Fairness, Accountability and Transparency. PMLR, 119\u2013133."},{"key":"e_1_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.1145\/2783258.2783311"},{"key":"e_1_3_2_36_2","doi-asserted-by":"publisher","DOI":"10.1056\/NEJMc2104626"},{"key":"e_1_3_2_37_2","first-page":"2325","volume-title":"Proceedings of the International Conference on Artificial Intelligence and Statistics","author":"Fogliato Riccardo","year":"2020","unstructured":"Riccardo Fogliato, Alexandra Chouldechova, and Max G\u2019sell. 2020. Fairness evaluation in presence of biased noisy labels. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR, 2325\u20132336."},{"key":"e_1_3_2_38_2","doi-asserted-by":"publisher","DOI":"10.1137\/1.9781611976236.48"},{"key":"e_1_3_2_39_2","doi-asserted-by":"publisher","DOI":"10.1111\/jofi.13090"},{"key":"e_1_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1145\/3437963.3441824"},{"key":"e_1_3_2_41_2","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-00257-z"},{"key":"e_1_3_2_42_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Giguere Stephen","year":"2021","unstructured":"Stephen Giguere, Blossom Metevier, Bruno Castro da Silva, Yuriy Brun, Philip S. Thomas, and Scott Niekum. 2021. Fairness guarantees under demographic shift. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_43_2","first-page":"2415","volume-title":"Advances in Neural Information Processing Systems","author":"Goh Gabriel","year":"2016","unstructured":"Gabriel Goh, Andrew Cotter, Maya Gupta, and Michael P. Friedlander. 2016. Satisfying real-world goals with dataset constraints. In Advances in Neural Information Processing Systems. 2415\u20132423."},{"key":"e_1_3_2_44_2","first-page":"290","volume-title":"Proceedings of the Conference on Causal Learning and Reasoning","author":"Goldstein Mark","year":"2022","unstructured":"Mark Goldstein, J\u00f6rn-Henrik Jacobsen, Olina Chau, Adriel Saporta, Aahlad Manas Puli, Rajesh Ranganath, and Andrew Miller. 2022. Learning invariant representations with missing data. In Proceedings of the Conference on Causal Learning and Reasoning. PMLR, 290\u2013301."},{"key":"e_1_3_2_45_2","article-title":"Generative adversarial nets","volume":"27","author":"Goodfellow Ian","year":"2014","unstructured":"Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. Adv. Neural Info. Process. Syst. 27 (2014).","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_46_2","doi-asserted-by":"publisher","DOI":"10.5555\/2188385.2188410"},{"key":"e_1_3_2_47_2","first-page":"3315","article-title":"Equality of opportunity in supervised learning","volume":"29","author":"Hardt Moritz","year":"2016","unstructured":"Moritz Hardt, Eric Price, and Nati Srebro. 2016. Equality of opportunity in supervised learning. Adv. Neural Info. Process. Syst. 29 (2016), 3315\u20133323.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_48_2","doi-asserted-by":"publisher","DOI":"10.1145\/3178876.3186044"},{"key":"e_1_3_2_49_2","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287600"},{"key":"e_1_3_2_50_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-61527-7_11"},{"key":"e_1_3_2_51_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-27615-7_20"},{"key":"e_1_3_2_52_2","doi-asserted-by":"publisher","DOI":"10.1145\/3306618.3314267"},{"key":"e_1_3_2_53_2","first-page":"702","volume-title":"Proceedings of the International Conference on Artificial Intelligence and Statistics","author":"Jiang Heinrich","year":"2020","unstructured":"Heinrich Jiang and Ofir Nachum. 2020. Identifying and correcting label bias in machine learning. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR, 702\u2013712."},{"key":"e_1_3_2_54_2","doi-asserted-by":"publisher","DOI":"10.1109\/ISIT50566.2022.9834475"},{"key":"e_1_3_2_55_2","doi-asserted-by":"publisher","DOI":"10.1038\/sdata.2016.35"},{"key":"e_1_3_2_56_2","first-page":"2439","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Kallus Nathan","year":"2018","unstructured":"Nathan Kallus and Angela Zhou. 2018. Residual unfairness in fair machine learning from prejudiced data. In Proceedings of the International Conference on Machine Learning. PMLR, 2439\u20132448."},{"key":"e_1_3_2_57_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-011-0463-8"},{"key":"e_1_3_2_58_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-33486-3_3"},{"key":"e_1_3_2_59_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58574-7_34"},{"key":"e_1_3_2_60_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbankfin.2010.06.001"},{"key":"e_1_3_2_61_2","first-page":"5637","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Koh Pang Wei","year":"2021","unstructured":"Pang Wei Koh, Shiori Sagawa, Henrik Marklund, Sang Michael Xie, Marvin Zhang, Akshay Balsubramani, Weihua Hu, Michihiro Yasunaga, Richard Lanas Phillips, Irena Gao et\u00a0al. 2021. Wilds: A benchmark of in-the-wild distribution shifts. In Proceedings of the International Conference on Machine Learning. PMLR, 5637\u20135664."},{"key":"e_1_3_2_62_2","article-title":"Fairness-aware PAC learning from corrupted data","volume":"23","author":"Konstantinov Nikola H.","year":"2022","unstructured":"Nikola H. Konstantinov and Christoph Lampert. 2022. Fairness-aware PAC learning from corrupted data. J. Mach. Learn. Res. 23 (2022), 1\u201360.","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_2_63_2","volume-title":"Fairness: Algorithmic Fairness Metrics","author":"Kozodoi Nikita","year":"2021","unstructured":"Nikita Kozodoi and Tibor V. Varga. 2021. Fairness: Algorithmic Fairness Metrics. Retrieved from https:\/\/CRAN.R-project.org\/package=fairnessR package version 1.2.1."},{"key":"e_1_3_2_64_2","article-title":"Noise-tolerant fair classification","volume":"32","author":"Lamy Alex","year":"2019","unstructured":"Alex Lamy, Ziyuan Zhong, Aditya K. Menon, and Nakul Verma. 2019. Noise-tolerant fair classification. Adv. Neural Info. Process. Syst. 32 (2019).","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_65_2","volume-title":"Proceedings of the Workshop on Fairness, Accountability and Transparency in Machine Learning (FAT\/ML\u201917).","author":"Lan Chao","year":"2017","unstructured":"Chao Lan and Jun Huan. 2017. Discriminatory transfer. In Proceedings of the Workshop on Fairness, Accountability and Transparency in Machine Learning (FAT\/ML\u201917)."},{"key":"e_1_3_2_66_2","article-title":"Does mitigating ML\u2019s impact disparity require treatment disparity?","volume":"31","author":"Lipton Zachary","year":"2018","unstructured":"Zachary Lipton, Julian McAuley, and Alexandra Chouldechova. 2018. Does mitigating ML\u2019s impact disparity require treatment disparity? Adv. Neural Info. Process. Syst. 31 (2018).","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_67_2","first-page":"3150","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Liu Lydia T.","year":"2018","unstructured":"Lydia T. Liu, Sarah Dean, Esther Rolf, Max Simchowitz, and Moritz Hardt. 2018. Delayed impact of fair machine learning. In Proceedings of the International Conference on Machine Learning. PMLR, 3150\u20133158."},{"key":"e_1_3_2_68_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Louizos Christos","year":"2016","unstructured":"Christos Louizos, Kevin Swersky, Yujia Li, Max Welling, and Richard Zemel. 2016. The variational fair autoencoder. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_69_2","first-page":"3384","volume-title":"Proceedings of the 35th International Conference on Machine Learning","volume":"80","author":"Madras David","year":"2018","unstructured":"David Madras, Elliot Creager, Toniann Pitassi, and Richard Zemel. 2018. Learning adversarially fair and transferable representations. In Proceedings of the 35th International Conference on Machine Learning, Vol. 80. PMLR, Stockholm, Sweden, 3384\u20133393."},{"key":"e_1_3_2_70_2","first-page":"25773","article-title":"Does enforcing fairness mitigate biases caused by subpopulation shift?","volume":"34","author":"Maity Subha","year":"2021","unstructured":"Subha Maity, Debarghya Mukherjee, Mikhail Yurochkin, and Yuekai Sun. 2021. Does enforcing fairness mitigate biases caused by subpopulation shift? Adv. Neural Info. Process. Syst. 34 (2021), 25773\u201325784.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_71_2","first-page":"739","volume-title":"Proceedings of the International Conference on Artificial Intelligence and Statistics","author":"Makar Maggie","year":"2022","unstructured":"Maggie Makar, Ben Packer, Dan Moldovan, Davis Blalock, Yoni Halpern, and Alexander D\u2019Amour. 2022. Causally motivated shortcut removal using auxiliary labels. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR, 739\u2013766."},{"key":"e_1_3_2_72_2","first-page":"18445","article-title":"Ensuring fairness beyond the training data","volume":"33","author":"Mandal Debmalya","year":"2020","unstructured":"Debmalya Mandal, Samuel Deng, Suman Jana, Jeannette Wing, and Daniel J. Hsu. 2020. Ensuring fairness beyond the training data. Adv. Neural Info. Process. Syst. 33 (2020), 18445\u201318456.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_73_2","first-page":"1273","volume-title":"Proceedings of the International Conference on Artificial Intelligence and Statistics","author":"McMahan Brendan","year":"2017","unstructured":"Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas. 2017. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR, 1273\u20131282."},{"key":"e_1_3_2_74_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i10.17080"},{"key":"e_1_3_2_75_2","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445887"},{"key":"e_1_3_2_76_2","first-page":"107","volume-title":"Proceedings of the Conference on Fairness, Accountability and Transparency","author":"Menon Aditya Krishna","year":"2018","unstructured":"Aditya Krishna Menon and Robert C. Williamson. 2018. The cost of fairness in binary classification. In Proceedings of the Conference on Fairness, Accountability and Transparency. PMLR, 107\u2013118."},{"issue":"99","key":"e_1_3_2_77_2","first-page":"1","article-title":"Joint causal inference from multiple contexts","volume":"21","author":"Mooij Joris M.","year":"2020","unstructured":"Joris M. Mooij, Sara Magliacane, and Tom Claassen. 2020. Joint causal inference from multiple contexts. J. Mach. Learn. Res. 21, 99 (2020), 1\u2013108.","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_2_78_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2011.06.019"},{"key":"e_1_3_2_79_2","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287599"},{"key":"e_1_3_2_80_2","first-page":"28902","article-title":"Domain adaptation meets individual fairness. And they get along.","volume":"35","author":"Mukherjee Debarghya","year":"2022","unstructured":"Debarghya Mukherjee, Felix Petersen, Mikhail Yurochkin, and Yuekai Sun. 2022. Domain adaptation meets individual fairness. And they get along. Adv. Neural Info. Process. Syst. 35 (2022), 28902\u201328913.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_81_2","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445910"},{"key":"e_1_3_2_82_2","first-page":"922","volume-title":"Proceedings of the International Conference on Artificial Intelligence and Statistics","author":"Olmin Amanda","year":"2022","unstructured":"Amanda Olmin and Fredrik Lindsten. 2022. Robustness and reliability when training with noisy labels. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR, 922\u2013942."},{"key":"e_1_3_2_83_2","doi-asserted-by":"publisher","DOI":"10.1109\/DSAA49011.2020.00015"},{"key":"e_1_3_2_84_2","doi-asserted-by":"publisher","DOI":"10.1145\/3494672"},{"key":"e_1_3_2_85_2","doi-asserted-by":"publisher","DOI":"10.1145\/2783258.2788629"},{"key":"e_1_3_2_86_2","unstructured":"Aahlad Puli Lily H. Zhang Eric K. Oermann and Rajesh Ranganath. 2021. Out-of-distribution generalization in the presence of nuisance-induced spurious correlations. Retrieved from https:\/\/arXiv:2107.00520"},{"key":"e_1_3_2_87_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2018.04.007"},{"key":"e_1_3_2_88_2","first-page":"677","volume-title":"Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems","author":"Quadrianto Novi","year":"2017","unstructured":"Novi Quadrianto and Viktoriia Sharmanska. 2017. Recycling privileged learning and distribution matching for fairness. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems, Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (Eds.). 677\u2013688."},{"key":"e_1_3_2_89_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00842"},{"key":"e_1_3_2_90_2","doi-asserted-by":"publisher","DOI":"10.2307\/j.ctv31xf5v0"},{"key":"e_1_3_2_91_2","volume-title":"A Theory of Justice","author":"Rawls John","year":"2009","unstructured":"John Rawls. 2009. A Theory of Justice. Harvard University Press, Cambridge, MA."},{"key":"e_1_3_2_92_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.6002"},{"key":"e_1_3_2_93_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i11.17135"},{"key":"e_1_3_2_94_2","first-page":"8147","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Roh Yuji","year":"2020","unstructured":"Yuji Roh, Kangwook Lee, Steven Whang, and Changho Suh. 2020. FR-Train: A mutual information-based approach to fair and robust training. In Proceedings of the International Conference on Machine Learning. PMLR, 8147\u20138157."},{"key":"e_1_3_2_95_2","unstructured":"Pedro Saleiro Benedict Kuester Loren Hinkson Jesse London Abby Stevens Ari Anisfeld Kit T. Rodolfa and Rayid Ghani. 2018. Aequitas: A bias and fairness audit toolkit. Retrieved from https:\/\/arXiv:1811.05577"},{"key":"e_1_3_2_96_2","first-page":"19304","article-title":"Diagnosing failures of fairness transfer across distribution shift in real-world medical settings","volume":"35","author":"Schrouff Jessica","year":"2022","unstructured":"Jessica Schrouff, Natalie Harris, Sanmi Koyejo, Ibrahim M. Alabdulmohsin, Eva Schnider, Krista Opsahl-Ong, Alexander Brown, Subhrajit Roy, Diana Mincu, Christina Chen et\u00a0al. 2022. Diagnosing failures of fairness transfer across distribution shift in real-world medical settings. Adv. Neural Info. Process. Syst. 35 (2022), 19304\u201319318.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_97_2","unstructured":"Candice Schumann Xuezhi Wang Alex Beutel Jilin Chen Hai Qian and Ed H. Chi. 2019. Transfer of machine learning fairness across domains. Retrieved from https:\/\/arXiv:1906.09688"},{"key":"e_1_3_2_98_2","unstructured":"Marco Scutari. 2023. fairml: A statistician\u2019s take on fair machine learning modelling. Retrieved from https:\/\/arXiv:2305.02009"},{"key":"e_1_3_2_99_2","doi-asserted-by":"publisher","DOI":"10.1142\/9789811232701_0022"},{"key":"e_1_3_2_100_2","volume-title":"Proceedings of the NIPS Workshop on Machine Learning for the Developing World","author":"Shankar Shreya","year":"2017","unstructured":"Shreya Shankar, Yoni Halpern, Eric Breck, James Atwood, Jimbo Wilson, and D. Sculley. 2017. No classification without representation: Assessing geodiversity issues in open data sets for the developing world. In Proceedings of the NIPS Workshop on Machine Learning for the Developing World."},{"key":"e_1_3_2_101_2","doi-asserted-by":"publisher","unstructured":"Yuxin Shi Han Yu and Cyril Leung. 2023. Towards fairness-aware federated learning. IEEE Transactions on Neural Networks and Learning Systems (2023) 1\u201317. DOI:10.1109\/tnnls.2023.3263594","DOI":"10.1109\/tnnls.2023.3263594"},{"key":"e_1_3_2_102_2","volume-title":"Proceedings of the NeurIPS Workshop on Fair ML for Health","author":"Singh Harvineet","year":"2019","unstructured":"Harvineet Singh, Rina Singh, Vishwali Mhasawade, and Rumi Chunara. 2019. Fair predictors under distribution shift. In Proceedings of the NeurIPS Workshop on Fair ML for Health. Curran Associates."},{"key":"e_1_3_2_103_2","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445865"},{"key":"e_1_3_2_104_2","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372839"},{"key":"e_1_3_2_105_2","first-page":"596","article-title":"Fixmatch: Simplifying semi-supervised learning with consistency and confidence","volume":"33","author":"Sohn Kihyuk","year":"2020","unstructured":"Kihyuk Sohn, David Berthelot, Nicholas Carlini, Zizhao Zhang, Han Zhang, Colin A. Raffel, Ekin Dogus Cubuk, Alexey Kurakin, and Chun-Liang Li. 2020. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. Adv. Neural Info. Process. Syst. 33 (2020), 596\u2013608.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_106_2","unstructured":"Bahar Taskesen Viet Anh Nguyen Daniel Kuhn and Jose Blanchet. 2020. A distributionally robust approach to fair classification. Retrieved from https:\/\/arXiv:2007.09530"},{"key":"e_1_3_2_107_2","doi-asserted-by":"publisher","DOI":"10.1126\/science.aag3311"},{"key":"e_1_3_2_108_2","first-page":"6373","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Ustun Berk","year":"2019","unstructured":"Berk Ustun, Yang Liu, and David Parkes. 2019. Fairness without harm: Decoupled classifiers with preference guarantees. In Proceedings of the International Conference on Machine Learning. PMLR, 6373\u20136382."},{"key":"e_1_3_2_109_2","first-page":"16196","article-title":"Counterfactual invariance to spurious correlations in text classification","volume":"34","author":"Veitch Victor","year":"2021","unstructured":"Victor Veitch, Alexander D\u2019Amour, Steve Yadlowsky, and Jacob Eisenstein. 2021. Counterfactual invariance to spurious correlations in text classification. Adv. Neural Info. Process. Syst. 34 (2021), 16196\u201316208.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_110_2","doi-asserted-by":"publisher","DOI":"10.1145\/3194770.3194776"},{"key":"e_1_3_2_111_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.clsr.2021.105567"},{"key":"e_1_3_2_112_2","article-title":"How robust is your fairness? Evaluating and sustaining fairness under unseen distribution shifts","author":"Wang Haotao","year":"2023","unstructured":"Haotao Wang, Junyuan Hong, Jiayu Zhou, and Zhangyang Wang. 2023. How robust is your fairness? Evaluating and sustaining fairness under unseen distribution shifts. Trans. Mach. Learn. Res. (2023). Retrieved from https:\/\/openreview.net\/forum?id=11pGlecTz2","journal-title":"Trans. Mach. Learn. Res."},{"key":"e_1_3_2_113_2","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445915"},{"key":"e_1_3_2_114_2","first-page":"5190","article-title":"Robust optimization for fairness with noisy protected groups","volume":"33","author":"Wang Serena","year":"2020","unstructured":"Serena Wang, Wenshuo Guo, Harikrishna Narasimhan, Andrew Cotter, Maya Gupta, and Michael Jordan. 2020. Robust optimization for fairness with noisy protected groups. Adv. Neural Info. Process. Syst. 33 (2020), 5190\u20135203.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_115_2","doi-asserted-by":"publisher","DOI":"10.1145\/3368555.3384469"},{"key":"e_1_3_2_116_2","unstructured":"Hilde Weerts Miroslav Dud\u00edk Richard Edgar Adrin Jalali Roman Lutz and Michael Madaio. 2023. Fairlearn: Assessing and improving fairness of AI systems. Retrieved from https:\/\/arXiv:2303.16626"},{"key":"e_1_3_2_117_2","doi-asserted-by":"publisher","DOI":"10.1136\/amiajnl-2013-002162"},{"key":"e_1_3_2_118_2","doi-asserted-by":"crossref","unstructured":"Jakub Wi\u015bniewski and Przemys\u0142aw Biecek. 2021. fairmodels: A flexible tool for bias detection visualization and mitigation. Retrieved from https:\/\/arXiv:2104.00507","DOI":"10.32614\/CRAN.package.fairmodels"},{"key":"e_1_3_2_119_2","first-page":"1920","volume-title":"Proceedings of the Conference on Learning Theory","author":"Woodworth Blake","year":"2017","unstructured":"Blake Woodworth, Suriya Gunasekar, Mesrob I. Ohannessian, and Nathan Srebro. 2017. Learning non-discriminatory predictors. In Proceedings of the Conference on Learning Theory. PMLR, 1920\u20131953."},{"key":"e_1_3_2_120_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3005987"},{"key":"e_1_3_2_121_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Yurochkin Mikhail","year":"2020","unstructured":"Mikhail Yurochkin, Amanda Bower, and Yuekai Sun. 2020. Training individually fair ML models with sensitive subspace robustness. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_122_2","doi-asserted-by":"publisher","DOI":"10.1145\/3038912.3052660"},{"key":"e_1_3_2_123_2","first-page":"962","volume-title":"Proceedings of the International Conference on Artificial Intelligence and Statistics","author":"Zafar Muhammad Bilal","year":"2017","unstructured":"Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rogriguez, and Krishna P. Gummadi. 2017. Fairness constraints: Mechanisms for fair classification. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR, 962\u2013970."},{"key":"e_1_3_2_124_2","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pmed.1002683"},{"key":"e_1_3_2_125_2","first-page":"325","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Zemel Rich","year":"2013","unstructured":"Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. 2013. Learning fair representations. In Proceedings of the International Conference on Machine Learning. PMLR, 325\u2013333."},{"key":"e_1_3_2_126_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-61527-7_12"},{"key":"e_1_3_2_127_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-75765-6_20"},{"key":"e_1_3_2_128_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICTAI50040.2020.00069"},{"key":"e_1_3_2_129_2","first-page":"18457","article-title":"How do fair decisions fare in long-term qualification?","volume":"33","author":"Zhang Xueru","year":"2020","unstructured":"Xueru Zhang, Ruibo Tu, Yang Liu, Mingyan Liu, Hedvig Kjellstrom, Kun Zhang, and Cheng Zhang. 2020. How do fair decisions fare in long-term qualification? Adv. Neural Info. Process. Syst. 33 (2020), 18457\u201318469.","journal-title":"Adv. Neural Info. Process. Syst."},{"key":"e_1_3_2_130_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.106994"},{"key":"e_1_3_2_131_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2019.00102"},{"key":"e_1_3_2_132_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICBK50248.2020.00029"},{"key":"e_1_3_2_133_2","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467389"},{"key":"e_1_3_2_134_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM50108.2020.00091"},{"key":"e_1_3_2_135_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICBK50248.2020.00047"},{"key":"e_1_3_2_136_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1521"},{"key":"e_1_3_2_137_2","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3470814"},{"key":"e_1_3_2_138_2","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2020.3004555"}],"container-title":["ACM Computing Surveys"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3637438","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3637438","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T23:43:44Z","timestamp":1750290224000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3637438"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,10]]},"references-count":137,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,6,30]]}},"alternative-id":["10.1145\/3637438"],"URL":"https:\/\/doi.org\/10.1145\/3637438","relation":{},"ISSN":["0360-0300","1557-7341"],"issn-type":[{"value":"0360-0300","type":"print"},{"value":"1557-7341","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2,10]]},"assertion":[{"value":"2022-10-10","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-12-01","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-02-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}