skip to main content
10.1145/3178876.3186133acmotherconferencesArticle/Chapter ViewAbstractPublication PagesthewebconfConference Proceedingsconference-collections
research-article
Free access

Adaptive Sensitive Reweighting to Mitigate Bias in Fairness-aware Classification

Published: 23 April 2018 Publication History

Abstract

Machine learning bias and fairness have recently emerged as key issues due to the pervasive deployment of data-driven decision making in a variety of sectors and services. It has often been argued that unfair classifications can be attributed to bias in training data, but previous attempts to 'repair' training data have led to limited success. To circumvent shortcomings prevalent in data repairing approaches, such as those that weight training samples of the sensitive group (e.g. gender, race, financial status) based on their misclassification error, we present a process that iteratively adapts training sample weights with a theoretically grounded model. This model addresses different kinds of bias to better achieve fairness objectives, such as trade-offs between accuracy and disparate impact elimination or disparate mistreatment elimination. We show that, compared to previous fairness-aware approaches, our methodology achieves better or similar trades-offs between accuracy and unfairness mitigation on real-world and synthetic datasets.

References

[1]
Solon Barocas and Andrew D Selbst. 2016. Big data's disparate impact. (2016).
[2]
Dan Biddle. 2006. Adverse impact and test validation: A practitioner's guide to valid and defensible employment testing. Gower Publishing, Ltd.
[3]
Toon Calders, Faisal Kamiran, and Mykola Pechenizkiy. 2009. Building classifiers with independency constraints Data mining workshops, 2009. ICDMW'09. IEEE international conference on. IEEE, 13--18.
[4]
Toon Calders, Asim Karim, Faisal Kamiran, Wasif Ali, and Xiangliang Zhang. 2013. Controlling attribute effect in linear regression. Data Mining (ICDM), 2013 IEEE 13th International Conference on. IEEE, 71--80.
[5]
Toon Calders and Sicco Verwer. 2010. Three naive Bayes approaches for discrimination-free classification. Data Mining and Knowledge Discovery Vol. 21, 2 (2010), 277--292.
[6]
L Elisa Celis, Damian Straszak, and Nisheeth K Vishnoi. 2017. Ranking with Fairness Constraints. arXiv preprint arXiv:1704.06840 (2017).
[7]
Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. arXiv preprint arXiv:1703.00056 (2017).
[8]
Sam Corbett-Davies, Emma Pierson, Avi Feller, Sharad Goel, and Aziz Huq. 2017. Algorithmic decision making and the cost of fairness. arXiv preprint arXiv:1701.08230 (2017).
[9]
Georges Dionne and Casey Rothschild. 2014. Economic effects of risk classification bans. The Geneva Risk and Insurance Review Vol. 39, 2 (2014), 184--221.
[10]
Neil A Doherty, Anastasia V Kartasheva, and Richard D Phillips. 2012. Information effect of entry into credit ratings market: The case of insurers' ratings. Journal of Financial Economics Vol. 106, 2 (2012), 308--330.
[11]
Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness through awareness. In Proceedings of the 3rd Innovations in Theoretical Computer Science Conference. ACM, 214--226.
[12]
Michael Feldman. 2015. Computational Fairness: Preventing Machine-Learned Discrimination. (2015).
[13]
Michael Feldman, Sorelle A Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. 2015. Certifying and removing disparate impact. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 259--268.
[14]
Daniel E Finkel. 2003. DIRECT optimization algorithm user guide. Center for Research in Scientific Computation, North Carolina State University Vol. 2 (2003).
[15]
Daniel E Finkel and CT Kelley. 2004. Convergence analysis of the DIRECT algorithm. Optimization Online, Vol. 14, 2 (2004), 1--10.
[16]
Benjamin Fish, Jeremy Kun, and Adám D Lelkes. 2015. Fair boosting: a case study. In Workshop on Fairness, Accountability, and Transparency in Machine Learning.
[17]
Benjamin Fish, Jeremy Kun, and Ádám D Lelkes. 2016. A confidence-based approach for balancing fairness and accuracy Proceedings of the 2016 SIAM International Conference on Data Mining. SIAM, 144--152.
[18]
Kazuto Fukuchi and Jun Sakuma. 2015. Fairness-Aware Learning with Restriction of Universal Dependency using f-Divergences. arXiv preprint arXiv:1506.07721 (2015).
[19]
Gabriel Goh, Andrew Cotter, Maya Gupta, and Michael P Friedlander. 2016. Satisfying Real-world Goals with Dataset Constraints Advances in Neural Information Processing Systems. 2415--2423.
[20]
Moritz Hardt, Eric Price, Nati Srebro, and others. 2016. Equality of opportunity in supervised learning. In Advances in Neural Information Processing Systems. 3315--3323.
[21]
Qinghua Hu, Pengfei Zhu, Yongbin Yang, and Daren Yu. 2011. Large-margin nearest neighbor classifiers via sample weight learning. Neurocomputing, Vol. 74, 4 (2011), 656--660.
[22]
Anatoli Iouditski and Yuri Nesterov. 2014. Primal-dual subgradient methods for minimizing uniformly convex functions. arXiv preprint arXiv:1401.1792 (2014).
[23]
Donald R Jones, Cary D Perttunen, and Bruce E Stuckman. 1993. Lipschitzian optimization without the Lipschitz constant. Journal of Optimization Theory and Applications, Vol. 79, 1 (1993), 157--181.
[24]
Faisal Kamiran and Toon Calders. 2009. Classifying without discriminating. In Computer, Control and Communication, 2009. IC4 2009. 2nd International Conference on. IEEE, 1--6.
[25]
Faisal Kamiran and Toon Calders. 2012. Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems Vol. 33, 1 (2012), 1--33.
[26]
Faisal Kamiran, Toon Calders, and others. 2011. Handling conditional discrimination. In Proc. of the 11th IEEE Int'l Conf. on Data Mining.
[27]
Faisal Kamiran, Toon Calders, and Mykola Pechenizkiy. 2010. Discrimination aware decision tree learning. In 2010 IEEE International Conference on Data Mining. IEEE, 869--874.
[28]
Toshihiro Kamishima, Shotaro Akaho, Hideki Asoh, and Jun Sakuma. 2012. Fairness-aware classifier with prejudice remover regularizer Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 35--50.
[29]
Toshihiro Kamishima, Shotaro Akaho, and Jun Sakuma. 2011. Fairness-aware learning through regularization approach Data Mining Workshops (ICDMW), 2011 IEEE 11th International Conference on. IEEE, 643--650.
[30]
Michael Kearns, Seth Neel, Aaron Roth, and Zhiwei Steven Wu. 2017. Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup Fairness. arXiv preprint arXiv:1711.05144 (2017).
[31]
Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan. 2016. Inherent trade-offs in the fair determination of risk scores. arXiv preprint arXiv:1609.05807 (2016).
[32]
Ron Kohavi. 1996. Scaling Up the Accuracy of Naive-Bayes Classifiers: A Decision-Tree Hybrid. KDD, Vol. Vol. 96. Citeseer, 202--207.
[33]
J. Larson, S. Mattu, L. Kirchner, and J. Angwin. 2017. COMPAS dataset. (2017). https://github.com/propublica/compas-analysis
[34]
Yuan Li, Chang Huang, and Ram Nevatia. 2009. Learning to associate: Hybridboosted multi-target tracker for crowded scene Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on. IEEE, 2953--2960.
[35]
M. Lichman. 2013. UCI Machine Learning Repository. (2013). http://archive.ics.uci.edu/ml
[36]
Enrico Miglierina and Elena Molho. 2002. Scalarization and stability in vector optimization. Journal of Optimization Theory and Applications, Vol. 114, 3 (2002), 657--670.
[37]
Sergio Moro, Raul Laureano, and Paulo Cortez. 2011. Using data mining for bank direct marketing: An application of the crisp-dm methodology Proceedings of European Simulation and Modelling Conference-ESM'2011. Eurosis, 117--121.
[38]
Shelly L Peffer. 2009. Title VII and disparate-treatment discrimination versus disparate-impact discrimination: The Supreme Court's decision in Ricci v. DeStefano. Review of Public Personnel Administration Vol. 29, 4 (2009), 402--410.
[39]
Andrea Romei and Salvatore Ruggieri. 2014. A multidisciplinary survey on discrimination analysis. The Knowledge Engineering Review Vol. 29, 05 (2014), 582--638.
[40]
Robert E Schapire. 2003. The boosting approach to machine learning: An overview. Nonlinear estimation and classification. Springer, 149--171.
[41]
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P. Gummadi. 2017. Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification Without Disparate Mistreatment. In Proceedings of the 26th International Conference on World Wide Web (WWW '17). International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, Switzerland, 1171--1180.
[42]
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P Gummadi. 2015 a. Fairness constraints: A mechanism for fair classification. stat Vol. 1050 (2015), 19.
[43]
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P Gummadi. 2015 b. Learning Fair Classifiers. stat Vol. 1050 (2015), 29.
[44]
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rogriguez, and Krishna P Gummadi. 2017. Fairness Constraints: Mechanisms for Fair Classification Artificial Intelligence and Statistics. 962--970.

Cited By

View all
  • (2025)FLAC: Fairness-Aware Representation Learning by Suppressing Attribute-Class AssociationsIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2024.348725447:2(1148-1160)Online publication date: Feb-2025
  • (2025)Fairness is essential for robustness: fair adversarial training by identifying and augmenting hard examplesFrontiers of Computer Science: Selected Publications from Chinese Universities10.1007/s11704-024-3587-119:3Online publication date: 1-Mar-2025
  • (2025)Fairness in Optimization and ML: A Survey Part 2Dynamics of Information Systems10.1007/978-3-031-81010-7_17(273-290)Online publication date: 26-Feb-2025
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
WWW '18: Proceedings of the 2018 World Wide Web Conference
April 2018
2000 pages
ISBN:9781450356398
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

  • IW3C2: International World Wide Web Conference Committee

In-Cooperation

Publisher

International World Wide Web Conferences Steering Committee

Republic and Canton of Geneva, Switzerland

Publication History

Published: 23 April 2018

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. algorithm bias
  2. classification fairness
  3. reweighting

Qualifiers

  • Research-article

Funding Sources

  • European Commission

Conference

WWW '18
Sponsor:
  • IW3C2
WWW '18: The Web Conference 2018
April 23 - 27, 2018
Lyon, France

Acceptance Rates

WWW '18 Paper Acceptance Rate 170 of 1,155 submissions, 15%;
Overall Acceptance Rate 1,899 of 8,196 submissions, 23%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)1,097
  • Downloads (Last 6 weeks)69
Reflects downloads up to 01 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2025)FLAC: Fairness-Aware Representation Learning by Suppressing Attribute-Class AssociationsIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2024.348725447:2(1148-1160)Online publication date: Feb-2025
  • (2025)Fairness is essential for robustness: fair adversarial training by identifying and augmenting hard examplesFrontiers of Computer Science: Selected Publications from Chinese Universities10.1007/s11704-024-3587-119:3Online publication date: 1-Mar-2025
  • (2025)Fairness in Optimization and ML: A Survey Part 2Dynamics of Information Systems10.1007/978-3-031-81010-7_17(273-290)Online publication date: 26-Feb-2025
  • (2025)Towards Fair Face Verification: An In-depth Analysis of Demographic BiasesMachine Learning and Principles and Practice of Knowledge Discovery in Databases10.1007/978-3-031-74630-7_14(194-208)Online publication date: 8-Feb-2025
  • (2024)Investigating and Mitigating the Performance–Fairness Tradeoff via Protected-Category SamplingElectronics10.3390/electronics1315302413:15(3024)Online publication date: 31-Jul-2024
  • (2024)Strategies to improve fairness in artificial intelligence:A systematic literature reviewEducation for Information10.3233/EFI-240045(1-24)Online publication date: 25-Jul-2024
  • (2024)A Granular Approach to Optimal and Fair Patient Placement in Hospital Emergency DepartmentsProduction and Operations Management10.1177/10591478241240390Online publication date: 16-Apr-2024
  • (2024)Trustworthy AI Using Confidential Federated LearningCommunications of the ACM10.1145/367739067:9(48-53)Online publication date: 15-Aug-2024
  • (2024)Boosting Fair Classifier Generalization through Adaptive Priority ReweighingACM Transactions on Knowledge Discovery from Data10.1145/366589519:2(1-26)Online publication date: 23-May-2024
  • (2024)Trustworthy AI using Confidential Federated LearningQueue10.1145/366522022:2(87-107)Online publication date: 24-May-2024
  • Show More Cited By

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Login options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media