Skip to content

A Python package to assess and improve fairness of machine learning models.

License

Notifications You must be signed in to change notification settings

nnpereira/fairlearn

This branch is 1 commit behind fairlearn/fairlearn:main.

Folders and files

NameName
Last commit message
Last commit date
Oct 23, 2024
Jan 28, 2025
Feb 26, 2025
Feb 20, 2025
Mar 20, 2025
Jan 24, 2025
Mar 20, 2025
Feb 24, 2025
Jun 4, 2024
Nov 5, 2024
Jun 4, 2024
Sep 20, 2023
Nov 7, 2022
Nov 7, 2022
Feb 26, 2025
Mar 9, 2023
Jan 11, 2025
Dec 10, 2024
Oct 23, 2024
Jan 28, 2025

Repository files navigation

MIT license PyPI Discord StackOverflow

Fairlearn

Fairlearn is a Python package that empowers developers of artificial intelligence (AI) systems to assess their system's fairness and mitigate any observed unfairness issues. Fairlearn contains mitigation algorithms as well as metrics for model assessment. Besides the source code, this repository also contains Jupyter notebooks with examples of Fairlearn usage.

Website: https://fairlearn.org/

What we mean by fairness

An AI system can behave unfairly for a variety of reasons. In Fairlearn, we define whether an AI system is behaving unfairly in terms of its impact on people – i.e., in terms of harms. We focus on two kinds of harms:

  • Allocation harms. These harms can occur when AI systems extend or withhold opportunities, resources, or information. Some of the key applications are in hiring, school admissions, and lending.
  • Quality-of-service harms. Quality of service refers to whether a system works as well for one person as it does for another, even if no opportunities, resources, or information are extended or withheld.

We follow the approach known as group fairness, which asks: Which groups of individuals are at risk for experiencing harms? The relevant groups need to be specified by the data scientist and are application specific.

Group fairness is formalized by a set of constraints, which require that some aspect (or aspects) of the AI system's behavior be comparable across the groups. The Fairlearn package enables assessment and mitigation of unfairness under several common definitions. To learn more about our definitions of fairness, please visit our user guide on Fairness of AI Systems.

Note: Fairness is fundamentally a sociotechnical challenge. Many aspects of fairness, such as justice and due process, are not captured by quantitative fairness metrics. Furthermore, there are many quantitative fairness metrics which cannot all be satisfied simultaneously. Our goal is to enable humans to assess different mitigation strategies and then make trade-offs appropriate to their scenario.

Overview of Fairlearn

The Fairlearn Python package has two components:

  • Metrics for assessing which groups are negatively impacted by a model, and for comparing multiple models in terms of various fairness and accuracy metrics.
  • Algorithms for mitigating unfairness in a variety of AI tasks and along a variety of fairness definitions.

Fairlearn metrics

Check out our in-depth guide on the Fairlearn metrics.

Fairlearn algorithms

For an overview of our algorithms please refer to our website.

Getting Started with Fairlearn

First steps

For Users & Practitioners

Please note that notebooks downloaded from main may not be compatible with pip-installed versions.

For Contributors

Maintainers

A list of current maintainers is on our website.

Code of conduct

Fairlearn follows the Fairlearn Organization's Code of Conduct.

Issues

Regular (non-security) issues

Issues are meant for bugs, feature requests, and documentation improvements. Please submit a report through GitHub issues. A maintainer will respond promptly as appropriate.

Maintainers will try to link duplicate issues when possible.

Reporting security issues

To report security issues please send an email to [email protected].

About

A Python package to assess and improve fairness of machine learning models.

Resources

License

Citation

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 99.5%
  • Other 0.5%