A comprehensive set of fairness metrics for datasets and machine learning models, explanations for these metrics, and algorithms to mitigate bias in datasets and models.
-
Updated
May 6, 2024 - Python
A comprehensive set of fairness metrics for datasets and machine learning models, explanations for these metrics, and algorithms to mitigate bias in datasets and models.
A Python package to assess and improve fairness of machine learning models.
Responsible AI Toolbox is a suite of tools providing model and data exploration and assessment user interfaces and libraries that enable a better understanding of AI systems. These interfaces and libraries empower developers and stakeholders of AI systems to develop and monitor AI more responsibly, and take better data-driven actions.
🐢 Open-Source Evaluation & Testing for LLMs and ML models
Fairness Aware Machine Learning. Bias detection and mitigation for datasets and models.
The LinkedIn Fairness Toolkit (LiFT) is a Scala/Spark library that enables the measurement of fairness in large scale machine learning workflows.
Toolkit for Auditing and Mitigating Bias and Fairness of Machine Learning Systems 🔎🤖🧰
Talks & Workshops by the CODAIT team
[ACL 2020] Towards Debiasing Sentence Representations
WEFE: The Word Embeddings Fairness Evaluation Framework. WEFE is a framework that standardizes the bias measurement and mitigation in Word Embeddings models. Please feel welcome to open an issue in case you have any questions or a pull request if you want to contribute to the project!
Deploy a Custom Machine Learning engine and Monitor Payload Logging and Fairness using AI OpenScale
FairPut - Machine Learning Fairness Framework with LightGBM — Explainability, Robustness, Fairness (by @firmai)
SIREN: A Simulation Framework for Understanding the Effects of Recommender Systems in Online News Environments
Slides, videos and other potentially useful artifacts from various presentations on responsible machine learning.
Data and Model-based approaches for Mitigating Bias in Machine Learning Applications
This GitHub repository offers resources to create fair and unbiased AI systems, including libraries, tools and tutorials on identifying and mitigating bias in machine learning models and implementing fairness in AI.
R package for computing and visualizing fair ML metrics
[ICML 2021] Towards Understanding and Mitigating Social Biases in Language Models
A tool for gender bias identification in text. Part of Microsoft's Responsible AI toolbox.
Add a description, image, and links to the fairness-ai topic page so that developers can more easily learn about it.
To associate your repository with the fairness-ai topic, visit your repo's landing page and select "manage topics."