The repository contains software library for Data Augmentation Services
-
Updated
Aug 1, 2018 - Python
The repository contains software library for Data Augmentation Services
Location-adjusted Wald statistics
A method to preprocess the training data, producing an adjusted dataset that is independent of the group variable with minimum information loss.
Tensorflow implementation of Learning Not to Learn (CVPR 2019)
Critical questions to help you gain useful information, clarify the context, figure out the pain points, and overcome biases.
unbiased toxicity detection from comments
NeurIPS 2019 Paper: RUBi : Reducing Unimodal Biases for Visual Question Answering
Sampling algorithms and machine learning models to reduce bias and predict credit risk.
This repository contains the experiments conducted in the ICLR 2022 spotlight paper "On the Importance of Firth Bias Reduction in Few-Shot Classification".
This repository contains the code to replicate the numerical studies presented in the paper "A Flexible Bias Correction Method based on Inconsistent Estimators".
🔍In recent years the advancement of ML (machine learning) increased automation for tasks in different domains. One of the challanges was an issues with job recruitment systems that demonstrated bias toward female applicants [4]. This repo will investigate some of the techniques used to overcome these challenges. 👨🏽🔧
Methods for M-estimation of statistical models
A small and simple prototype designed to alert users of the bias of the news source.
This repository contains the firth bias reduction experiments on the few-shot distribution calibration method conducted in the ICLR 2022 spotlight paper "On the Importance of Firth Bias Reduction in Few-Shot Classification".
Bias reduction in quasi likelihood estimation
[ICML 2022] Channel Importance Matters in Few-shot Image Classification
Toolkit for Auditing and Mitigating Bias and Fairness of Machine Learning Systems 🔎🤖🧰
WEFE: The Word Embeddings Fairness Evaluation Framework. WEFE is a framework that standardizes the bias measurement and mitigation in Word Embeddings models. Please feel welcome to open an issue in case you have any questions or a pull request if you want to contribute to the project!
A comprehensive set of fairness metrics for datasets and machine learning models, explanations for these metrics, and algorithms to mitigate bias in datasets and models.
Pytorch implementation of 'Explaining text classifiers with counterfactual representations' (Lemberger & Saillenfest, 2024)
Add a description, image, and links to the bias-reduction topic page so that developers can more easily learn about it.
To associate your repository with the bias-reduction topic, visit your repo's landing page and select "manage topics."