Does Data Repair Lead to Fair Models?Curating Contextually Fair Data To Reduce Model BiasSharat Agarwal1* Sumanyu Muku2* Chetan Arora2 Saket Anand11IIIT Delhi, 2IIT Delhi, *equal contribution
In IEEE Winter Conference on Applications of Computer Vision (WACV), 2022 |
AbstractContextual information is a valuable cue for Deep Neural Networks (DNNs) to learn better representations and improve accuracy. However, co-occurrence bias in the training dataset may hamper a DNN model's generalizability to unseen scenarios in the real world. For example, in COCO, many object categories have a much higher co-occurrence with men compared to women, which can bias a DNN's prediction in favor of men. Recent works have focused on task-specific training strategies to handle bias in such scenarios, but fixing the available data is often ignored. In this paper, we propose a novel and more generic solution to address the contextual bias in the datasets by selecting a subset of the samples, which is fair in terms of the co-occurrence with various classes for a protected attribute. We introduce a data repair algorithm using the coefficient of variation, which can curate fair and contextually balanced data for a protected class(es). This helps in training a fair model irrespective of the task, architecture or training methodology. Our proposed solution is simple, effective and can even be used in an active learning setting where the data labels are not present or being generated incrementally. We demonstrate the effectiveness of our algorithm for the task of object detection and multi-label image classification across different datasets. Through a series of experiments, we validate that curating contextually fair data helps make model predictions fair by balancing the true positive rate for the protected class across groups without compromising on the model's overall performance. |
Bar plots in the left shows the count of men (
blue ) and women (red ) for different sports objects inCOCO dataset.
We propose to curate a fair data for a protected class (gender in this case) across its co-occurring classes.
As a result, the women which were detected as men with skis, tennis and surfboard in the middle column are now detected
as women with high confidence in the right column, with only 20% of the training data. The left-middle and left-bottom plots
show the co-occurrence frequency after selecting a subset using Repair , and our technique respectively.
|
COCO suffers from gender bias with its 79 object categories. We can see that males are biased in almost all the
categories by a significant margin as compared to females, thus this contextual bias is important to be fixed for real world application.
|
Comparison of ALOFT with other AL baselines. (a) Coefficient of variation for each selection,
lower the value fairer the selection
(b) Representational bias of the model.
|
Average precision of detecting (a) Male, (b) Female face in the presence of 39 different attributes, at different budgets in CelebA dataset.
AP of female has significantly improved compromising a bit in Male because of fair sampling.
|
Qualitative ResultsDetection of
CUP class by different Active Learning techniques on Dollar Street and ObjectNet dataset.
Green box denotes predicted bounding box of CUP when model was trained using only 20% of training data. |
Quantitative ResultsWe compare the average precision of detecting
CUP on OpenImages dataset when model was trained on CUP images from COCO
dataset by different methods in supervised and active learning setting. |