HOME | ABSTRACT | METHODS | CITATION | DEMO | CONTACT-US
Facial attribute classification algorithms frequently manifest demographic biases, disproportionately impacting specific gender and racial groups. Existing bias mitigation techniques often lack generalizability, require demographically annotated training sets, exhibit application-specific limitations, and entail a trade-off between fairness and classification accuracy. This trade-off implies that achieving fairness often results in diminished classification accuracy for the most proficient demographic sub-group. In this paper, we propose a continual learning framework designed to mitigate bias in facial attribute classification algorithms by integrating human-machine partnership, especially during the deployment stage. Our methodology harnesses the expertise of human annotators to label uncertain data samples, subsequently used to fine-tune a deep neural network over a time period. Through iterative refinement of the network's predictions with human guidance, we seek to enhance the accuracy and fairness of facial attribute classification. Extensive experimentation on gender and smile attribute classification tasks validates the efficacy of our approach, supported by empirical results from four datasets. The outcomes consistently demonstrate accuracy improvements and reduced bias across both classification tasks.
Method | Acceptance | Accuracy | Standard deviation |
---|---|---|---|
Shannon's Entropy | ` | s | d |
Dirichlet's Uncertainty Estimation | | | ||
Boundary Proximity Confidence-based Outlier Detection | ` | s | d |
Ensemble-based Outlier Detection | | | ||
Multiview Disagreemen | | |
We discussed two methods in the paper
For this method, we defined the uncertain data as the test data which was misclassified with high confidence score. To order to capture the definition, we proposed the following, The data is called to be uncertain when,
If there is a disagreement between two scores obtained from the main classifier model, and the combination of different pruned and quantized models, then the data is considered as uncertain.
Data points are considered as uncertain when it lies close to the decision boundary.