מושגי ליבה
Improving fairness in image classification models through pre-processing and in-processing methods is crucial for achieving a balance between accuracy and fairness.
תקציר
The study focuses on evaluating 13 state-of-the-art fairness improving techniques across three datasets. It compares the performance of pre-processing, in-processing, and post-processing methods, highlighting the importance of balancing model accuracy and fairness. The results show variations in method effectiveness across different datasets and metrics.
Abstract:
- Fairness challenges in deep learning models.
- Lack of systematic evaluation among existing methods.
- Large-scale empirical study to compare fairness improvement techniques.
- Pre-processing and in-processing methods outperform post-processing.
Introduction:
- AI systems' discriminatory tendencies raise ethical concerns.
- Existing approaches categorized into pre-processing, in-processing, and post-processing.
- Need for comprehensive comparison under the same setup.
Background:
- Definitions of individual fairness and group fairness.
- Limitations of existing studies: incomplete datasets, inconsistent metrics.
- Growing focus on fairness issues necessitates large-scale empirical study.
Studied Methods:
- Pre-processing: Undersampling (US), Oversampling (OS), Upweighting (UW), Bias Mimicking (BM).
- In-processing: Adversarial Training (Adv), Domain Independent Training (DI), Bias-Contrastive and Bias-Balanced Learning (BC+BB), FLAC, MMD-based Fair Distillation (MFD), Fair Deep Feature Reweighting (FDR).
- Post-processing: FairReprogram variants (FR-B, FR-P), Fairness-Aware Adversarial Perturbation (FAAP).
Experimental Setup:
- Dataset selection based on diversity and adaptability criteria.
- Measurement metrics include fairness metrics like SPD, DEO, EOD, AAOD, AED; performance metrics like Accuracy and Balanced Accuracy.
- Implementation details using ResNet-18 architecture with optimal configurations from respective papers.
Research Questions:
- Overall effectiveness of fairness improving methods?
- Influence of evaluation metrics on DL models' evaluation results?
- Influence of dataset settings on fairness improvements?
- Efficiency analysis of different fairness improving methods?
סטטיסטיקה
Fairness has been a critical issue that affects the adoption of deep learning models in real practice.
To improve model fairness, many existing methods have been proposed and evaluated to be effective in their own contexts.
Our findings reveal substantial variations in the performance of each method across different datasets and sensitive attributes.
ציטוטים
"Pre-processing methods outperform post-processing methods."
"While the best-performing method does not exist."