Core Concepts
Incorporating influence functions as feedback improves model performance in subjective tasks.
Abstract
Abstract:
Influence functions quantify perturbations of individual train instances impacting test predictions.
INFFEED uses influence functions to improve model performance and identify data points for annotation.
Introduction:
Real-world data distribution imbalance affects model performance.
Transparency enhances model explainability and trustworthiness.
Data Extraction:
"INFFEED outperforms state-of-the-art baselines by 4% for hate speech classification, 3.5% for stance classification, and 3% for irony and sarcasm detection."
Related Work:
Interpretability issues in deep learning models.
Influence functions provide insights into training data impact on model predictions.
Methodology:
INFFEED framework adjusts labels based on influencers to enhance model performance.
Evaluation on subjective tasks shows significant F1 score improvements over baselines.
Experimental Setup:
Three setups with varying dataset sizes demonstrate improved performance with increased data.
Results:
INFFEED outperforms baselines across multiple datasets, showing statistically significant results.
Stats
INFFEEDはヘイトスピーチ分類で最大マクロF1スコアの改善を示す。