Adversarial datasets, collected by exploiting model weaknesses, can improve the robustness of hate speech detection models.