Language imbalance during training can boost cross-lingual generalization in multilingual language models, leading to better performance on less frequent languages.