Multilingual Knowledge Editing Benchmark for Large Language Models
MLaKE is a novel benchmark for evaluating the multilingual knowledge editing capabilities of large language models, comprising 5,360 single-hop and 4,072 multi-hop questions across five languages (English, Chinese, Japanese, French, German).