@inproceedings{8c66715879c44a7bb01483bd2e16adf6,
title = "Information Parity: Measuring and Predicting the Multilingual Capabilities of Language Models",
abstract = "Large Language Models (LLMs) are increasingly deployed in user-facing applications worldwide, necessitating handling multiple languages across various tasks.We propose a metric called Information Parity (IP) that can predict an LLM's capabilities across multiple languages in a task-agnostic manner.IP is well-motivated from an information theoretic perspective: it is associated with the LLM's efficiency of compressing the text in a given language compared to a reference language.We evaluate IP and other popular metrics such as Tokenization Parity (TP) and Tokenizer Fertility (TF) on several variants of open-sourced LLMs (Llama2, Gemma, Mistral).Among all metrics known to us, IP is better correlated with existing task-specific benchmark scores from the literature and thus better predicts such scores in a certain language.These findings show that IP may be useful for ranking multilingual LLMs' capabilities regardless of the downstream task.",
author = "Alexander Tsvetkov and Alon Kipnis",
note = "Publisher Copyright: {\textcopyright} 2024 Association for Computational Linguistics.; 2024 Findings of the Association for Computational Linguistics, EMNLP 2024 ; Conference date: 12-11-2024 Through 16-11-2024",
year = "2024",
month = jan,
day = "1",
doi = "10.18653/v1/2024.findings-emnlp.468",
language = "English",
series = "EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2024",
publisher = "Association for Computational Linguistics (ACL)",
pages = "7971--7989",
editor = "Yaser Al-Onaizan and Mohit Bansal and Yun-Nung Chen",
booktitle = "EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2024",
address = "United States",
}