Home > Research > Publications & Outputs > AfriMTE and AfriCOMET

Electronic data

  • 2311.09828v1

    Final published version, 1.7 MB, PDF document

    Available under license: CC BY: Creative Commons Attribution 4.0 International License

Links

Keywords

View graph of relations

AfriMTE and AfriCOMET: Empowering COMET to Embrace Under-resourced African Languages

Research output: Working paperPreprint

Published
  • Jiayi Wang
  • David Ifeoluwa Adelani
  • Sweta Agrawal
  • Ricardo Rei
  • Eleftheria Briakou
  • Marine Carpuat
  • Marek Masiak
  • Xuanli He
  • Sofia Bourhim
  • Andiswa Bukula
  • Muhidin Mohamed
  • Temitayo Olatoye
  • Hamam Mokayede
  • Christine Mwase
  • Wangui Kimotho
  • Foutse Yuehgoh
  • Anuoluwapo Aremu
  • Jessica Ojo
  • Shamsuddeen Hassan Muhammad
  • Salomey Osei
  • Abdul-Hakeem Omotayo
  • Perez Ogayo
  • Oumaima Hourrane
  • Salma El Anigri
  • Lolwethu Ndolela
  • Thabiso Mangwana
  • Shafie Abdi Mohamed
  • Ayinde Hassan
  • Oluwabusayo Olufunke Awoyomi
  • Lama Alkhaled
  • Sana Al-Azzawi
  • Naome A. Etori
  • Millicent Ochieng
  • Clemencia Siro
  • Samuel Njoroge
  • Eric Muchiri
  • Wangari Kimotho
  • Lyse Naomi Wamba Momo
  • Daud Abolade
  • Simbiat Ajao
  • Tosin Adewumi
  • Iyanuoluwa Shode
  • Ricky Macharm
  • Ruqayya Nasir Iro
  • Saheed S. Abdullahi
  • Stephen E. Moore
  • Bernard Opoku
  • Zainab Akinjobi
  • Abeeb Afolabi
  • Nnaemeka Obiefuna
  • Onyekachi Raphael Ogbu
  • Sam Brian
  • Verrah Akinyi Otiende
  • Chinedu Emmanuel Mbonu
  • Sakayo Toadoum Sari
  • Pontus Stenetorp
Close
Publication date16/11/2023
PublisherArxiv
<mark>Original language</mark>English

Abstract

Despite the progress we have recorded in scaling multilingual machine translation (MT) models and evaluation data to several under-resourced African languages, it is difficult to measure accurately the progress we have made on these languages because evaluation is often performed on n-gram matching metrics like BLEU that often have worse correlation with human judgments. Embedding-based metrics such as COMET correlate better; however, lack of evaluation data with human ratings for under-resourced languages, complexity of annotation guidelines like Multidimensional Quality Metrics (MQM), and limited language coverage of multilingual encoders have hampered their applicability to African languages. In this paper, we address these challenges by creating high-quality human evaluation data with a simplified MQM guideline for error-span annotation and direct assessment (DA) scoring for 13 typologically diverse African languages. Furthermore, we develop AfriCOMET, a COMET evaluation metric for African languages by leveraging DA training data from high-resource languages and African-centric multilingual encoder (AfroXLM-Roberta) to create the state-of-the-art evaluation metric for African languages MT with respect to Spearman-rank correlation with human judgments (+0.406).