Skip to content

MLRS/MELABench

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MELABench logo A Maltese Evaluation Language Benchmark 🇲🇹

MELABench is an evaluation benchmark of model capabilities on Maltese. We have a publicly available benchmark which is periodically updated: https://huggingface.co/spaces/MLRS/MELABench

To run evaluation on this benchmark, we provide code to do this in various ways:

  • Prompting: runs models by prompting them with pre-defined instructions.
  • Fine-Tuning: trains models first before evaluating them.

We also release small fine-tuned models on each task:

Citation

This work was introduced in MELABenchv1: Benchmarking Large Language Models against Smaller Fine-Tuned Models for Low-Resource Maltese NLP. Cite as follows:

@inproceedings{micallef-borg-2025-melabenchv1,
    title = "{MELAB}enchv1: Benchmarking Large Language Models against Smaller Fine-Tuned Models for Low-Resource {M}altese {NLP}",
    author = "Micallef, Kurt  and
      Borg, Claudia",
    editor = "Che, Wanxiang  and
      Nabende, Joyce  and
      Shutova, Ekaterina  and
      Pilehvar, Mohammad Taher",
    booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
    month = jul,
    year = "2025",
    address = "Vienna, Austria",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2025.findings-acl.1053/",
    doi = "10.18653/v1/2025.findings-acl.1053",
    pages = "20505--20527",
    ISBN = "979-8-89176-256-5",
}

About

A Maltese Evaluation Language Benchmark

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages