Abstract
Large Language Models (LLMs) have demonstrated remarkable success as general-purpose task solvers across various fields. However, their capabilities remain limited when addressing domain-specific problems, particularly in downstream NLP tasks. Research has shown that models fine-tuned on instruction-based downstream NLP datasets outperform those that are not fine-tuned. While most efforts in this area have primarily focused on resource-rich languages like English and broad domains, little attention has been given to multilingual settings and specific domains. To address this gap, this study focuses on developing a specialized LLM, LlamaLens, for analyzing news and social media content in a multilingual context. To the best of our knowledge, this is the first attempt to tackle both domain specificity and multilinguality, with a particular focus on news and social media. Our experimental setup includes 18 tasks, represented by 52 datasets covering Arabic, English, and Hindi. We demonstrate that LlamaLens outperforms the current state-of-the-art (SOTA) on 23 testing sets, and achieves comparable performance on 8 sets. We make the models and resources publicly available for the research community (https://huggingface.co/QCRI).
| Original language | English |
|---|---|
| Pages | 5627-5649 |
| Number of pages | 23 |
| DOIs | |
| Publication status | Published - Apr 2025 |
| Event | Findings of the Association for Computational Linguistics: NAACL 2025 - Albuquerque, Unknown Duration: 29 Apr 2025 → 4 May 2025 |
Conference
| Conference | Findings of the Association for Computational Linguistics: NAACL 2025 |
|---|---|
| Country/Territory | Unknown |
| City | Albuquerque |
| Period | 29/04/25 → 4/05/25 |