Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/127920
Title: Large language model for Maltese
Authors: Bonnici, Kelsey (2024)
Keywords: Maltese language
Natural language processing (Computer science)
Natural language generation (Computer science) -- Computer programs
Issue Date: 2024
Citation: Bonnici, K. (2024). Large language model for Maltese (Bachelor's dissertation).
Abstract: Language models are essential components in natural language processing, facilitating tasks such as text generation and comprehension across diverse linguistic landscapes. However, tailored models for less prevalent languages, like Maltese, are scarce, presenting challenges in accessing language‐specific applications. It is also widely recognised that the development of language models of this nature is associated with a considerable financial investment. This study addresses these challenges by presenting Gendus, a low‐cost, instruction‐tuned Maltese language model. This study contributes to existing knowledge by demonstrating the effectiveness of instruction‐tuning and cost‐cutting techniques in developing language models for languages like Maltese. By creating a tailored Maltese language model, we not only open avenues for diverse applications for Maltese speakers but also continue developing the framework for creating models for other underrepresented languages. Our methodology is adapted from established practices used for developing such language models. We begin by constructing a dataset comprising 52,000 instructions translated into Maltese using machine translation. Subsequently, we employ an English base language model, specifically Llama 2 7B, a decoder‐only model, and fine‐tune it on the instructions using PEFT and LoRA, thereby imbuing it with knowledge of the Maltese language. The results of a comparative evaluation with BERTu, a Maltese encoder‐only language model, showcase a narrow performance margin with Gendus. For sentiment ana‐ lysis, Gendus achieved 75.41% while BERTu scored slightly higher at 78.96%. Similarly, in named‐entity recognition, Gendus attained 79.15% compared to BERTu’s 86.77%. However, despite not achieving superiority, our model demonstrates a 99.78% reduction in training costs, underscoring its cost‐effectiveness. The affordability of our approach makes it a compelling option, especially in projects with budget constraints, where sacrificing slight performance gains for significant cost savings is a viable trade‐ off. Moreover, our model demonstrates capabilities for open‐ended text generation, enhancing its versatility and potential for various natural language processing tasks.
Description: B.Sc. IT (Hons)(Melit.)
URI: https://www.um.edu.mt/library/oar/handle/123456789/127920
Appears in Collections:Dissertations - FacICT - 2024
Dissertations - FacICTAI - 2024

Files in This Item:
File Description SizeFormat 
2408ICTICT390900015921_1.PDF
  Restricted Access
11.32 MBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.