Please use this identifier to cite or link to this item:
https://www.um.edu.mt/library/oar/handle/123456789/127920| Title: | Large language model for Maltese |
| Authors: | Bonnici, Kelsey (2024) |
| Keywords: | Maltese language Natural language processing (Computer science) Natural language generation (Computer science) -- Computer programs |
| Issue Date: | 2024 |
| Citation: | Bonnici, K. (2024). Large language model for Maltese (Bachelor's dissertation). |
| Abstract: | Language models are essential components in natural language processing, facilitating tasks such as text generation and comprehension across diverse linguistic landscapes. However, tailored models for less prevalent languages, like Maltese, are scarce, presenting challenges in accessing language‐specific applications. It is also widely recognised that the development of language models of this nature is associated with a considerable financial investment. This study addresses these challenges by presenting Gendus, a low‐cost, instruction‐tuned Maltese language model. This study contributes to existing knowledge by demonstrating the effectiveness of instruction‐tuning and cost‐cutting techniques in developing language models for languages like Maltese. By creating a tailored Maltese language model, we not only open avenues for diverse applications for Maltese speakers but also continue developing the framework for creating models for other underrepresented languages. Our methodology is adapted from established practices used for developing such language models. We begin by constructing a dataset comprising 52,000 instructions translated into Maltese using machine translation. Subsequently, we employ an English base language model, specifically Llama 2 7B, a decoder‐only model, and fine‐tune it on the instructions using PEFT and LoRA, thereby imbuing it with knowledge of the Maltese language. The results of a comparative evaluation with BERTu, a Maltese encoder‐only language model, showcase a narrow performance margin with Gendus. For sentiment ana‐ lysis, Gendus achieved 75.41% while BERTu scored slightly higher at 78.96%. Similarly, in named‐entity recognition, Gendus attained 79.15% compared to BERTu’s 86.77%. However, despite not achieving superiority, our model demonstrates a 99.78% reduction in training costs, underscoring its cost‐effectiveness. The affordability of our approach makes it a compelling option, especially in projects with budget constraints, where sacrificing slight performance gains for significant cost savings is a viable trade‐ off. Moreover, our model demonstrates capabilities for open‐ended text generation, enhancing its versatility and potential for various natural language processing tasks. |
| Description: | B.Sc. IT (Hons)(Melit.) |
| URI: | https://www.um.edu.mt/library/oar/handle/123456789/127920 |
| Appears in Collections: | Dissertations - FacICT - 2024 Dissertations - FacICTAI - 2024 |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| 2408ICTICT390900015921_1.PDF Restricted Access | 11.32 MB | Adobe PDF | View/Open Request a copy |
Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.
