Assessing interpretability in machine translation models for low-resource languages

dc.contributor.advisorMarivate, Vukosi
dc.contributor.emailu17391718@tuks.co.zaen_US
dc.contributor.postgraduateGomba, Tsholofelo
dc.date.accessioned2025-01-22T08:57:42Z
dc.date.available2025-01-22T08:57:42Z
dc.date.created2025-04
dc.date.issued2024-12
dc.descriptionDissertation (MSc (Computer Science))--University of Pretoria, 2024.en_US
dc.description.abstractIn recent years, we have seen an increase in the adoption of Large Language Models (LLM) usage across many different applications. A practical example is OpenAI’s ChatGPT, a tool based on InstructGPT that uses pre-training combined with questioning answering and guidance with reinforcement learning with human feedback. A gap that still exists, the need for better coverage of low resource languages, has led to a substantial amount of research focused on multilingual LLMs in the Natural Language Processing (NLP) domain bringing about models such as NLLB-200, Glot500-m, and BLOOM. However, most of these black box multilingual LLMs fail at representing low resource languages, especially when applied to translation tasks, as their internal logic remain hidden from the user. This leaves one unable to account for or explain reasons for failures in real-life translations tasks. This research investigates the performance and interpretability of two models, a LLM and a small-scale model, trained on low-resource language pairs Xhosa Zulu and Tswana-Zulu. Both models make use of the transformer architecture. The research aims to evaluate the differences in translation quality and interpretability between the models, examining the role of attention mechanisms in capturing context and ensuring correct translations. The research aims to evaluate the (1) differences in translation quality and interpretability between models of different scales, (2) the impact of training dataset sizes on translation quality, and (3) the effectiveness of post-model eXplainable AI (XAI) methods in evaluating generated translations and model efficiency in low-resource language settings. The post-model methods used are attention pattern analysis, BLEU scores, MMD scores and human evaluation methods. We conclude that larger models handle linguistic complexities better, training on larger datasets generally improves translation quality, and diverse post-hoc evaluation methods are essential for a comprehensive assessment. This analysis contributes to a better understanding of the strengths and weaknesses of different model scales in machine translation, guiding future developments in XAI for machine translation of languages such as Swati, Tshiluba, Yoruba and other low-resource languages.en_US
dc.description.availabilityUnrestricteden_US
dc.description.degreeMSc (Computer Science)en_US
dc.description.departmentComputer Scienceen_US
dc.description.facultyEngineering, Built Environment and Information Technologyen
dc.description.sdgSDG-04: Quality educationen_US
dc.identifier.citation*en_US
dc.identifier.doi10.25403/UPresearchdata.28248956en_US
dc.identifier.otherA2025en_US
dc.identifier.urihttp://hdl.handle.net/2263/100236
dc.identifier.uriDOI: https://doi.org/10.25403/UPresearchdata.28248956.v1
dc.language.isoenen_US
dc.publisherUniversity of Pretoria
dc.rights© 2023 University of Pretoria. All rights reserved. The copyright in this work vests in the University of Pretoria. No part of this work may be reproduced or transmitted in any form or by any means, without the prior written permission of the University of Pretoria.
dc.subjectUCTDen_US
dc.subjectSustainable Development Goals (SDGs)en_US
dc.subjectInterpretabilityen_US
dc.subjectMachine translationen_US
dc.subjectTransformersen_US
dc.subjectLow-resource languagesen_US
dc.titleAssessing interpretability in machine translation models for low-resource languagesen_US
dc.typeDissertationen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Gomba_Assessing_2024.pdf
Size:
5.09 MB
Format:
Adobe Portable Document Format
Description:
Dissertation

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: