Abstract
Meaning conflation deficiency (MCD) presents a continual obstacle in natural language processing (NLP), especially for low-resourced and morphologically complex languages, where polysemy and contextual ambiguity diminish model precision in word sense disambiguation (WSD) tasks. This paper examines the optimisation of contextual embedding models, namely XLNet, ELMo, BART, and their improved variations, to tackle MCD in linguistic settings. Utilising Sesotho sa Leboa as a case study, researchers devised an enhanced XLNet architecture with specific hyperparameter optimisation, dynamic padding, early termination, and class-balanced training. Comparative assessments reveal that the optimised XLNet attains an accuracy of 91% and exhibits balanced precision–recall metrics of 92% and 91%, respectively, surpassing both its baseline counterpart and competing models. Optimised ELMo attained the greatest overall metrics (accuracy: 92%, F1-score: 96%), whilst optimised BART demonstrated significant accuracy improvements (96%) despite a reduced recall. The results demonstrate that fine-tuning contextual embeddings using MCD-specific methodologies significantly improves semantic disambiguation for under-represented languages. This study offers a scalable and flexible optimisation approach suitable for additional low-resource language contexts.
| Original language | English |
|---|---|
| Article number | 402 |
| Journal | Computers |
| Volume | 14 |
| Issue number | 9 |
| DOIs | |
| Publication status | Published - Sept 2025 |
| Externally published | Yes |
Keywords
- BART
- ELMo
- Sesotho sa Leboa
- XLNet optimisation
- contextual embeddings
- hyperparameter tuning
- low-resourced languages
- meaning conflation deficiency
- morphologically rich languages
- word sense disambiguation