Linguistic entity masking to improve cross-lingual representation of multilingual language models for low-resource languages

dc.citation.volumeLatest Articles
dc.contributor.authorFernando A
dc.contributor.authorRanathunga S
dc.date.accessioned2025-07-29T21:10:32Z
dc.date.available2025-07-29T21:10:32Z
dc.date.issued2025-07-19
dc.description.abstractMultilingual Pre-trained Language models (multiPLMs), trained on the Masked Language Modelling (MLM) objective are commonly being used for cross-lingual tasks such as bitext mining. However, the performance of these models is still suboptimal for low-resource languages (LRLs). To improve the language representation of a given multiPLM, it is possible to further pre-train it. This is known as continual pre-training. Previous research has shown that continual pre-training with MLM and subsequently with Translation Language Modelling (TLM) improves the cross-lingual representation of multiPLMs. However, during masking, both MLM and TLM give equal weight to all tokens in the input sequence, irrespective of the linguistic properties of the tokens. In this paper, we introduce a novel masking strategy, Linguistic Entity Masking (LEM) to be used in the continual pre-training step to further improve the cross-lingual representations of existing multiPLMs. In contrast to MLM and TLM, LEM limits masking to the linguistic entity types nouns, verbs and named entities, which hold a higher prominence in a sentence. Secondly, we limit masking to a single token within the linguistic entity span thus keeping more context, whereas, in MLM and TLM, tokens are masked randomly. We evaluate the effectiveness of LEM using three downstream tasks, namely bitext mining, parallel data curation and code-mixed sentiment analysis using three low-resource language pairs English-Sinhala, English-Tamil, and Sinhala-Tamil. Experiment results show that continually pre-training a multiPLM with LEM outperforms a multiPLM continually pre-trained with MLM+TLM for all three tasks.
dc.description.confidentialfalse
dc.identifier.citationFernando A, Ranathunga S. (2025). Linguistic entity masking to improve cross-lingual representation of multilingual language models for low-resource languages. Knowledge and Information Systems. Latest Articles.
dc.identifier.doi10.1007/s10115-025-02520-4
dc.identifier.eissn0219-3116
dc.identifier.elements-typejournal-article
dc.identifier.issn0219-1377
dc.identifier.urihttps://mro.massey.ac.nz/handle/10179/73253
dc.languageEnglish
dc.publisherSpringer-Verlag London Ltd
dc.publisher.urihttps://link.springer.com/article/10.1007/s10115-025-02520-
dc.relation.isPartOfKnowledge and Information Systems
dc.rights(c) The author/sen
dc.rights.licenseCC BYen
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en
dc.subjectMasked language modelling
dc.subjectTranslation language modelling
dc.subjectMultilingual pre-trained language model
dc.subjectBitext mining
dc.subjectSentiment analysis
dc.subjectXLM-R
dc.subjectSinhala
dc.subjectTamil
dc.titleLinguistic entity masking to improve cross-lingual representation of multilingual language models for low-resource languages
dc.typeJournal article
pubs.elements-id501733
pubs.organisational-groupOther

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
501733 PDF.pdf
Size:
1.73 MB
Format:
Adobe Portable Document Format
Description:
501733 PDF.pdf

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
9.22 KB
Format:
Plain Text
Description:

Collections