Cross-lingual Transfer of Monolingual Models
2022 (English)In: 2022 Language Resources and Evaluation Conference, LREC 2022, European Language Resources Association (ELRA) , 2022, p. 948-955Conference paper, Published paper (Refereed)
Abstract [en]
Recent studies in cross-lingual learning using multilingual models have cast doubt on the previous hypothesis that shared vocabulary and joint pre-training are the keys to cross-lingual generalization. We introduce a method for transferring monolingual models to other languages through continuous pre-training and study the effects of such transfer from four different languages to English. Our experimental results on GLUE show that the transferred models outperform an English model trained from scratch, independently of the source language. After probing the model representations, we find that model knowledge from the source language enhances the learning of syntactic and semantic knowledge in English. © licensed under CC-BY-NC-4.0.
Place, publisher, year, edition, pages
European Language Resources Association (ELRA) , 2022. p. 948-955
Keywords [en]
Learning systems, Cross-lingual, Generalisation, Model knowledge, Model representation, Pre-training, Semantics knowledge, Source language, Semantics
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:ri:diva-62612Scopus ID: 2-s2.0-85144427655ISBN: 9791095546726 (electronic)OAI: oai:DiVA.org:ri-62612DiVA, id: diva2:1730361
Conference
13th International Conference on Language Resources and Evaluation Conference, LREC 2022, 20 June 2022 through 25 June 2022
Note
Funding details: VINNOVA, 2019-02996; Funding text 1: This work is supported by the Swedish innovation agency (Vinnova) under contract 2019-02996.
2023-01-242023-01-242023-01-24Bibliographically approved