To learn more, refer to the following papers:
- Cross-lingual Language Model Pretraining, by Guillaume Lample and Alexis Conneau, available at https://arxiv.org/pdf/1901.07291.pdf.
- Unsupervised Cross-lingual Representation Learning at Scale, by Alexis Conneau, Kartikay Khandelwal, et al., available at https://arxiv.org/pdf/1911.02116.pdf.
- FlauBERT: Unsupervised Language Model Pre-training for French, by Hang Le, Loic Vial, et al., available at https://arxiv.org/pdf/1912.05372.pdf.
- Spanish Pre-Trained BERT Model and Evaluation Data, by Jou-Hui Ho, Hojin Kang, et al., available at https://users.dcc.uchile.cl/~jperez/papers/pml4dc2020.pdf.
- BERTje: A Dutch BERT Model, by Wietse de Vries, Andreas van Cranenburgh, Arianna Bisazza, Tommaso Caselli, Gertjan van Noord, and Malvina Nissim, available at https://arxiv.org/pdf/1912.09582.pdf.
- Pre-Training with Whole Word Masking for Chinese BERT, by Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, and Guoping...