1. Alsentzer, E., Murphy, J., Boag, W., Weng, W.-H., Jindi, D., Naumann, T., et al. (2019). Publicly Available Clinical BERT Embeddings. In Proceedings of the 2nd clinical natural language processing workshop (pp. 72–78).
2. Antoun, W., Baly, F., & Hajj, H. (2020). AraBERT: Transformer-based Model for Arabic Language Understanding. In Proceedings of the 4th workshop on open-source arabic corpora and processing tools, with a shared task on offensive language detection (pp. 9–15).
3. Conneau, A., Rinott, R., Lample, G., Williams, A., Bowman, S., Schwenk, H., et al. (2018). XNLI: Evaluating Cross-lingual Sentence Representations. In Proceedings of the conference on empirical methods in natural language processing (pp. 2475–2485).
4. Cui, Y., Che, W., Liu, T., Qin, B., Wang, S., & Hu, G. (2020). Revisiting Pre-Trained Models for Chinese Natural Language Processing. In Findings of the association for computational linguistics: EMNLP (pp. 657–668).
5. Pre-training with whole word masking for Chinese BERT;Cui;IEEE/ACM Transactions on Audio, Speech, and Language Processing,2021