@inproceedings{linlin-coling-22,
abstract = {Cross-lingual word embeddings (CLWE) have been proven useful in many cross-lingual tasks. However, most existing approaches to learn CLWE including the ones with contextual embeddings are sense agnostic. In this work, we propose a novel framework to align contextual embeddings at the sense level by leveraging cross-lingual signal from bilingual dictionaries only. We operationalize our framework by first proposing a novel sense-aware cross entropy loss to model word senses explicitly. The monolingual ELMo and BERT models pretrained with our sense-aware cross entropy loss demonstrate significant performance improvement for word sense disambiguation tasks. We then propose a sense alignment objective on top of the sense-aware cross entropy loss for cross-lingual model pretraining, and pretrain cross-lingual models for several language pairs (English to German/Spanish/Japanese/Chinese). Compared with the best baseline results, our cross-lingual models achieve 0.52\%, 2.09\% and 1.29\% average performance improvements on zero-shot cross-lingual NER, sentiment classification and XNLI tasks, respectively. We will release our code.},
address = {Gyeongju, Republic of Korea},
author = {Linlin Liu and Thien Hai Nguyen and Shafiq Joty and Lidong Bing and Luo Si
},
booktitle = {Proceedings of the 29th International Conference on Computational Linguistics},
month = {October},
pages = {xx--xx},
series = {COLING'22},
title = {Towards Multi-Sense Cross-Lingual Alignment of Contextual Embeddings},
url = {https://aclanthology.org/2022.coling-1.386/},
year = {2022}
}