|
Evaluation of cross-lingual machine translation problems
Pre-trained mBERT, trained with cross-lingual sentences without parallel corpora
XMoverScore, calculated at a token level, not sentence level
Evaluators like BERTScore need manual layer selection
TeacherSim architecture
Cross-lingual sentence-level representation
Monolingual sentence representation as teacher
Fine-tuning with parallel corpora, delivering SOTA performance
Experiments
TeacherSim achieves SOTA performance for cross-lingual evaluation
TeacherSim is easy to use, with last layer always selected, like SBERT
TeacherSim is much more accurate for token-level alignment
|