Reading: Learning Cross-Lingual Sentence Representations via a Multi-task Dual-Encoder Model #8
Labels
Conf: ACL
Annual Meeting of the Association for Computational Linguistics
Representation
Transfer Learning
0. Paper
@inproceedings{chidambaram-etal-2019-learning,
title = "Learning Cross-Lingual Sentence Representations via a Multi-task Dual-Encoder Model",
author = "Chidambaram, Muthu and
Yang, Yinfei and
Cer, Daniel and
Yuan, Steve and
Sung, Yunhsuan and
Strope, Brian and
Kurzweil, Ray",
booktitle = "Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019)",
month = aug,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/W19-4330",
doi = "10.18653/v1/W19-4330",
pages = "250--259",
}
Article is here
1. What is it?
In this paper, the authors proposed a novel approach for cross-lingual representation learning using Universal Sentence Encoder.
2. What is amazing compared to previous studies?
They construct a multitask training scheme using
3. Where is the key to technologies and techniques?
The key is Multi-Task Dual-Encoder Model.
Input sentence
sIi
and response sentencesRi
, and seek to ranksRi
over all other possible response sentences.Maximize the log-likelihood,
P(sRi|sIi)
for each task.However,
P(sRi|sIi)
is hard to calculated, so they usedP'(sRi|sIi)
as below:They used 2 USE (Transformer encoder based) to embed each sentence.
To calculate the sentence representation, they used the average of each position of words in a sentence.
4. How did validate it?
They evaluated their learned representation using monolingual and cross-lingual tasks.
Their model achieved near-state-of-the-art or state-of-the-art performance on a variety of English tasks.
5. Is there a discussion?
6. Which paper should read next?
The text was updated successfully, but these errors were encountered: