Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reading: Learning Cross-Lingual Sentence Representations via a Multi-task Dual-Encoder Model #8

Open
a1da4 opened this issue Sep 5, 2019 · 3 comments
Labels
Conf: ACL Annual Meeting of the Association for Computational Linguistics Representation Transfer Learning

Comments

@a1da4
Copy link
Owner

a1da4 commented Sep 5, 2019

0. Paper

@inproceedings{chidambaram-etal-2019-learning,
title = "Learning Cross-Lingual Sentence Representations via a Multi-task Dual-Encoder Model",
author = "Chidambaram, Muthu and
Yang, Yinfei and
Cer, Daniel and
Yuan, Steve and
Sung, Yunhsuan and
Strope, Brian and
Kurzweil, Ray",
booktitle = "Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019)",
month = aug,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/W19-4330",
doi = "10.18653/v1/W19-4330",
pages = "250--259",
}

Article is here

1. What is it?

In this paper, the authors proposed a novel approach for cross-lingual representation learning using Universal Sentence Encoder.

2. What is amazing compared to previous studies?

They construct a multitask training scheme using

  • native source|target language tasks
  • bridging translation task

3. Where is the key to technologies and techniques?

The key is Multi-Task Dual-Encoder Model.
スクリーンショット 2019-09-05 21 59 18

Input sentence sIi and response sentence sRi, and seek to rank sRi over all other possible response sentences.
Maximize the log-likelihood, P(sRi|sIi) for each task.
スクリーンショット 2019-09-05 22 10 46
However, P(sRi|sIi) is hard to calculated, so they used P'(sRi|sIi) as below:
スクリーンショット 2019-09-05 22 10 59

They used 2 USE (Transformer encoder based) to embed each sentence.
To calculate the sentence representation, they used the average of each position of words in a sentence.

4. How did validate it?

They evaluated their learned representation using monolingual and cross-lingual tasks.
Their model achieved near-state-of-the-art or state-of-the-art performance on a variety of English tasks.

5. Is there a discussion?

6. Which paper should read next?

@a1da4 a1da4 added Conf: ACL Annual Meeting of the Association for Computational Linguistics Representation Transfer Learning labels Sep 5, 2019
@a1da4 a1da4 changed the title Learning Cross-Lingual Sentence Representations via a Multi-task Dual-Encoder Model Reading: Learning Cross-Lingual Sentence Representations via a Multi-task Dual-Encoder Model Sep 18, 2019
@thak123
Copy link

thak123 commented Dec 29, 2019

Did you get any reference code for this paper ?

@a1da4
Copy link
Owner Author

a1da4 commented Dec 30, 2019

Did you get any reference code for this paper ?

Sorry, I didn’t.
But you can use pre-trained model. [TensorFlow Hub]

@thak123
Copy link

thak123 commented Jan 7, 2020

No I wanted to train the model from scratch to reproduce the results

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Conf: ACL Annual Meeting of the Association for Computational Linguistics Representation Transfer Learning
Projects
None yet
Development

No branches or pull requests

2 participants