Skip to content
This repository has been archived by the owner on Sep 24, 2024. It is now read-only.

Fine-tuned Model #35

Closed
jiangix-paper opened this issue Jun 27, 2020 · 3 comments
Closed

Fine-tuned Model #35

jiangix-paper opened this issue Jun 27, 2020 · 3 comments

Comments

@jiangix-paper
Copy link

Hello, Thanks a lot for your great works. I want to know the principle of fine-tuned model. For example, if i want to fine-tune the pretrained model on supervised cnn/dailymail dataset, is the model seq2seq? And i only load the pretrained word embedding based on pegasus, and set it as the input of seq2seq?

@JingqingZ
Copy link
Contributor

If i want to fine-tune the pretrained model on supervised cnn/dailymail dataset, is the model seq2seq?

The model uses the encoder-decoder Transformer framework so yes it is seq2seq.

And i only load the pretrained word embedding based on pegasus, and set it as the input of seq2seq?

Sorry, I am not sure if I understand this question. Could you elaborate on your question?

@jiangix-paper
Copy link
Author

If i want to fine-tune the pretrained model on supervised cnn/dailymail dataset, is the model seq2seq?

The model uses the encoder-decoder Transformer framework so yes it is seq2seq.

And i only load the pretrained word embedding based on pegasus, and set it as the input of seq2seq?

Sorry, I am not sure if I understand this question. Could you elaborate on your question?

@jiangix-paper
Copy link
Author

Thanks for your reply. When I fine-tune the model with encoder-decoder transformer framework on cnn/dailymail dataset, whether the model structure and initial model parameter is the same as pretrained model. The model parameter of transformer encoder and decoder is updated during fine-tuning.

I want to sure whether the fine-tuned model are as follows:
IMG_20200628_151058

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants