Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

when the text is very short #4

Closed
08tjlys opened this issue Nov 10, 2021 · 2 comments
Closed

when the text is very short #4

08tjlys opened this issue Nov 10, 2021 · 2 comments

Comments

@08tjlys
Copy link

08tjlys commented Nov 10, 2021

hi,
first of all, really thanks for such great project.
When i tried your inference code on hugging face, i found that when the text is short, for example like just "weather is fine", it will generate a longer unrelated summary.
i am totally new to summarization, so is there any thing i should pay attention to in such case? for example, in such case, generating inputs_ids will be different?
really appreciate your guide if possible, thanks!

@Tahmid04
Copy link
Collaborator

Hi, thanks for appreciating our work. Since XL-Sum is a news article-summary dataset, models pretrained on it are expected to generate news-like summaries, even when the input is not from the news domain, e.g., paper abstracts, and for your case non-article inputs. If you want to generate summaries for non-news inputs, we recommend you further fine-tune the model checkpoint on such data. And if you just want to reduce the output length, using a higher length penalty may suit your use case. Hope that helps!

@08tjlys
Copy link
Author

08tjlys commented Nov 13, 2021

thanks a lot, it really helps

@08tjlys 08tjlys closed this as completed Nov 13, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants