Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sequence is shorter then the required number of steps #44

Closed
kaeflint opened this issue Sep 17, 2018 · 5 comments
Closed

Sequence is shorter then the required number of steps #44

kaeflint opened this issue Sep 17, 2018 · 5 comments

Comments

@kaeflint
Copy link

kaeflint commented Sep 17, 2018

No description provided.

@kaeflint kaeflint reopened this Sep 17, 2018
@kaeflint
Copy link
Author

Running
from nlgeval import NLGEval nlgeval = NLGEval() metrics_dict = nlgeval.compute_individual_metrics('The man and woman are on a vacation', 'The couple were on honeymoon.')
gave the ValueError: ('Sequence is shorter then the required number of steps : (n_steps, seq, seq.shape):', 1, array([], shape=(0, 1, 1), dtype=float32), (0, 1, 1))

@kaeflint kaeflint changed the title No such file or directory [File not found] Sequence is shorter then the required number of steps Sep 17, 2018
@juharris
Copy link
Member

Thanks for trying out our repo!
The first parameter (ref) should be a list like this:

from nlgeval import NLGEval
nlgeval = NLGEval()
metrics_dict = nlgeval.compute_individual_metrics(["The man and woman are on a vacation"], "The couple were on honeymoon.")

You can see some examples here: https://github.com/Maluuba/nlg-eval/blob/master/nlgeval/tests/test_nlgeval.py

@kaeflint
Copy link
Author

Thanks. working well now but I realized the ROUGE scores are different from what I obtained from using the official scripts

@juharris
Copy link
Member

Thanks for the feedback. Would you open a new issue for that with examples?

@kracwarlock
Copy link
Member

Our overlap-based metric implementations are derived from https://github.com/tylin/coco-caption. The ROUGE metric variant they implement is called ROUGE_L. Are you talking about the ROUGE_L metric or some other variant of it. If ROUGE_L doesn't match the official eval scripts then it would be great if you can open an issue for this.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants