Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Non deterministic NLU training on GPU #6040

Closed
fparga opened this issue Jun 17, 2020 · 4 comments
Closed

Non deterministic NLU training on GPU #6040

fparga opened this issue Jun 17, 2020 · 4 comments
Labels
area:rasa-oss 🎡 Anything related to the open source Rasa framework stale type:bug 🐛 Inconsistencies or issues which will cause an issue or problem for users or implementors.

Comments

@fparga
Copy link

fparga commented Jun 17, 2020

Rasa version: 1.9.4

Python version: 3.6.9

Operating system (windows, osx, ...): linux

Issue:
NLU training on GPU is non reproducible, each training with the same pipeline and the same train set gives a model that perform differently. I get that GPU training is by nature non-deterministic (although I think some progress have been made on that front), but the variations we see here are significant and makes it really hard to compare the performance impact of pipeline configuration/parameter tuning.

We don't have this problem when training on CPU.

Note: we have 2701 intent examples (5 distinct intents)

Command or request that led to the issue:

rasa train nlu  --nlu training_data.json
rasa test nlu -m models --successes --no-plot --nlu test_data.json

Content of configuration file (config.yml):

language: fr
pipeline:
  - name: WhitespaceTokenizer
    case_sensitive: False
  - name: RegexFeaturizer
  - name: LexicalSyntacticFeaturizer
  - name: CountVectorsFeaturizer
  - name: CountVectorsFeaturizer
    analyzer: "char_wb"
    min_ngram: 1
    max_ngram: 4
  - name: DIETClassifier
    epochs: 100
    random_seed: 4321
@fparga fparga added area:rasa-oss 🎡 Anything related to the open source Rasa framework type:bug 🐛 Inconsistencies or issues which will cause an issue or problem for users or implementors. labels Jun 17, 2020
@sara-tagger
Copy link
Collaborator

Thanks for the issue, @alwx will get back to you about it soon!

You may find help in the docs and the forum, too 🤗

@fparga
Copy link
Author

fparga commented Jun 18, 2020

Some more context, from our experiments, the predicted intents seems to be the same when running tests on two models trained with the same train set (and pipeline) on GPU. It's the confidence scores that varies wildly, whereas they are identical when the training is done on CPU.

@stale
Copy link

stale bot commented Sep 19, 2020

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

@stale stale bot added the stale label Sep 19, 2020
@stale
Copy link

stale bot commented Sep 29, 2020

This issue has been automatically closed due to inactivity. Please create a new issue if you need more help.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area:rasa-oss 🎡 Anything related to the open source Rasa framework stale type:bug 🐛 Inconsistencies or issues which will cause an issue or problem for users or implementors.
Projects
None yet
Development

No branches or pull requests

2 participants