-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Increase coverage #57
Comments
I checked |
I added the experiment with VQ architecture in branch 57-increase-coverage. I think we are good regarding |
Anything we need to do regarding this issue? |
The VQ test looks great! Let's make a PR for the branch, looks mergeable to me and it increases coverage by 2.36%, see https://app.codecov.io/gh/spokenlanguage/platalea/compare/master...57-increase-coverage/overview. I went through the coverage report in detail and other ways to increase coverage would be:
|
My guess is that actually 3. will give the biggest boost and is also most valuable to cover. All cost and validation paths are currently uncovered. |
My comments for each of the points above. In general, I am in favor of keeping the tests to a minimum and only for important parts, but let's discuss that during our next meeting. |
Ok great, agreed on the overkill parts, 100% is not a goal in itself. So to summarize, a list of reasonable/feasible actions to increase coverage:
Two remaining questions:
|
I'd keep alternative attention mechanisms which could serve in future experiments. |
@bhigy: For the SLT path, do we actually need a test dataset in a different language than English? Or does the actual data not matter for "smoke testing" (testing whether a run works at all)? In the first case, we could add some Japanese sentences to flickr1d. Is the stuff you used for those experiments public data? If it doesn't matter, we could just put some lorum ipsum there. |
It doesn't matter if the captions are actually in a different language but we need to set the language as 'jp' for |
Ok, the SLT path gave a nice +0.9% in the coverage. That was the last of the wish list! One final question @bhigy: we now get these warnings from that pip_seq test with Japanese texts:
What do these mean? Is there anything we can do about these? Or is it just because flickr1d is too small with only 50 sentences? |
I think the size of the dataset is the issue. Could you quickly check what happens when you run it in normal conditions? |
Turns out it's not the small dataset (the warnings also occur with the full dataset), but the small hidden layer size that I had set for the tests. Increasing that to 8 gets rid of the warnings, so I'll push that to the PR branch as well and then we can wrap it up. |
We should take a look at the coverage reports (now that they are fixed #19) and figure out why some parts of the code are not covered by the experiment runs on CI. It seems for instance that in
encoders.py
there are a lot of unused models. Are they still used in some other dependent package or can we remove them?Related to #41.
The text was updated successfully, but these errors were encountered: