-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Examples for training phase of validation task #30
Comments
In case 1. no implementation needed. Just need to add the examples in the ontology.json file I think the case 1. makes more sense imo, since we want to control the training phase as much as possible. Providing good examples might be essential. |
We can also default to case 2 when there are no examples chosen for case 1. |
I like @ffont idea, as well! It is nice for the "cold start" problem we face. For case 1 I propose computing a ranking of sounds given a punctuation considering PP and Freesound ratings. |
I guess you mean Freesound ratings? |
Yes, thanks. I amended my comment! :) |
For the crowdsource launch (ie, annotations have been validated by a single rater) it has been suggested:
The examples could be used for (see #27 to locate this steps in the annotation protocol):
How to choose these examples for every category? Currently, it is proposed to:
|
Here is a dictionary {<aso_id> : [<fs_id>, <fs_id>, ...] , ...} which provides PP examples for each AudioSet Ontology category ids. I have combined Freesound ratings and downloads to sort them. As you said, now we need to manually validate some of this examples, and put them in the ontology_preCrowd.json file in the field "positive_examples" for each categories. |
Definitely the best way would be to implement this functionality on the web platform. So I will leave the option of filling directly the json file with the freesound examples. Please start adding and checking examples! |
Tool is ready for adding the examples:
|
After inspecting the tool, it seems good for the task. I have few suggestions:
|
|
Providing examples for all the 398 categories seems to be complicated. I think we don't have time to gathered enough examples for allowing to have this ready for the platform launching. |
|
As mentioned in #27 , ideally, the annotation protocol will consist of a training phase followed by a validation phase. In the former, some representative audio examples should be presented to the rater. How to choose these examples for every category? Several options:
The text was updated successfully, but these errors were encountered: