Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to measure the effectiveness of tag "controls" (e.g. sentiment)? #33

Open
enzoampil opened this issue Apr 18, 2020 · 2 comments
Open
Labels
documentation Improvements or additions to documentation enhancement New feature or request question Further information is requested

Comments

@enzoampil
Copy link
Owner

I was thinking that simple methodology would be to generate sequences for sentiment spans and measure accuracy based on some overlap measure, e.g. Jaccard, or rougue.

While imperfect, this would be an initial approach to communicating the effectiveness of utilizing text generation controls, derived from pre-trained supervised models.

Tweets would be a good start.

Potential dataset

https://www.kaggle.com/c/tweet-sentiment-extraction/data

@enzoampil enzoampil added documentation Improvements or additions to documentation enhancement New feature or request question Further information is requested labels Apr 18, 2020
@enzoampil
Copy link
Owner Author

Can also try to get a pooled version of each predicted token vector based on the regular Tito Joker model, and another model that generates only positive statements. This new model should be a language model fine tuned on positive examples of a sentiment analysis dataset.

@enzoampil
Copy link
Owner Author

enzoampil commented Apr 23, 2020

OpenAI published exactly how to do the above with IMDB reviews!!

Interesting to see that the tag controls are quite similar to ours, the main difference seems to be the use of an additional reward function for the sentiment special tokens.

https://lvwerra.github.io/trl/05-gpt2-sentiment-control/

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation enhancement New feature or request question Further information is requested
Projects
None yet
Development

No branches or pull requests

1 participant