-
Notifications
You must be signed in to change notification settings - Fork 41
Textual Entailment using Bejamin Riedel's model #6
Comments
Run with the following command (GPU optional). Gets about 88% accuracy which is within 1% of the score reported in [Riedel et al 2017]. On our dataset, we're sitting about 60% accuracy for a 2-way classification 👎
This is without early stopping or a clever learning rate schedule. I also want to evaluate different vocab/NN sizes. Might try a grid search on Sharc over the weekend. |
Interesting! The random baseline would be 50% right? (same number of supported/refuted). I would say let's first complete the full task evaluation (and make sure we are happy with the metrics) and then we optimize the various components according to what the metrics tell us. |
Training with randomly sampled pages for the not enough info class: Accuracy on gold labels dev set is approx: 70%. A random baseline would be 33% Will try incorporating DRQA predictions into the training set too. |
Training on pages solely on pages retrieved from DRQA for the Not Enough Info class gives a dev accuracy of 0.37 Training on FNC (merging discuss and unrelated into not enough info) and testing on pages predicted with DRQA gives an accuracy of 0.35. |
Interesting! It is still the case that 0.33 is the random baseline, right?
While accuracy is the main metric (given the balanced dataset), it would be
good to look at the confusion matrix.
…On Tue, 5 Dec 2017 at 12:15 James Thorne ***@***.***> wrote:
Training on pages solely on pages retrieved from DRQA for the Not Enough
Info class gives a dev accuracy of 0.37
Training on FNC (merging discuss and unrelated into not enough info) and
testing on pages predicted with DRQA gives an accuracy of 0.35.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#6 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABbUhV71DHAHu0JV-Bxw8xfZLV8ld01-ks5s9TO9gaJpZM4QrfKr>
.
|
Got 2 families of experiments going for this MLP model for generating training data for the Oracle RS - no DrQA for the test-time predictions. Just using the annotator labeled pages. Accuracy 71%. Confusion matrix/classification report: https://pastebin.com/x43y9vR0 Oracle NN - using DrQA just to identify the nearest neighbour pages for DrQA selecting k pages for all claims Confusion matrix for k=1 RS model https://pastebin.com/UQtTHYvK |
PS: that has early stopping with patience=8. I think the reason the RS model is doing well is because the cosine similarity between TF-IDF vectors (one of the features) is going to be v. low for unrelated documents. this might work quite nicely as a document relatedness filter. |
Hey, yes, makes sense. Random is often hard to beat, and for a good reason (unless we know what mistakes we will make). Is it correct to say that the oracle RS is a bit more oracle than the oracle NN as the first uses the labeled pages by the annotators while the second one uses DRQA? |
Both use labeled pages for the support/refutes classes. It's just for the NEI class where we have no labeled pages. I think because the nearest neighbour pages are more semantically similar than randomly sampled, the classifier needs to be more sensitive which we cannot achieve with this mlp |
got it. The reason could be what you say. Maybe see if the NN chosen ones help if added to the RS chosen ones. |
No description provided.
The text was updated successfully, but these errors were encountered: