-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Loss not reducing, high validation and test metric values #11
Comments
Yes. I think the value of loss looks normal. As you mentioned, the rank loss is about 2. Remind that it is a listwise loss computed over 10 documents, so the cross-entropy on average is 0.2, which is OK. |
Thanks @rowedenny for the clarification. Another thing I wanted to ask is that the train output scores for the documents keep increasing with steps and even become in the range of 10^5. Since its log softmax, and relative difference of scores between documents matter, the loss remains almost same. |
Yes, I do observe that the scale of output scores keeps increasing. However, for ranking tasks trained with pairwise or listwise approaches, the scores between different queries are not comparable. In other words, we usually care about the order within a query, but barely worry about scores across different queries. In addition, L2 reg may help, yet the performance ranking model may suffer. |
Actually, there is no need to sort, because we have Recall we optimize the IPW or IRW with clicked items, so the multiplication here just ignores the items without click and assigns a small value. |
Thanks for the reply
|
Unbiased Learning to Rank uses clicks as positive labels while unclicks as negative labels.
We predict 1 to 10 because of the
No, it will not. |
I am still unsure on the underlying assumption here. As we know, the examination model doesn't depend on the documents features. Every-time we feed it a one hot encoded vector for all rank positions and get representation weights for them. |
I would like to refer you to the unbiased learning to rank tutorials as follow,
https://drive.google.com/file/d/1fyd3AbmtxTGLeIU6zPcYmaMFXfMQjk-D/view <https://drive.google.com/file/d/1fyd3AbmtxTGLeIU6zPcYmaMFXfMQjk-D/view>
https://www.youtube.com/watch?v=BEEfMrn9T9c&t=1511s&ab_channel=HarrieOosterhuis <https://www.youtube.com/watch?v=BEEfMrn9T9c&t=1511s&ab_channel=HarrieOosterhuis>
… On Jul 25, 2022, at 11:56 AM, Parth Shettiwar ***@***.***> wrote:
I am still unsure on the underlying assumption here. As we know, the examination model doesn't depend on the documents features. Every-time we feed it a one hot encoded vector for all rank positions and get representation weights for them.
So are we assuming some sort of ordering in the documents while loading the batch? Can we load the query-document in any order, even keeping the relevant documents in end?
Since in the formulation of observation probability for document, we denote it by variable o^x_q, for document x and query q. But in implementation I don't see any dependence of the observation probability on the document x or on its rank.
—
Reply to this email directly, view it on GitHub <#11 (comment)>, or unsubscribe <https://github.com/notifications/unsubscribe-auth/AHCCEBEUIZBG4FHVHRNAPTLVV22L7ANCNFSM53COM52Q>.
You are receiving this because you were mentioned.
|
Thanks, is there any particular time stamp or page you would like me to go through to address the above query? |
I tried to run the code with DLA algorithm on Yahoo dataset. Following is the output attached. I am not sure of the following observation where I am getting almost constant training loss of about 4 (with each rank loss and exam loss as about 2), and high validation and testing metric values of more than 0.9. I did try to observe the parameter values of 2 models, which are actually updating. Also the loss is just fluctuating in range of 3.9 to 4.5 always. Is there something I should do with hyperparameters, have kept the default learning rate of 0.05 and selection_bias_cutoff = 10. This is with respect to the pytorch implementation of the code
The text was updated successfully, but these errors were encountered: