-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is Cost actually a Layer? #1311
Comments
Yes, I agree with you. |
Hi @jacquesqiao , @helinwang and I learned something more from @emailweixu this afternoon that it is not flexible enough to express the to-be-trained network by a composition of model and cost. @emailweixu gave us an example problem: suppose that we are going to learn a text embedding, f, of our inputs in the way that if we have "search result A is closer to query Q than search result B" as a training instance, we should have sim{f(A), f(Q} > sim{f(B), f(Q)}. In order to learn the f, we need to construct the following 3-branch network:
In this example, the model is actually Actually, I remember that @reyoung once said that he thought that a model should include the cost. But a precise statement should be that the cost should be included in the to-be-trained network, instead of the model. After this discussion, we reached a point that
|
By this conclusion
I am closing this issue and open #1315 for the discussion of the concept network. |
It seems that it is.
Consider the following network for image classification:
If we are going to train it by minimizing the squared error, we can add
So it seems that the "model" should contain only the part from image to softmax, and at training time we add label and squared_error.
The text was updated successfully, but these errors were encountered: