-
Notifications
You must be signed in to change notification settings - Fork 166
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
the transformer to be applied to classification #18
Comments
Hi, I believe the most straight forward solution would be to keep the original architecture, and only change the output module. Currently, I have a linear transformation followed by a sigmoid activation, I would start by simply replacing the activation with a softmax, and see from there. |
I currently want to input 250 features, segment them, and output the categories of these 250 features, so I just need to change the output module to softmax? |
Yes, set |
Thank you very much |
If it is the category of these 250 features, the output is like 250*4 |
Hi @maxjcohen , thanks for your great repo! Is it possible to change the transformer to understand sequence classification (many-to-one)? |
Hi, nothing is stopping you from setting |
Thank you for your reply @maxjcohen ! How exactly do you mean its different? From the way a RNN-model would take hidden states as further input? |
RNN carry a memory-like hidden state across time steps, while the Transformer has no notion of memory and compute time steps in parallel instead. |
How should I change the transformer to be applied to classification, such as seq2seq (many to many), how should I change it in the last layer of the model
The text was updated successfully, but these errors were encountered: