-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU usage of model, if it's possible to increase the max length of the text to 256 or 512? #3
Comments
Yes,of course |
thanks for your reply!
|
the f1 increase slowly after 30 epoch trains on the joint model. using training data 30000 and dev data 100. if it's better to change the learning rate of bert and relation extraction network? Epoch: 1 |
2.if it's better to change the learning rate of bert and relation extraction network? 原论文作者在训练网络时,使用了更多的epochs,当然也有更多的数据。 |
您好,多谢回复。
·模型使用3000条训练数据,最后对测试集前100进行验证,最大F1值是81.8%; |
1.argmax([0.2,0.3,0.4,0.5,0.3]) = 3 |
`class Ner_model(tf.keras.Model):
您好,多谢回复 个人感觉您还是说的有点问题 这里tf.nn.softmax(x)之后的维度应该是batch * max_len * num_classes_of_entity 按照您的做法的前面应该是sigmod才对 如果是softmax的话,直接在num_classes_of_entity做softmax是最好的。 |
|
span based model suffer from to large max_len of text
The batch size of your model = 6
max_length of text = 128
num_classes of relation
you predict a 3d Matrix shaped 6 * 128 * 128 * 50
if it's possible to increase the max length of the text to 256 or 512?
The text was updated successfully, but these errors were encountered: