You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @luyang-huang96, thanks so much for posting the code.
Table 3 and 4 in your paper shows that the encoder variants (SINKHORN and LSH) can bring great performance.
To reproduce these results, I wonder if you use the hybrid attention in the encoder (how you set the input parameter encoder_not_hybrid)
If you use the hybrid attention in the encode (encoder_not_hybrid is false), I hope to know how you set the args.sw, args.encoder_linear, and args.encoder_kernel_linear.
If only use the SINKHORN for all encoder layers (encoder_not_hybrid is True), my result shows its performance can not compete with LED/Bigbird, when using the same-length inputs.
Hi @luyang-huang96, thanks so much for posting the code.
Table 3 and 4 in your paper shows that the encoder variants (SINKHORN and LSH) can bring great performance.
To reproduce these results, I wonder if you use the hybrid attention in the encoder (how you set the input parameter encoder_not_hybrid)
If you use the hybrid attention in the encode (encoder_not_hybrid is false), I hope to know how you set the args.sw, args.encoder_linear, and args.encoder_kernel_linear.
If only use the SINKHORN for all encoder layers (encoder_not_hybrid is True), my result shows its performance can not compete with LED/Bigbird, when using the same-length inputs.
LongDocSum/Model/longbart/longbartmodel.py
Lines 204 to 217 in d2b9bd0
The text was updated successfully, but these errors were encountered: