You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I've read SBERT paper but there's one thing I can not understand.
In section 7 Computational Efficiency in the paper, SBERT uses 'smart batching' for computation efficiency by reducing computational overhead from padding.
For improved computation of sentence embeddings, we implemented a smart batching strategy: Sentences with similar lengths are grouped together and are only padded to the longest element in a mini-batch. This drastically reduces computational overhead from padding tokens.
However, the code doesn't use this sentences with similar lengths information.
It seems to me that the collate function just extract text and label from example, and then tokenize the texts and return.
Could please somebody let me know what am I missing?
The text was updated successfully, but these errors were encountered:
Hi.
I've read SBERT paper but there's one thing I can not understand.
In section 7 Computational Efficiency in the paper, SBERT uses 'smart batching' for computation efficiency by reducing computational overhead from padding.
However, the code doesn't use this sentences with similar lengths information.
It seems to me that the collate function just extract text and label from
example
, and then tokenize the texts and return.Could please somebody let me know what am I missing?
The text was updated successfully, but these errors were encountered: