You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
If I understand this post correctly, in most BERT-related articles, when 12 layers are mentioned for the uncased model, this corresponds to layer_output_1 to layer_output_12 in output from RBERT::extract_features(), ya?
For completeness, would be good to return the bare token embeddings before any transformer layers along with the layer outputs.
The text was updated successfully, but these errors were encountered: