Skip to content
This repository has been archived by the owner on Nov 22, 2022. It is now read-only.

fp16 fixes for byte-lstm and distillation #1059

Closed
wants to merge 1 commit into from

Conversation

mwu1993
Copy link
Contributor

@mwu1993 mwu1993 commented Oct 17, 2019

Summary:
fixes:

  • sequence dimension of bytes for byte-lstm was set outside of precision; instead determine the dimension using precision.

  • class dimension of soft labels was determined by pad which depends on precision; instead always keep it the number of classes (just like for hard labels and model logits).

Reviewed By: chenyangyu1988

Differential Revision: D17967471

Summary:
fixes:

- sequence dimension of bytes for byte-lstm was set outside of `precision`; instead determine the dimension using `precision`.

- class dimension of soft labels was determined by `pad` which depends on precision; instead always keep it the number of classes (just like for hard labels and model logits).

Reviewed By: chenyangyu1988

Differential Revision: D17967471

fbshipit-source-id: 65748c3eedb79c525684930911bcc69a25bc00b3
@facebook-github-bot facebook-github-bot added the CLA Signed Do not delete this pull request or issue due to inactivity. label Oct 17, 2019
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D17967471

@facebook-github-bot
Copy link
Contributor

This pull request has been merged in 90c9c2c.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
CLA Signed Do not delete this pull request or issue due to inactivity. Merged
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants