Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add FP16 capablity #44

Merged
merged 25 commits into from Aug 1, 2019
Merged

Add FP16 capablity #44

merged 25 commits into from Aug 1, 2019

Conversation

marty1885
Copy link
Member

@marty1885 marty1885 commented Jul 19, 2019

WIP:

This PR aims to:

  • Add support to create FP16 tensor
  • Add FP16 tensor operators
  • Optimize HTM algorithm against FP16
  • Clean up how tensor's properties are checked in the backend
    • Reduce LoC, more readable

To both CPU and OpenCL backend.

And for OpenCL exclusively:

  • Check the given OpenCL device can process FP16

Tests for the type system

  • Check types have the correct size
  • Check resulting type of unitary operations
  • Check resulting type of binary operations
    • Besides comparison ops, I;m lazy
  • Check resulting type of general tensor operations
  • Somehow disable the fp16 tests when GPU doesn't support FP16
  • A way to convert SP/TM from float to float16
    • Just cast the permeance!

Serialize:

  • Save/load fp16 tensors

Ref #12

@marty1885 marty1885 marked this pull request as ready for review August 1, 2019 07:46
@marty1885
Copy link
Member Author

Done. Merging!

@marty1885 marty1885 merged commit d61c755 into etaler:master Aug 1, 2019
marty1885 added a commit to marty1885/Etaler that referenced this pull request Jun 19, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

1 participant