-
Notifications
You must be signed in to change notification settings - Fork 195
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
embedding_lookup相比tensorflow 速度变慢 #114
Comments
应该不会比tf慢吧。虽然没人会去单纯比较一个embedding层 |
请问大佬,使用GPU该如何设置? |
Python还是C++?python的话看example |
|
你如果已经把cpu model move to GPU了,然后再调用from_torch接口,那就自动在gpu上计算了。 |
|
首先你要让张量都放在cpu上,你可以调用Copy函数,显式转移数据。 |
解决了。embedding_lookup确实不是热点,我想通过embedding算子的替换做一点技术积累,之后看能不能做成一个算子库,可以灵活选择的那种。未来希望和硬件、编译走的更近些 |
TurboTransformers也需要人贡献算子库,比如conv1d这种算子还没实现呢,你愿不愿意involve进来? |
|
readme底下有联系方式,你加我企业微信吧 |
GPU型号
GPU 0: GeForce RTX 2080 Ti
怎样调节batch_size、vocab_size、seq_len大小都比tensoflow要慢,
请问这符合预期吗? 可能是什么原因呢
The text was updated successfully, but these errors were encountered: