Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[TODO] SIMD, BF16/FP16, INT8 optimization #79

Open
syoyo opened this issue Feb 6, 2023 · 1 comment
Open

[TODO] SIMD, BF16/FP16, INT8 optimization #79

syoyo opened this issue Feb 6, 2023 · 1 comment

Comments

@syoyo
Copy link
Contributor

syoyo commented Feb 6, 2023

Currently NanoRT does not utilize SIMD/AVX.

Also no quantized BVH support.

It'd be better to start to consider optimization and quantization.

Fortunately, recent CPU architecture(AlderLake, ZEN4) supports native BF16/FP16 and INT8 op support, which will boost quantized BVH construction/traversal.

@syoyo
Copy link
Contributor Author

syoyo commented Feb 22, 2023

We can utilize https://github.com/DLTcollab/sse2neon for SIMDizing code for SSE and NEON(Arm) target.
(TODO: RISC-V SIMD)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant