nanort icon indicating copy to clipboard operation
nanort copied to clipboard

[TODO] SIMD, BF16/FP16, INT8 optimization

Open syoyo opened this issue 2 years ago • 1 comments

Currently NanoRT does not utilize SIMD/AVX.

Also no quantized BVH support.

It'd be better to start to consider optimization and quantization.

Fortunately, recent CPU architecture(AlderLake, ZEN4) supports native BF16/FP16 and INT8 op support, which will boost quantized BVH construction/traversal.

syoyo avatar Feb 06 '23 11:02 syoyo

We can utilize https://github.com/DLTcollab/sse2neon for SIMDizing code for SSE and NEON(Arm) target. (TODO: RISC-V SIMD)

syoyo avatar Feb 22 '23 13:02 syoyo