llama.cpp
llama.cpp copied to clipboard
Add avx-512 support?
No clue but I think it may work faster
No clue but I think it may work faster
On i7-11700F, when I built with AVX512 flags in VS2022, my speed increased from ~0.37 s/token to ~0.34 s/token. Don't know if it was placebo or not. Maybe a proper assembly inline code will be better
No clue but I think it may work faster
On i7-11700F, when I built with AVX512 flags in VS2022, my speed increased from ~0.37 token/s to ~0.34 token/s. Don't know if it was placebo or not. Maybe a proper assembly inline code will be better
How a single flag can make 10% boost? Seems great!
How a single flag can make 10% boost? Seems great!
It's actually placebo :) Now I'm testing again, they're both running at the same speed
How a single flag can make 10% boost? Seems great!
It's actually placebo :) Now I'm testing again, they're both running at the same speed
😂hahah, that make more sense
This would require code changes, not just compiler flags.
Interesting https://stackoverflow.com/questions/63484266/enabling-avx512-support-on-compilation-significantly-decreases-performance