Add LlamaCppInferenceEngine
Related to #167
Hi Vijay, this would be a great addition - we need better support for CPU in exo!
Did you forget to commit some files?
@vs4vijay , any updates?
:eyes: @vs4vijay definitely interested in this PR, though based on prev comments some files are missing?
Hey @AlexCheema , @lexasub , @pcfreak30 - I got carried away with other work. I plan to pick this up this week.
@vs4vijay sorry to bug :), but I have an interest in this as well. Would love to look at using this to extend support to Intel iGPU/dGPU
https://github.com/intel/ipex-llm/blob/main/docs/mddocs/Quickstart/llama_cpp_quickstart.md