llama.cpp
llama.cpp copied to clipboard
On the edge llama?
Sorry to ask this... But is possible to get llama.cpp working on things like edge TPU?
https://coral.ai/products/accelerator-module/
That thing doesn't really have the memory needed for a LLM (8MB on-chip). You would have to transfer data in and out for processing and that would become the bottleneck.