Inference Systems for Foundation Models
Foundation Model Inference
Running large language models on a single GPU for throughput-oriented scenarios.
FMInference
[NeurIPS'23] H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models.