Sebastian Nicolas

Results 7 comments of Sebastian Nicolas

> > Getting this output now after > > ``` > exo --inference-engine pytorch --run-model llama-3.1-8b > ``` I'm assuming this comment was meant for #139 ?

@AlexCheema I added support for benchmarking on the MLX inference engine. Right now it only benchmarks f32 and f16 calculations because mlx doesn't support matrix multiplication for int8. Not sure...

@AlexCheema I created a benchmark for tinygrad, cleaned up the mlx benchmark and attempted to implement your requests from your last review. I also removed the dict from device_capabilites and...

@AlexCheema DeviceCapabilites are now lazily computed. PTAL

@AlexCheema resolved merge conflicts.