Video-LLaMA icon indicating copy to clipboard operation
Video-LLaMA copied to clipboard

Does video-llama or any finetuned version support types such as float16 or bfloat16 when inferencing? Thanks

Open luentong opened this issue 9 months ago • 0 comments

When I change input weights type either using model.half() or dtype = torch.float16/bfloat16, it gets much slower on CPU inferencing.

luentong avatar Feb 21 '25 03:02 luentong