TensorRT-LLM[Branchv0.12.0-jetson] Quick confirmation: Gemma2 not supported yet?
Greetings everyone. 0. I am trying to use TensorRT-LLM to deploy Gemma2 LLM on the Jetson AGX Orin platform.
- I am going through this instruction: https://github.com/NVIDIA/TensorRT-LLM/tree/v0.12.0-jetson/examples/gemma
- I downloaded the genuine checkpoint from huggingface source: huggingface-cli download --resume-download google/gemma-2-27b-it
- After run following code:
python3 ./convert_checkpoint.py
--ckpt-type jax
--model-dir /home/nvidia/.cache/huggingface/hub/models--google--gemma-2-27b-it/snapshots/aaf20e6b9f4c0fcf043f6fb2a2068419086d77b0
--dtype bfloat16
--world-size 1
--output-model-dir /home/nvidia/projects/TensorRT-LLM/examples/gemma/gemma-2-27b-it-bf16-cvt-ckpt
I got following error:
python3 ./convert_checkpoint.py
--ckpt-type jax
--model-dir /home/nvidia/.cache/huggingface/hub/models--google--gemma-2-27b-it/snapshots/aaf20e6b9f4c0fcf043f6fb2a2068419086d77b0
--dtype bfloat16
--world-size 1
--output-model-dir /home/nvidia/projects/TensorRT-LLM/examples/gemma/gemma-2-27b-it-bf16-cvt-ckpt
[TensorRT-LLM] TensorRT-LLM version: 0.12.0
Loading source parameters from /home/nvidia/.cache/huggingface/hub/models--google--gemma-2-27b-it/snapshots/aaf20e6b9f4c0fcf043f6fb2a2068419086d77b0
Traceback (most recent call last):
File "/home/nvidia/projects/TensorRT-LLM/examples/gemma/./convert_checkpoint.py", line 268, in
So I just want to quickly confirm that Gemma2 is not still supported by TensorRT-LLM[Branchv0.12.0-jetson] yet? Thank you very much for any hint or information.
My model conversion failed too (for a different reason), but it's hf checkpoint, isn't it?