convert NVILA with 0.16.0 #2706
Labels
bug
Something isn't working
Investigating
LLM API/Workflow
triaged
Issue has been triaged by maintainers
System Info
I can not convet NVILA checkpoint with tensorRT-LLM 0.16.0 follow examples/multimodal/README.md
Who can help?
@ncomly-nvidia
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
--model_dir tmp/hf_models/${MODEL_NAME}
--output_dir tmp/trt_models/${MODEL_NAME}/fp16/1-gpu
--dtype float16
Expected behavior
Total time of reading and converting: 1.589 s
Total time of saving checkpoint: 5.007 s
Total time of converting checkpoints: 00:00:06
actual behavior
--dtype float16
[TensorRT-LLM] TensorRT-LLM version: 0.16.0
0.16.0
[01/20/2025-12:06:37] [TRT-LLM] [W] AutoConfig cannot load the huggingface config.
Traceback (most recent call last):
File "/usr/local/lib/python3.12/dist-packages/transformers/models/auto/configuration_auto.py", line 1034, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/transformers/models/auto/configuration_auto.py", line 736, in getitem
raise KeyError(key)
KeyError: 'llava_llama'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/app/tensorrt_llm/examples/multimodal/../llama/convert_checkpoint.py", line 581, in
main()
File "/app/tensorrt_llm/examples/multimodal/../llama/convert_checkpoint.py", line 573, in main
convert_and_save_hf(args)
File "/app/tensorrt_llm/examples/multimodal/../llama/convert_checkpoint.py", line 514, in convert_and_save_hf
execute(args.workers, [convert_and_save_rank] * world_size, args)
File "/app/tensorrt_llm/examples/multimodal/../llama/convert_checkpoint.py", line 521, in execute
f(args, rank)
File "/app/tensorrt_llm/examples/multimodal/../llama/convert_checkpoint.py", line 496, in convert_and_save_rank
llama = LLaMAForCausalLM.from_hugging_face(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/tensorrt_llm/models/llama/model.py", line 393, in from_hugging_face
config = LLaMAConfig.from_hugging_face(hf_config_or_dir,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/tensorrt_llm/models/llama/config.py", line 108, in from_hugging_face
hf_config = transformers.AutoConfig.from_pretrained(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/transformers/models/auto/configuration_auto.py", line 1036, in from_pretrained
raise ValueError(
ValueError: The checkpoint you are trying to load has model type
llava_llama
but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.additional notes
no
The text was updated successfully, but these errors were encountered: