quantized model using AWQ and lora weights #2703
Labels
Investigating
Low Precision
Issue about lower bit quantization, including int8, int4, fp8
triaged
Issue has been triaged by maintainers
Hello:
Does TensorRT-LLM supports a model quantized with AWQ and the lora weights trained on the quantized weights?
The text was updated successfully, but these errors were encountered: