-
Notifications
You must be signed in to change notification settings - Fork 42
Issues: NVIDIA/TensorRT-Model-Optimizer
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
FP16 and FP32 shows 30% lower accuracy compared to INT8 for the ViT Example in ONNX_PTQ
#106
opened Nov 13, 2024 by
chjej202
In cache_diffusion example, can we use dynamic image shape & batch size?
#101
opened Nov 4, 2024 by
wxsms
test fail with compiling error AttributeError: _ARRAY_API not found
#87
opened Oct 11, 2024 by
braindevices
Whether the comfyui_tensorrt node can load the tensorrt plan model generated by the tool ?
#85
opened Oct 11, 2024 by
blacklong28
Do TP and PP parameters play a role in the quantitative calibration stage
#84
opened Oct 10, 2024 by
hadoop2xu
Bringing Back Effective Quantization: Using ModelOPT for YOLO and Similar Architectures
#83
opened Oct 9, 2024 by
levipereira
[LLM PTQ] non-fatal error during eval (UnicodeDecodeError: 'utf-8' codec can't decode byte)
#82
opened Oct 4, 2024 by
stas00
Please help, ModelOPT int8 quantized model runs slower than fp16 quantized model.
#80
opened Oct 4, 2024 by
Rajjeshwar
Previous Next
ProTip!
Adding no:label will show everything without a label.