Skip to content

Pull requests: NVIDIA/TensorRT-Model-Optimizer

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[OMNIML-2244] enable fp8 and int8 ONNX export
#594 opened Nov 21, 2025 by ajrasane Loading…
[2/N] Added KDLoss based AutoQuantize
#592 opened Nov 20, 2025 by realAsma Loading…
Product Rename: TensorRT Model Optimizer to Model Optimizer
#583 opened Nov 20, 2025 by kevalmorabia97 Loading…
1 of 2 tasks
support for newer checkpoints
#582 opened Nov 20, 2025 by binghanc Draft
Bump TRT-LLM docker to 1.2.0rc2 (CUDA 13)
#578 opened Nov 19, 2025 by kevalmorabia97 Loading…
1 task
Feat: SGL backend for online SD training
#564 opened Nov 14, 2025 by h-guo18 Loading…
Fix hf_quant_config with kv cache type
#557 opened Nov 14, 2025 by jenchen13 Loading…
GPTQ Lite implementation
#555 opened Nov 13, 2025 by sugunav14 Loading…
1 of 2 tasks
[OMNIML-3015]Add per tensor/per channel MSE calibrator
#540 opened Nov 12, 2025 by Fridah-nv Loading…
2 tasks
Optimize NVFP4 Triton kernel
#533 opened Nov 11, 2025 by mxinO Loading…
parallel eagle draft
#523 opened Nov 6, 2025 by yeyu-nvidia Draft
[Bug #193] fix fp8 blockwise real quantization
#522 opened Nov 6, 2025 by meenchen Loading…
Fix BMM style MoE export in fp8_pc_pt recipe
#515 opened Nov 5, 2025 by Edwardf0t1 Loading…
ProTip! Type g i on any issue or pull request to go back to the issue listing page.