TensorRT Model Optimizer
Getting Started
Overview
Installation
Quick Start: Quantization
Quick Start: Quantization (Windows)
Quick Start: Pruning
Quick Start: Distillation
Quick Start: Sparsity
Guides
Support Matrix
Quantization
Pruning
NAS
Distillation
Sparsity
Saving & Restoring
Speculative Decoding
AutoCast (ONNX)
Deployment
TensorRT-LLM
DirectML
Unified HuggingFace Checkpoint
Examples
All GitHub Examples
Reference
Changelog
modelopt API
deploy
onnx
torch
distill
export
convert_hf_config
distribute
hf_config_map
layer_utils
mcore_config_map
model_config
model_config_export
model_config_utils
model_utils
plugins
postprocess
quant_utils
tensorrt_llm_type
tensorrt_llm_utils
transformer_engine
unified_export_hf
unified_export_megatron
nas
opt
prune
quantization
sparsity
speculative
trace
utils
Support
Contact us
FAQs
TensorRT Model Optimizer
modelopt API
torch
export
plugins
View page source
plugins
Modules
Export package plugin.