Model Optimizer
Getting Started
Overview
Installation
Quick Start: PTQ - PyTorch
Quick Start: PTQ - ONNX
Quick Start: PTQ - PyTorch to ONNX
Quick Start: PTQ - Windows
Quick Start: QAT
Quick Start: Pruning
Quick Start: Distillation
Quick Start: Speculative Decoding
Quick Start: Sparsity
Guides
Support Matrix
Quantization
Saving & Restoring
Pruning
Distillation
Speculative Decoding
Sparsity
NAS
AutoCast (ONNX)
Deployment
TensorRT-LLM
Onnxruntime
Unified HuggingFace Checkpoint
Examples
All GitHub Examples
Reference
Changelog
modelopt API
deploy
onnx
torch
distill
export
convert_hf_config
diffusers_utils
distribute
hf_config_map
layer_utils
mcore_config_map
model_config
model_config_export
model_config_utils
model_utils
plugins
postprocess
quant_utils
tensorrt_llm_type
tensorrt_llm_utils
transformer_engine
unified_export_hf
unified_export_megatron
nas
opt
peft
prune
quantization
sparsity
speculative
trace
utils
Security Considerations
Support
Contact us
FAQs
Model Optimizer
modelopt API
torch
export
plugins
View page source
plugins
Modules
Export package plugin.