TensorRT Model Optimizer
Getting Started
Overview
Installation
Quick Start: Quantization
Quick Start: Quantization (Windows)
Quick Start: Pruning
Quick Start: Distillation
Quick Start: Sparsity
Guides
Quantization
Pruning
NAS
Distillation
Sparsity
Saving & Restoring
Speculative Decoding
Deployment
TensorRT-LLM Deployment
DirectML Deployment
Examples
All GitHub Examples
ResNet20 on CIFAR-10: Pruning
HF BERT: Prune, Distill & Quantize
Reference
Changelog
Model Optimizer Changelog (Linux)
Model Optimizer Changelog (Windows)
modelopt API
Support
Contact us
FAQs
TensorRT Model Optimizer
Changelog
View page source
Changelog
Model Optimizer Changelog (Linux)
Model Optimizer Changelog (Windows)