Release Notes#
All published functionality in the Release Notes has been fully tested and verified with known limitations documented. To share feedback about this release, access our NVIDIA Developer Forum.
TensorRT-LLM Release 0.20.0#
Key Features and Enhancements#
Model Support
Added Qwen3 support.Refer to “Qwen3” section in
examples/models/core/qwen/README.md.Added HyperCLOVAX-SEED-Vision support in PyTorch flow. Refer to
examples/models/contrib/hyperclovax/README.mdAdded Dynasor-CoT in scaffolding examples. Refer to
examples/scaffolding/contrib/Dynasor/README.mdAdded Mistral Small 3.1 24B VLM support in TRT workflow
Added Gemma3-1b-it support in PyTorch workflow
Added Nemotron-H model support
Added Eagle-3 support for LLAMA4
PyTorch workflow
Added lora support
Added return logits support
Adopt new logprob definition in PyTorch flow
Enabled per-request stats with PyTorch backend
Enabled LogitsProcessor in PyTorch backend
Benchmark:
Add beam width to low latency.
Fix trtllm-bench iter_stats and cuda_graph_batch_sizes errors.
Remove deprecated Python runtime benchmark
Add benchmark support for scaffolding
Multimodal models
Added support in trtllm-serve
Added support in trtllm-bench, the support is limited to image only for now
Supported DeepSeek-R1 W4A8 on Hopper
Add the RTX Pro 6000 support on single GPU
Integrated Llama4 input processor
Added CGA reduction FHMA kernels on Blackwell
Enabled chunked context for FlashInfer
Supported KV cache reuse for MLA
Added Piecewise CUDA Graph support
Supported multiple LoRA adapters and TP
Added KV cache-aware router for disaggregated serving
Unfused attention for native support
Added group_rms_norm kernel to normalize multiple inputs in a single operator
Added smart router for the MoE module
Added head size 72 support for QKV preprocessing kernel
Added MNNVL MoE A2A support
Optimized Large Embedding Tables in Multimodal Models
Supported Top-K logprobs and prompt_logprobs in LLMAPI
Enabled overlap scheduler in TRT workflow via executor API
Infrastructure Changes#
TRT-LLM team formally releases docker image on NGC.
The pre-built TensorRT-LLM wheel on PyPI is linked against PyTorch 2.7.0 now, which uses the CXX11 ABI
The dependent TensorRT version is updated to 10.10.0
The dependent CUDA version is updated to 12.9.0
The dependent public PyTorch version is updated to 2.7.0
The dependent NVIDIA ModelOpt version is updated to 0.29.0
The dependent NCCL version is maintained at 2.25.1
Open-sourced XQA kernels
Dependent datasets version was upgraded to 3.1.0
Migrate Triton Backend to TensorRT LLM repo to TensorRT LLM submodule
Downgrade gcc toolset version from 13 to 11
API Changes#
[Breaking Change]:Enable scheduling overlap by default
Remove deprecated GptSession/V1 from TRT workflow
Set _AutoDeployLlmArgs as primary config object
Allow overriding CLI arguments with YAML file in trtllm-serve
Introduced multimodal embedding field in LlmRequest
Fixed Issues#
Fix hang bug when context server doesn’t have enough capacity for KV Cache (#3095)
Fix C++ decoder synchronization in PyTorch (#3106)
Fix bug of create cuda stream as default parameter which will be initialized during importing (#3764)
Fix bug related to creating CUDA stream as default parameter, which will be initialized during importing (#3764)
Fix attention DP bug on Qwen3 MoE model (#4141)
Fix illegal memory access when running LLaMA 4 with CUDA Graph enabled (#4101)
Reset planned states to avoid memory leak in TrtllmAttentionWrapper (#4227)
Known Issues#
multi-GPU model support on RTX Pro 6000
TensorRT-LLM Release 0.19.0#
Key Features and Enhancements#
The C++ runtime is now open sourced.
PyTorch workflow
Added DeepSeek V3/R1 support. Refer to
examples/deepseek_v3/README.md, also to the blogdocs/source/blogs/Best_perf_practice_on_DeepSeek-R1_in_TensorRT-LLM.md.Added Llava-Next support.
Added BERT support.
Added a C++ based decoder, which added support for:
TopK / TopP.
Bad words.
Stop words.
Embedding bias.
Added Autotuner for custom-op-compatible tuning process.
Added a Python-based Autotuner core framework for kernel tuning.
Applied the Autotuner to fused MoE and NVFP4 linear operators for concept and performance evaluations.
Added guided decoding support (XGrammar integration).
Added pipeline parallelism support for the overlap scheduler in
PyExecutor.Added Qwen2VL model support.
Added mixed precision quantization support.
Added pipeline parallelism with attention DP support.
Added no-cache attention support.
Added
PeftCacheManagersupport.Added Qwen2.5‑VL support and refactored Qwen2‑VL.
Added trtllm‑gen FP4 GEMM support.
Added Qwen2 MoE support.
Applied
AutoTunerto both Fused MoE and NVFP4 Linear operators.Introduced a
UserBuffersallocator.Added Deepseek eager mode AllReduce fusion support.
Added Multi-Token Prediction (MTP) support. Refer to the “Multi-Token Prediction (MTP)” section of
examples/deepseek_v3/README.md.Added FlashMLA support for SM90.
Added support for enabling MTP with CUDA graph padding.
Added initial EAGLE-3 implementation.
Added support for FP8 MLA on NVIDIA Hopper and Blackwell GPUs.
AutoDeploy for PyTorch workflow.
The AutoDeploy for PyTorch workflow is an experimental feature in
tensorrt_llm._torch.auto_deploy.AutoDeploy provides an automated path from off-the-shelf models to optimized deployment in the TensorRT-LLM runtime.
Check out
examples/auto_deploy/README.mdfor more details.
LLM API
[BREAKING CHANGE] Added dynamic logits processor support, and deprecated static logits processor.
Added batched logits processor support.
Added EAGLE support.
Added abort request support.
Added
get_statssupport.Added multi-node support for Slurm-based clusters, refer to
examples/llm-api/llm_mgmn_*.sh.
Added InternLM-XComposer2 support. Refer to “InternLM-XComposer2” section in
examples/multimodal/README.md.Added INT4-AWQ support for MoE models. Refer to the “AWQ Quantization” section in
examples/mixtral/README.md.Added Qwen2-Audio support. Refer to
examples/qwen2audio/README.md.Added Language-Adapter support. Refer to
examples/language_adapter/README.md.Added STDiT for OpenSoRA text-to-video support. Refer to
examples/stdit/README.md.Added vision encoders with tensor parallelism and context parallelism support. Refer to
examples/vit/README.md.Added EXAONE-Deep support. Refer to
examples/exaone/README.md.Added support for Phi-4-mini and Phi‑4‑MM.
Added Gemma3 text‑only model support. Refer to “Run Gemma 3” section at
examples/gemma/README.md.Added FP8 quantization support for Qwen2-VL.
Added batched inference support for the LLM API MMLU example
examples/mmlu_llmapi.py.Added FP4 quantization-layernorm fusion plugin support. (Llama models only)
Added Mamba-Hybrid support.
Added NVILA video support. The support includes 1 prompt - N media and N prompt - N media batching modes.
Added a
--quantize_lm_headoptionexamples/quantization/quantize.pyto supportlm_headquantization.Added batched tensor FP4 quantization support.
Added a
/metricsendpoint fortrtllm-serveto log iteration statistics.Added LoRA support for Phi-2 model.
Added returning context logits support for
trtllm-serve.Added one-shot version for UserBuffer AllReduce-Normalization on FP16/BF16.
Added request BW metric measurement for
disaggServerBenchmark.Updated logits bitmask kernel to v3.
Enabled CUDA graphs when attention DP was used and active requests on different GPUs were uneven.
Added iteration log support for
trtllm-bench.fp8_blockscale_gemmis now open-sourced.Added AWQ support for ModelOpt checkpoints.
Added Linear block scale layout support in FP4 quantization.
Added pre-quantized FP8 checkpoint support for Nemotron-mini-4b-instruct.
Added Variable-Beam-Width-Search (VBWS) support (part2).
Added LoRA support for Gemma.
Refactored scaffolding worker, added OpenAI API worker support.
Optionally split MoE inputs into chunks to reduce GPU memory usage.
Added UCX IP interface support.
[BREAKING CHANGE] Added output of first token to additional generation outputs.
Added FP8 support for SM120 architecture.
Registered
ENABLE_MULTI_DEVICEandENABLE_UCXas CMake options.Made the scaffolding Controller more generic.
Breaking change: Added individual gatherContext support for each additional output.
Enabled
PyExecutorinference flow to estimatemax_num_tokensforkv_cache_manager.Added
TLLM_OVERRIDE_LAYER_NUMandTLLM_TRACE_MODEL_FORWARDenvironment variables for debugging.Supported aborting disconnected requests.
Added an option to run disaggregated serving without context servers.
Fixed and improved allreduce and fusion kernels.
Enhanced the integrated robustness of scaffolding via
init.py.
API Changes#
Exposed
kc_cache_retention_configfrom C++executorAPI to the LLM API.Moved
BuildConfigarguments toLlmArgs.Removed speculative decoding parameters from stateful decoders.
Exposed
DecoderStatevia bindings and integrated it in decoder.Refactored the
LlmArgswithPydanticand migrated remaining pybinding configurations to Python.Refactored disaggregated serving scripts.
Added
numNodestoParallelConfig.Redesigned the multi‑stream API for DeepSeek.
Fixed Issues#
Fixed misused length argument of PluginField. Thanks to the contribution from @jl749 in #2712. This also fixes #2685.
Fixed a Llama-3.2 SmoothQuant convert checkpoint issue. (#2677)
Fixed a bug when loading an engine using LoRA through the LLM API. (#2782)
Fixed incorrect batch slot usage in
addCumLogProbskernel. Thanks to the contribution from @aotman in #2787.Fixed incorrect output for Llama-3.2-11B-Vision-Instruct. (#2796)
Removed the necessary of
--extra-index-url https://pypi.nvidia.comwhen runningpip install tensorrt-llm.
Infrastructure Changes#
The dependent NVIDIA ModelOpt version is updated to 0.27.
Known Issues#
The PyTorch workflow on SBSA is incompatible with bare metal environments like Ubuntu 24.04. Please use the PyTorch NGC Container for optimal support on SBSA platforms.
TensorRT-LLM Release 0.18.2#
Key Features and Enhancements#
This update addresses known security issues. For the latest NVIDIA Vulnerability Disclosure Information visit https://www.nvidia.com/en-us/security/.
TensorRT-LLM Release 0.18.1#
Key Features and Enhancements#
The 0.18.x series of releases builds upon the 0.17.0 release, focusing exclusively on dependency updates without incorporating features from the previous 0.18.0.dev pre-releases. These features will be included in future stable releases.
Infrastructure Changes#
The dependent
transformerspackage version is updated to 4.48.3.
TensorRT-LLM Release 0.18.0#
Key Features and Enhancements#
Features that were previously available in the 0.18.0.dev pre-releases are not included in this release.
[BREAKING CHANGE] Windows platform support is deprecated as of v0.18.0. All Windows-related code and functionality will be completely removed in future releases.
Known Issues#
The PyTorch workflow on SBSA is incompatible with bare metal environments like Ubuntu 24.04. Please use the PyTorch NGC Container for optimal support on SBSA platforms.
Infrastructure Changes#
The base Docker image for TensorRT-LLM is updated to
nvcr.io/nvidia/pytorch:25.03-py3.The base Docker image for TensorRT-LLM Backend is updated to
nvcr.io/nvidia/tritonserver:25.03-py3.The dependent TensorRT version is updated to 10.9.
The dependent CUDA version is updated to 12.8.1.
The dependent NVIDIA ModelOpt version is updated to 0.25 for Linux platform.
TensorRT-LLM Release 0.17.0#
Key Features and Enhancements#
Blackwell support
NOTE: pip installation is not supported for TRT-LLM 0.17 on Blackwell platforms only. Instead, it is recommended that the user build from source using NVIDIA NGC 25.01 PyTorch container.
Added support for B200.
Added support for GeForce RTX 50 series using Windows Subsystem for Linux (WSL) for limited models.
Added NVFP4 Gemm support for Llama and Mixtral models.
Added NVFP4 support for the
LLMAPI andtrtllm-benchcommand.GB200 NVL is not fully supported.
Added benchmark script to measure perf benefits of KV cache host offload with expected runtime improvements from GH200.
PyTorch workflow
The PyTorch workflow is an experimental feature in
tensorrt_llm._torch. The following is a list of supported infrastructure, models, and features that can be used with the PyTorch workflow.Added support for H100/H200/B200.
Added support for Llama models, Mixtral, QWen, Vila.
Added support for FP16/BF16/FP8/NVFP4 Gemm and fused Mixture-Of-Experts (MOE), FP16/BF16/FP8 KVCache.
Added custom context and decoding attention kernels support via PyTorch custom op.
Added support for chunked context (default off).
Added CudaGraph support for decoding only.
Added overlap scheduler support to overlap prepare inputs and model forward by decoding 1 extra token.
Added FP8 context FMHA support for the W4A8 quantization workflow.
Added ModelOpt quantized checkpoint support for the
LLMAPI.Added FP8 support for the Llama-3.2 VLM model. Refer to the “MLLaMA” section in
examples/multimodal/README.md.Added PDL support for
userbufferbased AllReduce-Norm fusion kernel.Added runtime support for seamless lookahead decoding.
Added token-aligned arbitrary output tensors support for the C++
executorAPI.
API Changes#
[BREAKING CHANGE] KV cache reuse is enabled automatically when
paged_context_fmhais enabled.Added
--concurrencysupport for thethroughputsubcommand oftrtllm-bench.
Known Issues#
Need
--extra-index-url https://pypi.nvidia.comwhen runningpip install tensorrt-llmdue to new third-party dependencies.The PYPI SBSA wheel is incompatible with PyTorch 2.5.1 due to a break in the PyTorch ABI/API, as detailed in the related GitHub issue.
The PyTorch workflow on SBSA is incompatible with bare metal environments like Ubuntu 24.04. Please use the PyTorch NGC Container for optimal support on SBSA platforms.
Fixed Issues#
Fixed incorrect LoRA output dimension. Thanks for the contribution from @akhoroshev in #2484.
Added NVIDIA H200 GPU into the
cluster_keyfor auto parallelism feature. (#2552)Fixed a typo in the
__post_init__function ofLLmArgsClass. Thanks for the contribution from @topenkoff in #2691.Fixed workspace size issue in the GPT attention plugin. Thanks for the contribution from @AIDC-AI.
Fixed Deepseek-V2 model accuracy.
Infrastructure Changes#
The base Docker image for TensorRT-LLM is updated to
nvcr.io/nvidia/pytorch:25.01-py3.The base Docker image for TensorRT-LLM Backend is updated to
nvcr.io/nvidia/tritonserver:25.01-py3.The dependent TensorRT version is updated to 10.8.0.
The dependent CUDA version is updated to 12.8.0.
The dependent ModelOpt version is updated to 0.23 for Linux platform, while 0.17 is still used on Windows platform.
TensorRT-LLM Release 0.16.0#
Key Features and Enhancements#
Added guided decoding support with XGrammar backend.
Added quantization support for RecurrentGemma. Refer to
examples/recurrentgemma/README.md.Added ulysses context parallel support. Refer to an example on building LLaMA 7B using 2-way tensor parallelism and 2-way context parallelism at
examples/llama/README.md.Added W4A8 quantization support to BF16 models on Ada (SM89).
Added PDL support for the FP8 GEMM plugins.
Added a runtime
max_num_tokensdynamic tuning feature, which can be enabled by setting--enable_max_num_tokens_tuningtogptManagerBenchmark.Added typical acceptance support for EAGLE.
Supported chunked context and sliding window attention to be enabled together.
Added head size 64 support for the XQA kernel.
Added the following features to the LLM API:
Lookahead decoding.
DeepSeek V1 support.
Medusa support.
max_num_tokensandmax_batch_sizearguments to control the runtime parameters.extended_runtime_perf_knob_configto enable various performance configurations.
Added LogN scaling support for Qwen models.
Added
AutoAWQcheckpoints support for Qwen. Refer to the “INT4-AWQ” section inexamples/qwen/README.md.Added
AutoAWQandAutoGPTQHugging Face checkpoints support for LLaMA. (#2458)Added
allottedTimeMsto the C++Requestclass to support per-request timeout.[BREAKING CHANGE] Removed NVIDIA V100 GPU support.
API Changes#
[BREAKING CHANGE] Removed
enable_xqaargument fromtrtllm-build.[BREAKING CHANGE] Chunked context is enabled by default when KV cache and paged context FMHA is enabled on non-RNN based models.
[BREAKING CHANGE] Enabled embedding sharing automatically when possible and remove the flag
--use_embedding_sharingfrom convert checkpoints scripts.[BREAKING CHANGE] The
if __name__ == "__main__"entry point is required for both single-GPU and multi-GPU cases when using theLLMAPI.[BREAKING CHANGE] Cancelled requests now return empty results.
Added the
enable_chunked_prefillflag to theLlmArgsof theLLMAPI.Integrated BERT and RoBERTa models to the
trtllm-buildcommand.
Model Updates#
Added Qwen2-VL support. Refer to the “Qwen2-VL” section of
examples/multimodal/README.md.Added multimodal evaluation examples. Refer to
examples/multimodal.Added Stable Diffusion XL support. Refer to
examples/sdxl/README.md. Thanks for the contribution from @Zars19 in #1514.
Fixed Issues#
Fixed unnecessary batch logits post processor calls. (#2439)
Fixed a typo in the error message. (#2473)
Fixed the in-place clamp operation usage in smooth quant. Thanks for the contribution from @StarrickLiu in #2485.
Fixed
sampling_paramsto only be setup ifend_idis None andtokenizeris not None in theLLMAPI. Thanks to the contribution from @mfuntowicz in #2573.
Infrastructure Changes#
Updated the base Docker image for TensorRT-LLM to
nvcr.io/nvidia/pytorch:24.11-py3.Updated the base Docker image for TensorRT-LLM Backend to
nvcr.io/nvidia/tritonserver:24.11-py3.Updated to TensorRT v10.7.
Updated to CUDA v12.6.3.
Added support for Python 3.10 and 3.12 to TensorRT-LLM Python wheels on PyPI.
Updated to ModelOpt v0.21 for Linux platform, while v0.17 is still used on Windows platform.
Known Issues#
There is a known AllReduce performance issue on AMD-based CPU platforms on NCCL 2.23.4, which can be workarounded by
export NCCL_P2P_LEVEL=SYS.
TensorRT-LLM Release 0.15.0#
Key Features and Enhancements#
Added support for EAGLE. Refer to
examples/eagle/README.md.Added functional support for GH200 systems.
Added AutoQ (mixed precision) support.
Added a
trtllm-servecommand to start a FastAPI based server.Added FP8 support for Nemotron NAS 51B. Refer to
examples/nemotron_nas/README.md.Added INT8 support for GPTQ quantization.
Added TensorRT native support for INT8 Smooth Quantization.
Added quantization support for Exaone model. Refer to
examples/exaone/README.md.Enabled Medusa for Qwen2 models. Refer to “Medusa with Qwen2” section in
examples/medusa/README.md.Optimized pipeline parallelism with ReduceScatter and AllGather for Mixtral models.
Added support for
Qwen2ForSequenceClassificationmodel architecture.Added Python plugin support to simplify plugin development efforts. Refer to
examples/python_plugin/README.md.Added different rank dimensions support for LoRA modules when using the Hugging Face format. Thanks for the contribution from @AlessioNetti in #2366.
Enabled embedding sharing by default. Refer to “Embedding Parallelism, Embedding Sharing, and Look-Up Plugin” section in
docs/source/performance/perf-best-practices.mdfor information about the required conditions for embedding sharing.Added support for per-token per-channel FP8 (namely row-wise FP8) on Ada.
Extended the maximum supported
beam_widthto256.Added FP8 and INT8 SmoothQuant quantization support for the InternVL2-4B variant (LLM model only). Refer to
examples/multimodal/README.md.Added support for prompt-lookup speculative decoding. Refer to
examples/prompt_lookup/README.md.Integrated the QServe w4a8 per-group/per-channel quantization. Refer to “w4aINT8 quantization (QServe)” section in
examples/llama/README.md.Added a C++ example for fast logits using the
executorAPI. Refer to “executorExampleFastLogits” section inexamples/cpp/executor/README.md.[BREAKING CHANGE] NVIDIA Volta GPU support is removed in this and future releases.
Added the following enhancements to the LLM API:
[BREAKING CHANGE] Moved the runtime initialization from the first invocation of
LLM.generatetoLLM.__init__for better generation performance without warmup.Added
nandbest_ofarguments to theSamplingParamsclass. These arguments enable returning multiple generations for a single request.Added
ignore_eos,detokenize,skip_special_tokens,spaces_between_special_tokens, andtruncate_prompt_tokensarguments to theSamplingParamsclass. These arguments enable more control over the tokenizer behavior.Added support for incremental detokenization to improve the detokenization performance for streaming generation.
Added the
enable_prompt_adapterargument to theLLMclass and theprompt_adapter_requestargument for theLLM.generatemethod. These arguments enable prompt tuning.
Added support for a
gpt_variantargument to theexamples/gpt/convert_checkpoint.pyfile. This enhancement enables checkpoint conversion with more GPT model variants. Thanks to the contribution from @tonylek in #2352.
API Changes#
[BREAKING CHANGE] Moved the flag
builder_force_num_profilesintrtllm-buildcommand to theBUILDER_FORCE_NUM_PROFILESenvironment variable.[BREAKING CHANGE] Modified defaults for
BuildConfigclass so that they are aligned with thetrtllm-buildcommand.[BREAKING CHANGE] Removed Python bindings of
GptManager.[BREAKING CHANGE]
autois used as the default value for--dtypeoption in quantize and checkpoints conversion scripts.[BREAKING CHANGE] Deprecated
gptManagerAPI path ingptManagerBenchmark.[BREAKING CHANGE] Deprecated the
beam_widthandnum_return_sequencesarguments to theSamplingParamsclass in the LLM API. Use then,best_ofanduse_beam_searcharguments instead.Exposed
--trust_remote_codeargument to the OpenAI API server. (#2357)
Model Updates#
Added support for Llama 3.2 and llama 3.2-Vision model. Refer to
examples/mllama/README.mdfor more details on the llama 3.2-Vision model.Added support for Deepseek-v2. Refer to
examples/deepseek_v2/README.md.Added support for Cohere Command R models. Refer to
examples/commandr/README.md.Added support for Falcon 2, refer to
examples/falcon/README.md, thanks to the contribution from @puneeshkhanna in #1926.Added support for InternVL2. Refer to
examples/multimodal/README.md.Added support for Qwen2-0.5B and Qwen2.5-1.5B model. (#2388)
Added support for Minitron. Refer to
examples/nemotron.Added a GPT Variant - Granite(20B and 34B). Refer to “GPT Variant - Granite” section in
examples/gpt/README.md.Added support for LLaVA-OneVision model. Refer to “LLaVA, LLaVa-NeXT, LLaVA-OneVision and VILA” section in
examples/multimodal/README.md.
Fixed Issues#
Fixed a slice error in forward function. (#1480)
Fixed an issue that appears when building BERT. (#2373)
Fixed an issue that model is not loaded when building BERT. (2379)
Fixed the broken executor examples. (#2294)
Fixed the issue that the kernel
moeTopK()cannot find the correct expert when the number of experts is not a power of two. Thanks @dongjiyingdjy for reporting this bug.Fixed an assertion failure on
crossKvCacheFraction. (#2419)Fixed an issue when using smoothquant to quantize Qwen2 model. (#2370)
Fixed a PDL typo in
docs/source/performance/perf-benchmarking.md, thanks @MARD1NO for pointing it out in #2425.
Infrastructure Changes#
The base Docker image for TensorRT-LLM is updated to
nvcr.io/nvidia/pytorch:24.10-py3.The base Docker image for TensorRT-LLM Backend is updated to
nvcr.io/nvidia/tritonserver:24.10-py3.The dependent TensorRT version is updated to 10.6.
The dependent CUDA version is updated to 12.6.2.
The dependent PyTorch version is updated to 2.5.1.
The dependent ModelOpt version is updated to 0.19 for Linux platform, while 0.17 is still used on Windows platform.
Documentation#
Added a copy button for code snippets in the documentation. (#2288)
TensorRT-LLM Release 0.14.0#
Key Features and Enhancements#
Enhanced the
LLMclass in the LLM API.Added support for calibration with offline dataset.
Added support for Mamba2.
Added support for
finish_reasonandstop_reason.
Added FP8 support for CodeLlama.
Added
__repr__methods for classModule, thanks to the contribution from @1ytic in #2191.Added BFloat16 support for fused gated MLP.
Updated ReDrafter beam search logic to match Apple ReDrafter v1.1.
Improved
customAllReduceperformance.Draft model now can copy logits directly over MPI to the target model’s process in
orchestratormode. This fast logits copy reduces the delay between draft token generation and the beginning of target model inference.NVIDIA Volta GPU support is deprecated and will be removed in a future release.
API Changes#
[BREAKING CHANGE] The default
max_batch_sizeof thetrtllm-buildcommand is set to2048.[BREAKING CHANGE] Remove
builder_optfrom theBuildConfigclass and thetrtllm-buildcommand.Add logits post-processor support to the
ModelRunnerCppclass.Added
isParticipantmethod to the C++ExecutorAPI to check if the current process is a participant in the executor instance.
Model Updates#
Added support for NemotronNas, see
examples/nemotron_nas/README.md.Added support for Deepseek-v1, see
examples/deepseek_v1/README.md.Added support for Phi-3.5 models, see
examples/phi/README.md.
Fixed Issues#
Fixed a typo in
tensorrt_llm/models/model_weights_loader.py, thanks to the contribution from @wangkuiyi in #2152.Fixed duplicated import module in
tensorrt_llm/runtime/generation.py, thanks to the contribution from @lkm2835 in #2182.Enabled
share_embeddingfor the models that have nolm_headin legacy checkpoint conversion path, thanks to the contribution from @lkm2835 in #2232.Fixed
kv_cache_typeissue in the Python benchmark, thanks to the contribution from @qingquansong in #2219.Fixed an issue with SmoothQuant calibration with custom datasets. Thanks to the contribution by @Bhuvanesh09 in #2243.
Fixed an issue surrounding
trtllm-build --fast-buildwith fake or random weights. Thanks to @ZJLi2013 for flagging it in #2135.Fixed missing
use_fused_mlpwhen constructingBuildConfigfrom dict, thanks for the fix from @ethnzhng in #2081.Fixed lookahead batch layout for
numNewTokensCumSum. (#2263)
Infrastructure Changes#
The dependent ModelOpt version is updated to v0.17.
Documentation#
@Sherlock113 added a tech blog to the latest news in #2169, thanks for the contribution.
Known Issues#
Replit Code is not supported with the transformers 4.45+
TensorRT-LLM Release 0.13.0#
Key Features and Enhancements#
Supported lookahead decoding (experimental), see
docs/source/speculative_decoding.md.Added some enhancements to the
ModelWeightsLoader(a unified checkpoint converter, seedocs/source/architecture/model-weights-loader.md).Supported Qwen models.
Supported auto-padding for indivisible TP shape in INT4-wo/INT8-wo/INT4-GPTQ.
Improved performance on
*.binand*.pth.
Supported OpenAI Whisper in C++ runtime.
Added some enhancements to the
LLMclass.Supported LoRA.
Supported engine building using dummy weights.
Supported
trust_remote_codefor customized models and tokenizers downloaded from Hugging Face Hub.
Supported beam search for streaming mode.
Supported tensor parallelism for Mamba2.
Supported returning generation logits for streaming mode.
Added
curandandbfloat16support forReDrafter.Added sparse mixer normalization mode for MoE models.
Added support for QKV scaling in FP8 FMHA.
Supported FP8 for MoE LoRA.
Supported KV cache reuse for P-Tuning and LoRA.
Supported in-flight batching for CogVLM models.
Supported LoRA for the
ModelRunnerCppclass.Supported
head_size=48cases for FMHA kernels.Added FP8 examples for DiT models, see
examples/dit/README.md.Supported decoder with encoder input features for the C++
executorAPI.
API Changes#
[BREAKING CHANGE] Set
use_fused_mlptoTrueby default.[BREAKING CHANGE] Enabled
multi_block_modeby default.[BREAKING CHANGE] Enabled
strongly_typedby default inbuilderAPI.[BREAKING CHANGE] Renamed
maxNewTokens,randomSeedandminLengthtomaxTokens,seedandminTokensfollowing OpenAI style.The
LLMclass[BREAKING CHANGE] Updated
LLM.generatearguments to includePromptInputsandtqdm.
The C++
executorAPI[BREAKING CHANGE] Added
LogitsPostProcessorConfig.Added
FinishReasontoResult.
Model Updates#
Supported Gemma 2, see “Run Gemma 2” section in
examples/gemma/README.md.
Fixed Issues#
Fixed an accuracy issue when enabling remove padding issue for cross attention. (#1999)
Fixed the failure in converting qwen2-0.5b-instruct when using
smoothquant. (#2087)Matched the
exclude_modulespattern inconvert_utils.pyto the changes inquantize.py. (#2113)Fixed build engine error when
FORCE_NCCL_ALL_REDUCE_STRATEGYis set.Fixed unexpected truncation in the quant mode of
gpt_attention.Fixed the hang caused by race condition when canceling requests.
Fixed the default factory for
LoraConfig. (#1323)
Infrastructure Changes#
Base Docker image for TensorRT-LLM is updated to
nvcr.io/nvidia/pytorch:24.07-py3.Base Docker image for TensorRT-LLM Backend is updated to
nvcr.io/nvidia/tritonserver:24.07-py3.The dependent TensorRT version is updated to 10.4.0.
The dependent CUDA version is updated to 12.5.1.
The dependent PyTorch version is updated to 2.4.0.
The dependent ModelOpt version is updated to v0.15.
TensorRT-LLM Release 0.12.0#
Key Features and Enhancements#
Supported LoRA for MoE models.
The
ModelWeightsLoaderis enabled for LLaMA family models (experimental), seedocs/source/architecture/model-weights-loader.md.Supported FP8 FMHA for NVIDIA Ada Lovelace Architecture.
Supported GPT-J, Phi, Phi-3, Qwen, GPT, GLM, Baichuan, Falcon and Gemma models for the
LLMclass.Supported FP8 OOTB MoE.
Supported Starcoder2 SmoothQuant. (#1886)
Supported ReDrafter Speculative Decoding, see “ReDrafter” section in
docs/source/speculative_decoding.md.Supported padding removal for BERT, thanks to the contribution from @Altair-Alpha in #1834.
Added in-flight batching support for GLM 10B model.
Supported
gelu_pytorch_tanhactivation function, thanks to the contribution from @ttim in #1897.Added
chunk_lengthparameter to Whisper, thanks to the contribution from @MahmoudAshraf97 in #1909.Added
concurrencyargument forgptManagerBenchmark.Executor API supports requests with different beam widths, see
docs/source/executor.md#sending-requests-with-different-beam-widths.Added the flag
--fast_buildtotrtllm-buildcommand (experimental).
API Changes#
[BREAKING CHANGE]
max_output_lenis removed fromtrtllm-buildcommand, if you want to limit sequence length on engine build stage, specifymax_seq_len.[BREAKING CHANGE] The
use_custom_all_reduceargument is removed fromtrtllm-build.[BREAKING CHANGE] The
multi_block_modeargument is moved from build stage (trtllm-buildand builder API) to the runtime.[BREAKING CHANGE] The build time argument
context_fmha_fp32_accis moved to runtime for decoder models.[BREAKING CHANGE] The arguments
tp_size,pp_sizeandcp_sizeis removed fromtrtllm-buildcommand.The C++ batch manager API is deprecated in favor of the C++
executorAPI, and it will be removed in a future release of TensorRT-LLM.Added a version API to the C++ library, a
cpp/include/tensorrt_llm/executor/version.hfile is going to be generated.
Model Updates#
Supported LLaMA 3.1 model.
Supported Mamba-2 model.
Supported EXAONE model, see
examples/exaone/README.md.Supported Qwen 2 model.
Supported GLM4 models, see
examples/chatglm/README.md.Added LLaVa-1.6 (LLaVa-NeXT) multimodal support, see “LLaVA, LLaVa-NeXT and VILA” section in
examples/multimodal/README.md.
Fixed Issues#
Fixed wrong pad token for the CodeQwen models. (#1953)
Fixed typo in
cluster_infosdefined intensorrt_llm/auto_parallel/cluster_info.py, thanks to the contribution from @saeyoonoh in #1987.Removed duplicated flags in the command at
docs/source/reference/troubleshooting.md, thanks for the contribution from @hattizai in #1937.Fixed segmentation fault in TopP sampling layer, thanks to the contribution from @akhoroshev in #2039. (#2040)
Fixed the failure when converting the checkpoint for Mistral Nemo model. (#1985)
Propagated
exclude_modulesto weight-only quantization, thanks to the contribution from @fjosw in #2056.Fixed wrong links in README, thanks to the contribution from @Tayef-Shah in #2028.
Fixed some typos in the documentation, thanks to the contribution from @lfz941 in #1939.
Fixed the engine build failure when deduced
max_seq_lenis not an integer. (#2018)
Infrastructure Changes#
Base Docker image for TensorRT-LLM is updated to
nvcr.io/nvidia/pytorch:24.07-py3.Base Docker image for TensorRT-LLM Backend is updated to
nvcr.io/nvidia/tritonserver:24.07-py3.The dependent TensorRT version is updated to 10.3.0.
The dependent CUDA version is updated to 12.5.1.
The dependent PyTorch version is updated to 2.4.0.
The dependent ModelOpt version is updated to v0.15.0.
Known Issues#
On Windows, installation of TensorRT-LLM may succeed, but you might hit
OSError: exception: access violation reading 0x0000000000000000when importing the library in Python. See Installing on Windows for workarounds.
TensorRT-LLM Release 0.11.0#
Key Features and Enhancements#
Supported very long context for LLaMA (see “Long context evaluation” section in
examples/llama/README.md).Low latency optimization
Added a reduce-norm feature which aims to fuse the ResidualAdd and LayerNorm kernels after AllReduce into a single kernel, which is recommended to be enabled when the batch size is small and the generation phase time is dominant.
Added FP8 support to the GEMM plugin, which benefits the cases when batch size is smaller than 4.
Added a fused GEMM-SwiGLU plugin for FP8 on SM90.
LoRA enhancements
Supported running FP8 LLaMA with FP16 LoRA checkpoints.
Added support for quantized base model and FP16/BF16 LoRA.
SQ OOTB (- INT8 A/W) + FP16/BF16/FP32 LoRA
INT8/ INT4 Weight-Only (INT8 /W) + FP16/BF16/FP32 LoRA
Weight-Only Group-wise + FP16/BF16/FP32 LoRA
Added LoRA support to Qwen2, see “Run models with LoRA” section in
examples/qwen/README.md.Added support for Phi-3-mini/small FP8 base + FP16/BF16 LoRA, see “Run Phi-3 with LoRA” section in
examples/phi/README.md.Added support for starcoder-v2 FP8 base + FP16/BF16 LoRA, see “Run StarCoder2 with LoRA” section in
examples/gpt/README.md.
Encoder-decoder models C++ runtime enhancements
Supported paged KV cache and inflight batching. (#800)
Supported tensor parallelism.
Supported INT8 quantization with embedding layer excluded.
Updated default model for Whisper to
distil-whisper/distil-large-v3, thanks to the contribution from @IbrahimAmin1 in #1337.Supported HuggingFace model automatically download for the Python high level API.
Supported explicit draft tokens for in-flight batching.
Supported local custom calibration datasets, thanks to the contribution from @DreamGenX in #1762.
Added batched logits post processor.
Added Hopper qgmma kernel to XQA JIT codepath.
Supported tensor parallelism and expert parallelism enabled together for MoE.
Supported the pipeline parallelism cases when the number of layers cannot be divided by PP size.
Added
numQueuedRequeststo the iteration stats log of the executor API.Added
iterLatencyMilliSecto the iteration stats log of the executor API.Add HuggingFace model zoo from the community, thanks to the contribution from @matichon-vultureprime in #1674.
API Changes#
[BREAKING CHANGE]
trtllm-buildcommandMigrated Whisper to unified workflow (
trtllm-buildcommand), see documents: examples/whisper/README.md.max_batch_sizeintrtllm-buildcommand is switched to 256 by default.max_num_tokensintrtllm-buildcommand is switched to 8192 by default.Deprecated
max_output_lenand addedmax_seq_len.Removed unnecessary
--weight_only_precisionargument fromtrtllm-buildcommand.Removed
attention_qk_half_accumulationargument fromtrtllm-buildcommand.Removed
use_context_fmha_for_generationargument fromtrtllm-buildcommand.Removed
strongly_typedargument fromtrtllm-buildcommand.The default value of
max_seq_lenreads from the HuggingFace mode config now.
C++ runtime
[BREAKING CHANGE] Renamed
free_gpu_memory_fractioninModelRunnerCpptokv_cache_free_gpu_memory_fraction.[BREAKING CHANGE] Refactored
GptManagerAPIMoved
maxBeamWidthintoTrtGptModelOptionalParams.Moved
schedulerConfigintoTrtGptModelOptionalParams.
Added some more options to
ModelRunnerCpp, includingmax_tokens_in_paged_kv_cache,kv_cache_enable_block_reuseandenable_chunked_context.
[BREAKING CHANGE] Python high-level API
Removed the
ModelConfigclass, and all the options are moved toLLMclass.Refactored the
LLMclass, please refer toexamples/high-level-api/README.mdMoved the most commonly used options in the explicit arg-list, and hidden the expert options in the kwargs.
Exposed
modelto accept either HuggingFace model name or local HuggingFace model/TensorRT-LLM checkpoint/TensorRT-LLM engine.Support downloading model from HuggingFace model hub, currently only Llama variants are supported.
Support build cache to reuse the built TensorRT-LLM engines by setting environment variable
TLLM_LLMAPI_BUILD_CACHE=1or passingenable_build_cache=TruetoLLMclass.Exposed low-level options including
BuildConfig,SchedulerConfigand so on in the kwargs, ideally you should be able to configure details about the build and runtime phase.
Refactored
LLM.generate()andLLM.generate_async()API.Removed
SamplingConfig.Added
SamplingParamswith more extensive parameters, seetensorrt_llm/llmapi/utils.py.The new
SamplingParamscontains and manages fields from Python bindings ofSamplingConfig,OutputConfig, and so on.
Refactored
LLM.generate()output asRequestOutput, seetensorrt_llm/llmapi/llm.py.
Updated the
appsexamples, specially by rewriting bothchat.pyandfastapi_server.pyusing theLLMAPIs, please refer to theexamples/apps/README.mdfor details.Updated the
chat.pyto support multi-turn conversation, allowing users to chat with a model in the terminal.Fixed the
fastapi_server.pyand eliminate the need formpirunin multi-GPU scenarios.
[BREAKING CHANGE] Speculative decoding configurations unification
Introduction of
SpeculativeDecodingMode.hto choose between different speculative decoding techniques.Introduction of
SpeculativeDecodingModule.hbase class for speculative decoding techniques.Removed
decodingMode.h.
gptManagerBenchmark[BREAKING CHANGE]
apiingptManagerBenchmarkcommand isexecutorby default now.Added a runtime
max_batch_size.Added a runtime
max_num_tokens.
[BREAKING CHANGE] Added a
biasargument to theLayerNormmodule, and supports non-bias layer normalization.[BREAKING CHANGE] Removed
GptSessionPython bindings.
Model Updates#
Supported Jais, see
examples/jais/README.md.Supported DiT, see
examples/dit/README.md.Supported VILA 1.5.
Supported Video NeVA, see
Video NeVAsection inexamples/multimodal/README.md.Supported Grok-1, see
examples/grok/README.md.Supported Qwen1.5-110B with FP8 PTQ.
Supported Phi-3 small model with block sparse attention.
Supported InternLM2 7B/20B, thanks to the contribution from @RunningLeon in #1392.
Supported Phi-3-medium models, see
examples/phi/README.md.Supported Qwen1.5 MoE A2.7B.
Supported phi 3 vision multimodal.
Fixed Issues#
Fixed brokens outputs for the cases when batch size is larger than 1. (#1539)
Fixed
top_ktype inexecutor.py, thanks to the contribution from @vonjackustc in #1329.Fixed stop and bad word list pointer offset in Python runtime, thanks to the contribution from @fjosw in #1486.
Fixed some typos for Whisper model, thanks to the contribution from @Pzzzzz5142 in #1328.
Fixed export failure with CUDA driver < 526 and pynvml >= 11.5.0, thanks to the contribution from @CoderHam in #1537.
Fixed an issue in NMT weight conversion, thanks to the contribution from @Pzzzzz5142 in #1660.
Fixed LLaMA Smooth Quant conversion, thanks to the contribution from @lopuhin in #1650.
Fixed
qkv_biasshape issue for Qwen1.5-32B (#1589), thanks to the contribution from @Tlntin in #1637.Fixed the error of Ada traits for
fpA_intB, thanks to the contribution from @JamesTheZ in #1583.Update
examples/qwenvl/requirements.txt, thanks to the contribution from @ngoanpv in #1248.Fixed rsLoRA scaling in
lora_manager, thanks to the contribution from @TheCodeWrangler in #1669.Fixed Qwen1.5 checkpoint convert failure #1675.
Fixed Medusa safetensors and AWQ conversion, thanks to the contribution from @Tushar-ml in #1535.
Fixed
convert_hf_mpt_legacycall failure when the function is called in other than global scope, thanks to the contribution from @bloodeagle40234 in #1534.Fixed
use_fp8_context_fmhabroken outputs (#1539).Fixed pre-norm weight conversion for NMT models, thanks to the contribution from @Pzzzzz5142 in #1723.
Fixed random seed initialization issue, thanks to the contribution from @pathorn in #1742.
Fixed stop words and bad words in python bindings. (#1642)
Fixed the issue that when converting checkpoint for Mistral 7B v0.3, thanks to the contribution from @Ace-RR: #1732.
Fixed broken inflight batching for fp8 Llama and Mixtral, thanks to the contribution from @bprus: #1738
Fixed the failure when
quantize.pyis export data to config.json, thanks to the contribution from @janpetrov: #1676Raise error when autopp detects unsupported quant plugin #1626.
Fixed the issue that
shared_embedding_tableis not being set when loading Gemma #1799, thanks to the contribution from @mfuntowicz.Fixed stop and bad words list contiguous for
ModelRunner#1815, thanks to the contribution from @Marks101.Fixed missing comment for
FAST_BUILD, thanks to the support from @lkm2835 in #1851.Fixed the issues that Top-P sampling occasionally produces invalid tokens. #1590
Fixed #1424.
Fixed #1529.
Fixed
benchmarks/cpp/README.mdfor #1562 and #1552.Fixed dead link, thanks to the help from @DefTruth, @buvnswrn and @sunjiabin17 in: https://github.com/triton-inference-server/tensorrtllm_backend/pull/478, https://github.com/triton-inference-server/tensorrtllm_backend/pull/482 and https://github.com/triton-inference-server/tensorrtllm_backend/pull/449.
Infrastructure Changes#
Base Docker image for TensorRT-LLM is updated to
nvcr.io/nvidia/pytorch:24.05-py3.Base Docker image for TensorRT-LLM backend is updated to
nvcr.io/nvidia/tritonserver:24.05-py3.The dependent TensorRT version is updated to 10.2.0.
The dependent CUDA version is updated to 12.4.1.
The dependent PyTorch version is updated to 2.3.1.
The dependent ModelOpt version is updated to v0.13.0.
Known Issues#
In a conda environment on Windows, installation of TensorRT-LLM may succeed. However, when importing the library in Python, you may receive an error message of
OSError: exception: access violation reading 0x0000000000000000. This issue is under investigation.
TensorRT-LLM Release 0.10.0#
Announcements#
TensorRT-LLM supports TensorRT 10.0.1 and NVIDIA NGC 24.03 containers.
Key Features and Enhancements#
The Python high level API
Added embedding parallel, embedding sharing, and fused MLP support.
Enabled the usage of the
executorAPI.
Added a weight-stripping feature with a new
trtllm-refitcommand. For more information, refer toexamples/sample_weight_stripping/README.md.Added a weight-streaming feature. For more information, refer to
docs/source/advanced/weight-streaming.md.Enhanced the multiple profiles feature;
--multiple_profilesargument intrtllm-buildcommand builds more optimization profiles now for better performance.Added FP8 quantization support for Mixtral.
Added support for pipeline parallelism for GPT.
Optimized
applyBiasRopeUpdateKVCachekernel by avoiding re-computation.Reduced overheads between
enqueuecalls of TensorRT engines.Added support for paged KV cache for enc-dec models. The support is limited to beam width 1.
Added W4A(fp)8 CUTLASS kernels for the NVIDIA Ada Lovelace architecture.
Added debug options (
--visualize_networkand--dry_run) to thetrtllm-buildcommand to visualize the TensorRT network before engine build.Integrated the new NVIDIA Hopper XQA kernels for LLaMA 2 70B model.
Improved the performance of pipeline parallelism when enabling in-flight batching.
Supported quantization for Nemotron models.
Added LoRA support for Mixtral and Qwen.
Added in-flight batching support for ChatGLM models.
Added support to
ModelRunnerCppso that it runs with theexecutorAPI for IFB-compatible models.Enhanced the custom
AllReduceby adding a heuristic; fall back to use native NCCL kernel when hardware requirements are not satisfied to get the best performance.Optimized the performance of checkpoint conversion process for LLaMA.
Benchmark
[BREAKING CHANGE] Moved the request rate generation arguments and logic from prepare dataset script to
gptManagerBenchmark.Enabled streaming and support
Time To the First Token (TTFT)latency andInter-Token Latency (ITL)metrics forgptManagerBenchmark.Added the
--max_attention_windowoption togptManagerBenchmark.
API Changes#
[BREAKING CHANGE] Set the default
tokens_per_blockargument of thetrtllm-buildcommand to 64 for better performance.[BREAKING CHANGE] Migrated enc-dec models to the unified workflow.
[BREAKING CHANGE] Renamed
GptModelConfigtoModelConfig.[BREAKING CHANGE] Added speculative decoding mode to the builder API.
[BREAKING CHANGE] Refactor scheduling configurations
Unified the
SchedulerPolicywith the same name inbatch_schedulerandexecutor, and renamed it toCapacitySchedulerPolicy.Expanded the existing configuration scheduling strategy from
SchedulerPolicytoSchedulerConfigto enhance extensibility. The latter also introduces a chunk-based configuration calledContextChunkingPolicy.
[BREAKING CHANGE] The input prompt was removed from the generation output in the
generate()andgenerate_async()APIs. For example, when given a prompt asA B, the original generation result could be<s>A B C D Ewhere onlyC D Eis the actual output, and now the result isC D E.[BREAKING CHANGE] Switched default
add_special_tokenin the TensorRT-LLM backend toTrue.Deprecated
GptSessionandTrtGptModelV1.
Model Updates#
Support DBRX
Support Qwen2
Support CogVLM
Support ByT5
Support LLaMA 3
Support Arctic (w/ FP8)
Support Fuyu
Support Persimmon
Support Deplot
Support Phi-3-Mini with long Rope
Support Neva
Support Kosmos-2
Support RecurrentGemma
Fixed Issues#
Fixed some unexpected behaviors in beam search and early stopping, so that the outputs are more accurate.
Fixed segmentation fault with pipeline parallelism and
gather_all_token_logits. (#1284)Removed the unnecessary check in XQA to fix code Llama 70b Triton crashes. (#1256)
Fixed an unsupported ScalarType issue for BF16 LoRA. (https://github.com/triton-inference-server/tensorrtllm_backend/issues/403)
Eliminated the load and save of prompt table in multimodal. (https://github.com/NVIDIA/TensorRT-LLM/discussions/1436)
Fixed an error when converting the models weights of Qwen 72B INT4-GPTQ. (#1344)
Fixed early stopping and failures on in-flight batching cases of Medusa. (#1449)
Added support for more NVLink versions for auto parallelism. (#1467)
Fixed the assert failure caused by default values of sampling config. (#1447)
Fixed a requirement specification on Windows for nvidia-cudnn-cu12. (#1446)
Fixed MMHA relative position calculation error in
gpt_attention_pluginfor enc-dec models. (#1343)
Infrastructure changes#
Base Docker image for TensorRT-LLM is updated to
nvcr.io/nvidia/pytorch:24.03-py3.Base Docker image for TensorRT-LLM backend is updated to
nvcr.io/nvidia/tritonserver:24.03-py3.The dependent TensorRT version is updated to 10.0.1.
The dependent CUDA version is updated to 12.4.0.
The dependent PyTorch version is updated to 2.2.2.
TensorRT-LLM Release 0.9.0#
Announcements#
TensorRT-LLM requires TensorRT 9.3 and 24.02 containers.
Key Features and Enhancements#
[BREAKING CHANGES] TopP sampling optimization with deterministic AIR TopP algorithm is enabled by default
[BREAKING CHANGES] Added support for embedding sharing for Gemma
Added support for context chunking to work with KV cache reuse
Enabled different rewind tokens per sequence for Medusa
Added BART LoRA support (limited to the Python runtime)
Enabled multi-LoRA for BART LoRA
Added support for
early_stopping=Falsein beam search for C++ RuntimeAdded support for logits post processor to the batch manager
Added support for import and convert HuggingFace Gemma checkpoints
Added support for loading Gemma from HuggingFace
Added support for auto parallelism planner for high-level API and unified builder workflow
Added support for running
GptSessionwithout OpenMPIAdded support for Medusa IFB
[Experimental] Added support for FP8 FMHA, note that the performance is not optimal, and we will keep optimizing it
Added support for more head sizes for LLaMA-like models
NVIDIA Ampere (SM80, SM86), NVIDIA Ada Lovelace (SM89), NVIDIA Hopper (SM90) all support head sizes [32, 40, 64, 80, 96, 104, 128, 160, 256]
Added support for OOTB functionality
T5
Mixtral 8x7B
Benchmark features
Added emulated static batching in
gptManagerBenchmarkAdded support for arbitrary dataset from HuggingFace for C++ benchmarks
Added percentile latency report to
gptManagerBenchmark
Performance features
Optimized
gptDecoderBatchto support batched samplingEnabled FMHA for models in BART, Whisper, and NMT family
Removed router tensor parallelism to improve performance for MoE models
Improved custom all-reduce kernel
Infrastructure features
Base Docker image for TensorRT-LLM is updated to
nvcr.io/nvidia/pytorch:24.02-py3The dependent PyTorch version is updated to 2.2
Base Docker image for TensorRT-LLM backend is updated to
nvcr.io/nvidia/tritonserver:24.02-py3The dependent CUDA version is updated to 12.3.2 (12.3 Update 2)
API Changes#
Added C++
executorAPIAdded Python bindings
Added advanced and multi-GPU examples for Python binding of
executorC++ APIAdded documents for C++
executorAPIMigrated Mixtral to high-level API and unified builder workflow
[BREAKING CHANGES] Moved LLaMA convert checkpoint script from examples directory into the core library
Added support for
LLM()API to accept engines built bytrtllm-buildcommand[BREAKING CHANGES] Removed the
modelparameter fromgptManagerBenchmarkandgptSessionBenchmark[BREAKING CHANGES] Refactored GPT with unified building workflow
[BREAKING CHANGES] Refactored the Qwen model to the unified build workflow
[BREAKING CHANGES] Removed all the LoRA related flags from
convert_checkpoint.pyscript and the checkpoint content totrtllm-buildcommand to generalize the feature better to more models[BREAKING CHANGES] Removed the
use_prompt_tuningflag, options from theconvert_checkpoint.pyscript, and the checkpoint content to generalize the feature better to more models. Usetrtllm-build --max_prompt_embedding_table_sizeinstead.[BREAKING CHANGES] Changed the
trtllm-build --world_sizeflag to the--auto_parallelflag. The option is used for auto parallel planner only.[BREAKING CHANGES]
AsyncLLMEngineis removed. Thetensorrt_llm.GenerationExecutorclass is refactored to work with both explicitly launching withmpirunin the application level and accept an MPI communicator created bympi4py.[BREAKING CHANGES]
examples/serverare removed.[BREAKING CHANGES] Removed LoRA related parameters from the convert checkpoint scripts.
[BREAKING CHANGES] Simplified Qwen convert checkpoint script.
[BREAKING CHANGES] Reused the
QuantConfigused intrtllm-buildtool to support broader quantization features.Added support for TensorRT-LLM checkpoint as model input.
Refined
SamplingConfigused inLLM.generateorLLM.generate_asyncAPIs, with the support of beam search, a variety of penalties, and more features.Added support for the
StreamingLLMfeature. Enable it by settingLLM(streaming_llm=...).
Model Updates#
Added support for distil-whisper
Added support for HuggingFace StarCoder2
Added support for VILA
Added support for Smaug-72B-v0.1
Migrate BLIP-2 examples to
examples/multimodal
Limitations#
openai-tritonexamples are not supported on Windows.
Fixed Issues#
Fixed a weight-only quant bug for Whisper to make sure that the
encoder_input_len_rangeis not0. (#992)Fixed an issue that log probabilities in Python runtime are not returned. (#983)
Multi-GPU fixes for multimodal examples. (#1003)
Fixed a wrong
end_idissue for Qwen. (#987)Fixed a non-stopping generation issue. (#1118, #1123)
Fixed a wrong link in
examples/mixtral/README.md. (#1181)Fixed LLaMA2-7B bad results when INT8 kv cache and per-channel INT8 weight only are enabled. (#967)
Fixed a wrong
head_sizewhen importing a Gemma model from HuggingFace Hub. (#1148)Fixed ChatGLM2-6B building failure on INT8. (#1239)
Fixed a wrong relative path in Baichuan documentation. (#1242)
Fixed a wrong
SamplingConfigtensor inModelRunnerCpp. (#1183)Fixed an error when converting SmoothQuant LLaMA. (#1267)
Fixed an issue that
examples/run.pyonly load one line from--input_file.Fixed an issue that
ModelRunnerCppdoes not transferSamplingConfigtensor fields correctly. (#1183)
TensorRT-LLM Release 0.8.0#
Key Features and Enhancements#
Chunked context support (see docs/source/advanced/gpt-attention.md#chunked-context)
LoRA support for C++ runtime (see docs/source/lora.md)
Medusa decoding support (see examples/medusa/README.md)
The support is limited to Python runtime for Ampere or newer GPUs with fp16 and bf16 accuracy, and the
temperatureparameter of sampling configuration should be 0
StreamingLLM support for LLaMA (see docs/source/advanced/gpt-attention.md#streamingllm)
Support for batch manager to return logits from context and/or generation phases
Include support in the Triton backend
Support AWQ and GPTQ for QWEN
Support ReduceScatter plugin
Support for combining
repetition_penaltyandpresence_penalty#274Support for
frequency_penalty#275OOTB functionality support:
Baichuan
InternLM
Qwen
BART
LLaMA
Support enabling INT4-AWQ along with FP8 KV Cache
Support BF16 for weight-only plugin
Baichuan
P-tuning support
INT4-AWQ and INT4-GPTQ support
Decoder iteration-level profiling improvements
Add
masked_selectandcumsumfunction for modelingSmooth Quantization support for ChatGLM2-6B / ChatGLM3-6B / ChatGLM2-6B-32K
Add Weight-Only Support To Whisper #794, thanks to the contribution from @Eddie-Wang1120
Support FP16 fMHA on NVIDIA V100 GPU
Note
Some features are not enabled for all models listed in the examples folder.
Model Updates#
Phi-1.5/2.0
Mamba support (see examples/mamba/README.md)
The support is limited to beam width = 1 and single-node single-GPU
Nougat support (see examples/multimodal/README.md#nougat)
Qwen-VL support (see examples/qwenvl/README.md)
RoBERTa support, thanks to the contribution from @erenup
Skywork model support
Add example for multimodal models (BLIP with OPT or T5, LlaVA)
Refer to the Software section for a list of supported models.
API
Add a set of LLM APIs for end-to-end generation tasks (see examples/llm-api/README.md)
[BREAKING CHANGES] Migrate models to the new build workflow, including LLaMA, Mistral, Mixtral, InternLM, ChatGLM, Falcon, GPT-J, GPT-NeoX, Medusa, MPT, Baichuan and Phi (see docs/source/new_workflow.md)
[BREAKING CHANGES] Deprecate
LayerNormandRMSNormplugins and removed corresponding build parameters[BREAKING CHANGES] Remove optional parameter
maxNumSequencesfor GPT manager
Fixed Issues
Fix the first token being abnormal issue when
--gather_all_token_logitsis enabled #639Fix LLaMA with LoRA enabled build failure #673
Fix InternLM SmoothQuant build failure #705
Fix Bloom int8_kv_cache functionality #741
Fix crash in
gptManagerBenchmark#649Fix Blip2 build error #695
Add pickle support for
InferenceRequest#701Fix Mixtral-8x7b build failure with custom_all_reduce #825
Fix INT8 GEMM shape #935
Minor bug fixes
Performance
[BREAKING CHANGES] Increase default
freeGpuMemoryFractionparameter from 0.85 to 0.9 for higher throughput[BREAKING CHANGES] Disable
enable_trt_overlapargument for GPT manager by defaultPerformance optimization of beam search kernel
Add bfloat16 and paged kv cache support for optimized generation MQA/GQA kernels
Custom AllReduce plugins performance optimization
Top-P sampling performance optimization
LoRA performance optimization
Custom allreduce performance optimization by introducing a ping-pong buffer to avoid an extra synchronization cost
Integrate XQA kernels for GPT-J (beamWidth=4)
Documentation
Batch manager arguments documentation updates
Add documentation for best practices for tuning the performance of TensorRT-LLM (See docs/source/perf_best_practices.md)
Add documentation for Falcon AWQ support (See examples/falcon/README.md)
Update to the
docs/source/new_workflow.mddocumentationUpdate AWQ INT4 weight only quantization documentation for GPT-J
Add blog: Speed up inference with SOTA quantization techniques in TRT-LLM
Refine TensorRT-LLM backend README structure #133
Typo fix #739
TensorRT-LLM Release 0.7.1#
Key Features and Enhancements#
Speculative decoding (preview)
Added a Python binding for
GptManagerAdded a Python class
ModelRunnerCppthat wraps C++gptSessionSystem prompt caching
Enabled split-k for weight-only cutlass kernels
FP8 KV cache support for XQA kernel
New Python builder API and
trtllm-buildcommand (already applied to blip2 and OPT)Support
StoppingCriteriaandLogitsProcessorin Python generate APIFHMA support for chunked attention and paged KV cache
Performance enhancements include:
MMHA optimization for MQA and GQA
LoRA optimization: cutlass grouped GEMM
Optimize Hopper warp specialized kernels
Optimize
AllReducefor parallel attention on Falcon and GPT-JEnable split-k for weight-only cutlass kernel when SM>=75
Added Workflow documentation
Model Updates#
BART and mBART support in encoder-decoder models
FairSeq Neural Machine Translation (NMT) family
Mixtral-8x7B model
Support weight loading for HuggingFace Mixtral model
OpenAI Whisper
Mixture of Experts support
MPT - Int4 AWQ / SmoothQuant support
Baichuan FP8 quantization support
Fixed Issues#
Fixed tokenizer usage in
quantize.py#288Fixed LLaMa with LoRA error
Fixed LLaMA GPTQ failure
Fixed Python binding for InferenceRequest issue
Fixed CodeLlama SQ accuracy issue
Known Issues#
The hang reported in issue #149 has not been reproduced by the TensorRT-LLM team. If it is caused by a bug in TensorRT-LLM, that bug may be present in that release.