LLM Examples# Basics# Generate text Generate text asynchronously Generate text in streaming Distributed LLM Generation Customization# Generate text with guided decoding Control generated text using logits processor Generate text with multiple LoRA adapters Speculative Decoding KV Cache Connector KV Cache Offloading Runtime Configuration Examples Sampling Techniques Showcase Slurm# Run LLM-API with pytorch backend on Slurm Run trtllm-bench with pytorch backend on Slurm Run trtllm-serve with pytorch backend on Slurm