modelopt.torch.quantization.backends.fp8_per_tensor_gemm

fp8_per_tensor_gemm(quant_module, input, bias=None)

GEMM function for fp8 per tensor quantization.