======================================================= LLM Examples Introduction ======================================================= Here is a simple example to show how to use the LLM with TinyLlama. .. literalinclude:: ../../../examples/llm-api/quickstart_example.py :language: python :linenos: The LLM API can be used for both offline or online usage. See more examples of the LLM API here: .. toctree:: :maxdepth: 1 :caption: LLM API Examples llm_inference_async llm_inference_async_streaming llm_inference_distributed llm_logits_processor llm_multilora llm_lookahead_decoding llm_guided_decoding llm_inference llm_inference_customize llm_medusa_decoding llm_quantization llm_auto_parallel For more details on how to fully utilize this API, check out: * `Common customizations `_ * `LLM API Reference <../llm-api/index.html>`_