NVIDIA Generative AI Examples

Generative AI enables users to quickly generate new content based on a variety of inputs and is a powerful tool for streamlining the workflow of creatives, engineers, researchers, scientists, and more. The use cases and possibilities span all industries and individuals. Generative AI models can produce novel content like stories, emails, music, images, and videos.

Generative AI starts with foundational models trained on vast quantities of unlabeled data. Large language models (LLMs) are trained on an extensive range of textual data online. These LLMs can understand prompts and generate novel, human-like responses. Businesses can build applications to leverage this capability of LLMs. Some uses are creative writing assistants for marketing, document summarization for legal teams, and code writing for software development.

The NVIDIA Generative AI Examples use Docker Compose run Retrieval Augmented Generation (RAG) Large Language Model (LLM) pipelines.

All the example pipelines deploy a sample chat bot application for question and answering that is enhanced with RAG. The chat bot also supports uploading documents to create a knowledge base.

Developer RAG Examples

Model
Embedding
Framework
Description
Model
Location
NIM
for
LLMs
Vector
Database

ai-llama3-70b

snowflake-arctic-embed-l

LangChain

Using the NVIDIA API Catalog

API Catalog

No

Milvus or pgvector

ai-llama3-70b

snowflake-arctic-embed-l

LangChain

Query Decomposition

API Catalog

No

Milvus or pgvector

meta/llama3-70b-instruct for response generation

meta/llama3-70b-instruct for PandasAI

Not Applicable

PandasAI

Structured Data

API Catalog

No

Not Applicable

ai-llama3-8b for response generation

ai-google-Deplot for graph to text conversion

ai-Neva-22B for image to text conversion

snowflake-arctic-embed-l

Custom Python

Multimodal Data

API Catalog

No

Milvus or pgvector

ai-llama3-8b

snowflake-arctic-embed-l

LangChain

Multi-Turn Conversational Chat Bot

API Catalog

No

Milvus or pgvector

meta-llama3-8b-instruct

nv-embed-qa:4

LangChain

Using NVIDIA NIM for LLMs

Local LLM

Yes

Milvus or pgvector

Open Source Connectors