The notebooks show how to use the langchain-nvidia-ai-endpoints and llama-index-embeddings-nvidia Python packages.
These packages provide the basics for developing a RAG application and performing inference either from NVIDIA API Catalog endpoints or a local deployment of NVIDIA microservices.
- You have Python 3 installed.
- Complete the common prerequisites.
-
Export your NVIDIA API key as an environment variable:
export NVIDIA_API_KEY="nvapi-<...>" -
Create a virtual environment:
python3 -m venv .venv source .venv/bin/activate
-
Install JupyterLab in the virtual environment:
pip3 install jupyterlab -
Start the JupyterLab server:
jupyter lab --allow-root --ip=0.0.0.0 --NotebookApp.token='' --port=8889 -
Open a web browser and access http://localhost:8889/lab.
Browse to the
RAG/notebooksdirectory to open an execute the cells of the notebooks.