Skip to content

Commit 23704d2

Browse files
committed
feat: add new cuda profile
1 parent 1e82ac9 commit 23704d2

File tree

2 files changed

+35
-0
lines changed

2 files changed

+35
-0
lines changed

docker-compose.yaml

Lines changed: 20 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -48,6 +48,26 @@ services:
4848
profiles:
4949
- llamacpp-cpu
5050

51+
# Private-GPT service for the local mode (with CUDA support)
52+
# This service builds from a local Dockerfile and runs the application in local mode.
53+
private-gpt-llamacpp-cuda:
54+
image: ${PGPT_IMAGE:-zylonai/private-gpt}${PGPT_TAG:-0.6.1}-llamacpp-cuda
55+
build:
56+
context: .
57+
dockerfile: Dockerfile.llamacpp-cuda
58+
volumes:
59+
- ./local_data/:/home/worker/app/local_data
60+
- ./models/:/home/worker/app/models
61+
entrypoint: sh -c ".venv/bin/python scripts/setup && .venv/bin/python -m private_gpt"
62+
ports:
63+
- "8001:8001"
64+
environment:
65+
PORT: 8001
66+
PGPT_PROFILES: local
67+
HF_TOKEN: ${HF_TOKEN}
68+
profiles:
69+
- llamacpp-cuda
70+
5171
#-----------------------------------
5272
#---- Ollama services --------------
5373
#-----------------------------------

fern/docs/pages/quickstart/quickstart.mdx

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -82,6 +82,21 @@ HF_TOKEN=<your_hf_token> docker-compose --profile llamacpp-cpu up
8282
```
8383
Replace `<your_hf_token>` with your actual Hugging Face token.
8484

85+
#### 2. LlamaCPP CUDA
86+
87+
**Description:**
88+
This profile runs the Private-GPT services locally using `llama-cpp` and Hugging Face models.
89+
90+
**Requirements:**
91+
A **Hugging Face Token (HF_TOKEN)** is required for accessing Hugging Face models. Obtain your token following [this guide](/installation/getting-started/troubleshooting#downloading-gated-and-private-models).
92+
93+
**Run:**
94+
Start the services with your Hugging Face token using pre-built images:
95+
```sh
96+
HF_TOKEN=<your_hf_token> docker-compose --profile llamacpp-cuda up
97+
```
98+
Replace `<your_hf_token>` with your actual Hugging Face token.
99+
85100
## Building Locally
86101

87102
If you prefer to build Docker images locally, which is useful when making changes to the codebase or the Dockerfiles, follow these steps:

0 commit comments

Comments
 (0)