Skip to content

Commit

Permalink
docs: LLAMA_CUBLAS -> LLAMA_CUDA
Browse files Browse the repository at this point in the history
  • Loading branch information
abetlen committed Apr 4, 2024
1 parent 612e78d commit c50309e
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -102,10 +102,10 @@ CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install llama-cpp-
<details>
<summary>cuBLAS (CUDA)</summary>

To install with cuBLAS, set the `LLAMA_CUBLAS=on` environment variable before installing:
To install with cuBLAS, set the `LLAMA_CUDA=on` environment variable before installing:

```bash
CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python
CMAKE_ARGS="-DLLAMA_CUDA=on" pip install llama-cpp-python
```

</details>
Expand Down Expand Up @@ -569,7 +569,7 @@ python3 -m llama_cpp.server --model models/7B/llama-model.gguf
Similar to Hardware Acceleration section above, you can also install with GPU (cuBLAS) support like this:

```bash
CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install 'llama-cpp-python[server]'
CMAKE_ARGS="-DLLAMA_CUDA=on" FORCE_CMAKE=1 pip install 'llama-cpp-python[server]'
python3 -m llama_cpp.server --model models/7B/llama-model.gguf --n_gpu_layers 35
```

Expand Down

0 comments on commit c50309e

Please sign in to comment.