Skip to content

Commit 7efa6e0

Browse files
authored
Update README.md (#322)
Updated to move to ROCm 6.3 and post the issue with saving Tunable Ops due to PyTorch bug.
1 parent a1aaa74 commit 7efa6e0

File tree

1 file changed

+5
-6
lines changed

1 file changed

+5
-6
lines changed

docs/dev-docker/README.md

Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -10,11 +10,11 @@ This documentation shows some reference performance numbers and the steps to rep
1010

1111
It includes:
1212

13-
- ROCm™ 6.2.2
13+
- ROCm™ 6.3
1414

1515
- vLLM 0.6.3
1616

17-
- PyTorch 2.5dev (nightly)
17+
- PyTorch 2.6dev (nightly)
1818

1919
## System configuration
2020

@@ -23,7 +23,7 @@ The performance data below was measured on a server with MI300X accelerators wit
2323
| System | MI300X with 8 GPUs |
2424
|---|---|
2525
| BKC | 24.13 |
26-
| ROCm | version ROCm 6.2.2 |
26+
| ROCm | version ROCm 6.3 |
2727
| amdgpu | build 2009461 |
2828
| OS | Ubuntu 22.04 |
2929
| Linux Kernel | 5.15.0-117-generic |
@@ -45,9 +45,8 @@ You can pull the image with `docker pull rocm/vllm-dev:main`
4545

4646
### What is New
4747

48-
- MoE optimizations for Mixtral 8x22B, FP16
49-
- Llama 3.2 stability improvements
50-
- Llama 3.3 support
48+
- ROCm 6.3 support
49+
- Potential bug with Tunable Ops not saving due to a PyTorch issue
5150

5251

5352
Gemms are tuned using PyTorch's Tunable Ops feature (https://github.com/pytorch/pytorch/blob/main/aten/src/ATen/cuda/tunable/README.md)

0 commit comments

Comments
 (0)