Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
use bfloat16 in grad reduction (= reduce-scatter / all-reduce)
ref: https://github.com/pytorch/pytorch/blob/762b1b4/torch/distributed/_composable/fsdp/_fsdp_api.py#L9 Signed-off-by: Masaki Kozuki <mkozuki@nvidia.com>
- Loading branch information