Skip to content

Training large models (>=7B) on multi-gpu distributed setups using technologies like FSDP, DeepSpeed, HF Accelerate

Notifications You must be signed in to change notification settings

aashrith-madasu/Train-Large-Models-Distributed

About

Training large models (>=7B) on multi-gpu distributed setups using technologies like FSDP, DeepSpeed, HF Accelerate

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published