Replies: 15 comments 13 replies
-
#625 multimodal models (e.g. Llama 3.2) (work in progress) |
Beta Was this translation helpful? Give feedback.
-
#184 MoE and Expert Parallel (work in progress) |
Beta Was this translation helpful? Give feedback.
-
state space models (e.g. Mamba) |
Beta Was this translation helpful? Give feedback.
-
diffusion models (e.g. DiT) |
Beta Was this translation helpful? Give feedback.
-
What about varlen ring attention? There are a few variants which are not stable enough for production use; its engineering/testing obstacles make it suitable for the strengths of the PyTorch folks. |
Beta Was this translation helpful? Give feedback.
-
Evaluations implementation with support for 4D parallelism |
Beta Was this translation helpful? Give feedback.
-
I'd like to see a Deepspeed-Ulysses-Style sequence parallel used with DTensor. |
Beta Was this translation helpful? Give feedback.
-
Qwen 2.5 support. Features needed on top of
|
Beta Was this translation helpful? Give feedback.
-
Parallel training is non-intrusive to model implementation. |
Beta Was this translation helpful? Give feedback.
-
Would be nice to have a native (and tested) transformer export script as well. There is some trick to do with the complex number rope implementation in the conversion. We have a script to do this that should be compatible with the torchtitan implementation as it was originally copied from it : https://github.com/PrimeIntellect-ai/prime/blob/main/scripts/export_dcp.py |
Beta Was this translation helpful? Give feedback.
-
Would be nice as well to have an implementation of sequence packing / document masking using flex attention like in torchtune |
Beta Was this translation helpful? Give feedback.
-
It seems like the biggest thing missing from torchtitan that most production gaint LLM training runs have is fault tolerance training I.e. can quickly recover from a failure by removing just 1 data replicas instead of stopping the whole workload and restarting from ckpt |
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
Hi torchtitanists,
Thank you for your interests in torchtitan!
Please upvote on what features you would like to see next, and add one if it's not already there. We'll try to prioritize on the most requested features.
Beta Was this translation helpful? Give feedback.
All reactions