The current MLXOptimizers implementation does not support saving and loading the optimizer's internal state (e.g., m and v for Adam), as the stateStorage is internal and inaccessible.
This prevents resuming training from checkpoints with preserved optimizer momentum, requiring restarts from zero state.
Will a feature be added to manipulate and serialize the optimizer state, such as a public state property or dedicated loadState/saveState methods?