Replies: 1 comment 1 reply
-
We usually do Jit of individual training steps to keep code clean and easy to understand. Overall, if you Jit multiple training steps then you probably will get some performance boost. I think people don't do this because for any complex model and non-trivial dataset this performance boost would be relatively small and at the same time code will become more complex. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Would it be better to Jit a lax.scan of the entire optimization loop versus jitting the loss and using a regular python loop?
I see that usually the latter is done in most examples
Beta Was this translation helpful? Give feedback.
All reactions