This repository was archived by the owner on Mar 31, 2025. It is now read-only.
Replies: 1 comment 1 reply
-
|
We usually do Jit of individual training steps to keep code clean and easy to understand. Overall, if you Jit multiple training steps then you probably will get some performance boost. I think people don't do this because for any complex model and non-trivial dataset this performance boost would be relatively small and at the same time code will become more complex. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Would it be better to Jit a lax.scan of the entire optimization loop versus jitting the loss and using a regular python loop?
I see that usually the latter is done in most examples
Beta Was this translation helpful? Give feedback.
All reactions