-
Notifications
You must be signed in to change notification settings - Fork 86
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
When comparing Thunder Torch Executor to Torch Eager, the ResNet18 gradients are not close for FP32. #655
Comments
run this script
has GradcheckError:
with float64 it can pass |
https://pytorch.org/docs/stable/generated/torch.autograd.gradcheck.gradcheck.html says
however, the values above seem very far off, so I'm wondering whether the operators we call have some bug / input assumptions not satisfied etc. |
Comparison of fp64 and fp32 results: Torch eager fp64 vs thunder torchex fp32: Torch eager fp64 vs torch eager fp32: Torch eager fp32 vs thunder torchex fp32: script
|
triage review:
|
Note: If you have a model or program that is not supported yet but should be, please use the program coverage template.
🐛 Bug
To Reproduce
Steps to reproduce the behavior:
lightning-thunder/thunder/tests/test_inplace_functionalization.py
Line 184 in 6320b2f
to
if train and executor == TorchExecutor:
pytest thunder/tests/test_inplace_functionalization.py -k test_parse_resnet18_torch_cuda_float32[True]
see error:
The text was updated successfully, but these errors were encountered: