Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Curious about the torch.compile baseline in benchmark_litgpt #123

Open
mitkotak opened this issue Apr 3, 2024 · 3 comments
Open

Curious about the torch.compile baseline in benchmark_litgpt #123

mitkotak opened this issue Apr 3, 2024 · 3 comments
Labels
question Further information is requested

Comments

@mitkotak
Copy link

mitkotak commented Apr 3, 2024

Was curious to know why the reduce-overhead option was not used as a baseline and if there's a comparison somewhere using reduce-overhead ?

Thanks !

cc @apaz-cli

@mitkotak mitkotak added enhancement New feature or request help wanted Extra attention is needed labels Apr 3, 2024
@carmocca
Copy link
Contributor

carmocca commented Apr 3, 2024

It should be because we are also not enabling CUDAGraphs with Thunder too ("reduce-overhead" simply enables CUDAGraphs). So this way is more of an apples-to-apples comparison. Maybe @parthmannan can correct me here

Either way, from my personal experience in running this, cudagraphs won't help during training (especially if your model is large enough) because you are not overhead bound at all

@carmocca carmocca added question Further information is requested and removed enhancement New feature or request help wanted Extra attention is needed labels Apr 3, 2024
@mitkotak
Copy link
Author

mitkotak commented Apr 3, 2024

Thanks for the reply ! How about max-autotune-no-cudagraphs ?

@carmocca
Copy link
Contributor

carmocca commented Apr 3, 2024

Sure, it could be used. Anecdotically, I found it to provide tiny speedups (if any) for a whole lot of compilation time, but it's a valid suggestion.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants