-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hyperparam tuning #54
base: main
Are you sure you want to change the base?
Conversation
…for learning rate hyperparameter tuning
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you paste the error with tensorboard and also which line you comment to solve the issue? Have you checked that this PR does not brake our existing build? Maybe add a parameter --ray_tune so that if we don't specify it, the behavior is the same as currently (I'm a bit skeptical regarding the new parameters).
Also could you give us some instructions on how to install raytune (It seems like you struggled, so I would love to get your opinion on how to do this :) )
To address your points:
was found to be non-serializable. There may be multiple other undetected variables that were non-serializable. Once I commented out all references to writer (eg. writer.add_scalar) in train_multitask and model_eval_multitask (which is called by train_multitask), the code runs fine.
Raytune was actually simple enough to install on AWS with pip. But here's the command: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is it possible to refactor the tuning code in another file than the multitask_classifier in the main ?
Hi @marie-huynh raytune serves as a wrapper function and requires the training function (train_multitask) to be modified to accept a config dictionary passed by ray and use the arguments within it. So I don't think I can run raytune without directly editing multitask_classifier. However, I have not made functional edits to train_multitask (only editing the input argument to be a config dictionary passed by ray instead of args). Internally, I've also edited train_multitask to use arguments from this config dictionary. Finally, I've added two new command line flags:
|
I've updated my script to avoid out of memory errors by saving directly to the disk rather than RAM. As long as you update your disk size, it should run ok. |
Hi guys, I've made some edits to hyperparameter tuning with raytune. Would love to get your feedback.
Some notes: