-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Nf test tune #166
Nf test tune #166
Conversation
…e, weigths are now set in a reproducible and deterministic manner
…he correct tune_run path
…de of the process dir
elif user_tune_config["tune"]["scheduler"]["name"] == "FIFOScheduler": | ||
user_tune_config["tune"]["run_params"]["stop"]["training_iteration"] = 1 | ||
|
||
# TODO future schedulers specific info will go here as well. maybe find a cleaner way. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Most likely this needs to become a stimulus class in the same way experiment is
results.save_best_model(output) | ||
results.save_best_config(best_config_path) | ||
results.save_best_metrics_dataframe(best_metrics_path) | ||
results.save_best_optimizer(best_optimizer_path) | ||
|
||
# debug section. predict the validation data using the best model. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why does the behavior of debug load the best model and tests on the validation data, shouldn't this be reserved to the analysis module? How is this helping us with debug tuning?
@@ -179,12 +210,34 @@ def setup(self, config: dict, training: object, validation: object) -> None: | |||
self.training = DataLoader(training, batch_size=self.batch_size, shuffle=True) # TODO need to check the reproducibility of this shuffling | |||
self.validation = DataLoader(validation, batch_size=self.batch_size, shuffle=True) | |||
|
|||
# debug section, first create a dedicated directory for each worker inside Ray_results/<tune_model_run_specific_dir> location |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
In the future, I believe this should be done regardless of debug or not (saving seed and initial model), this would be a "robustness mode" toggled "on" by default!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Or a "reproducibility" mode, as I believe the formula "model + initial state + seed + training code + training data" is our "deep learning" container
Changes aported:
TODO for the future is decide wich flavor of the nf-test tune should be put as guthub action to test reproducibility.