-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Access to current_training_job_name before .train() #5047
Comments
cc @benieric |
Hi @discort, I wonder if better solution would be to have the By default, ModelTrainer could set the s3_output_path to follow same contract as the rest of artifacts like:
So user could provide a
This way user will be able to call |
Hi @benieric ,
I think it could resolve the case I described for |
Yeah, so if we went with a solution like this one, would make sense to also have the s3 path be optional for both OutputDataConfig and CheckpointConfig and have the ModelTrainer resolve the path to be under same rules |
That makes total sense to me. |
Describe the feature you'd like
I want to keep training artifacts and tensorboard logs for a training job in the same s3 folder.
How would this feature be used? Please describe.
This feature allows to keep my artifacts and tensorboard logs organized. For instance, I can easily find my logs by a job name.
results on s3://:
Describe alternatives you've considered
The only alternative that's coming to my mind is using timestamp in
base_job_name
. However, the drawback of this approach results in getting unpleasant training job name likebase_job_name-<my-timestamp>-<generated-timestamp>
Additional context
The text was updated successfully, but these errors were encountered: