Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Proper hyperparameter handling #53

Merged
merged 2 commits into from
Jan 30, 2025
Merged

Proper hyperparameter handling #53

merged 2 commits into from
Jan 30, 2025

Conversation

Flova
Copy link
Member

@Flova Flova commented Jan 28, 2025

Proposed changes

Until now, all parameters related to model training and inference were hardcoded in the respective scripts. Now we have yaml config files for the training. The parameters are subsequently stored in the model files itself and are automatically loaded later during inference to configure the model and dataset.

Additionally we also store the optimizer and lr_sheduler state which enables effortless restarting.

Old models were converted using a small script.

Related issues

Close #21

Checklist

  • Write documentation
  • Create issues for future work
  • This PR is on our DDLitLab project board

@Flova Flova requested review from texhnolyze and jaagut January 28, 2025 18:42
@Flova Flova merged commit d4fe368 into main Jan 30, 2025
5 checks passed
@Flova Flova deleted the feature/proper_param_handling branch January 30, 2025 10:51
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: Done
Development

Successfully merging this pull request may close these issues.

Add hyperparameter configuration
3 participants