Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Allow self-tuning submissions to specify dropout_rate and aux_dropout_rate #851

Open
wants to merge 9 commits into
base: dev
Choose a base branch
from

Conversation

Niccolo-Ajroldi
Copy link
Contributor

Original issue

Self-tuning submissions use default values for dropout_rate and aux_dropout_rate (see #753).
We want to allow them to specify custom values for these hyperparameters.

Solution

Allow self-tuning submission to provide a JSON file specifying values for these hyperparameters.
The JSON file should look like this:

{
    "dropout_rate": 0.1,
    "aux_dropout_rate": 0.2,
}

This is passed to submission_runner.py through --tuning_search_space.

We read from such JSON file, and check that:

  1. it specifies a value only for the allowed hyperparameters, and that
  2. it specifies one single value for each of them.

We then proceed to call train_once with the correspondent fixed hyperparameters.

Discussion

I think this solution is pretty clean, I am just dubious about passing the JSON file path trough tuning_search_space., as this might create some confusion. Perhaps we shall change the name of this flag, or we can add a different flag?

Fix #753

@Niccolo-Ajroldi Niccolo-Ajroldi requested a review from a team as a code owner March 12, 2025 17:17
Copy link

github-actions bot commented Mar 12, 2025

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

@priyakasimbeg
Copy link
Contributor

I agree with you, I think we should add an additional flag that is optional for self-tuning and disallowed for external-tuning ruleset (so raise an error if the value of the flag is not none for external-tuning).
Perhaps we can call this flag self_tuning_init_hyperparemeters_config? Maybe that is a bit long, do you have any suggestions for names?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants