Hyperparameter Tuning with Weights & Biases

Neu.ro allows you to run model training in parallel with different hyperparameter combinations via integration with Weights & Biases. W&B is an experiment tracking tool for deep learning. The ML engineer only needs to initiate the process: prepare the code for the model training, set up the hyperparameters space, and start the search with just one command. Neu.ro is in charge of the rest.

To see this guide in action, check out our recipe, where we apply hyperparemeter tuning with W&B to image classification task.

Creating a Neu.ro project

The Neu.ro project template contains an integration with Weights and Biases. To create a new project from a template, you need to follow a couple of steps.

First, Sign up and install CLI client

Then, create a project:

neuro project init

This command will then ask you several simple questions:

project_name \[Neuro Project]: Hyperparameter tuning test
project_slug \[hyperparameter-tuning-test]:
code_directory \[modules]:

Press Enter if you do not want to change a suggested value.

Then, change the working directory:

cd hyperparameter-tuning-test

Connecting Weights & Biases

Now, connect your project with Weights & Biases:

  1. Find your API key (it is also called a token) on W&B’s settings page(section “API keys”). It should be a sequence like cf23df2207d99a74fbe169e3eba035e633b65d94.

  2. Save your API key (token) to a file in your local home directory ~ and protect it by setting appropriate permissions to make W&B available on Neu.ro platform:

export WANDB_SECRET_FILE=wandb-token.txt
echo "cf23df2207d99a74fbe169e3eba035e633b65d94" > ~/$WANDB_SECRET_FILE
chmod 600 ~/$WANDB_SECRET_FILE

After that, create a Neu.ro secret:

neuro secrets add wandb-token @~/$WANDB_SECRET_FILE

Next, add several changes to Makefile in order to enable experiment tracking:

# 1. Replace this line:
# With the following:
SECRETS?="-e WANDB_API_KEY=secrets:wandb-token"
# 2. Add this line after "N_JOBS?=3"
# 3. Add this line after ##### CONSTANTS #####

Also, add hypertrain , kill-hypertrain and kill-hypertrain-all Make targets from this file.

Please find instructions on using Weights & Biases in your code in W&B documentation and W&B example projects.

Using Weights & Biases for hyperparameter tuning

If you have completed the previous parts of the instruction, W&B is ready to use. To run hyperparameter tuning for the model, you need to:

  • define the list of hyperparameters (in a wandb-sweep.yaml file), and

  • send the metrics to W&B after each run (by using Makefile and make hypertrain command).

Makefile and wandb-sweep.yaml have links to train.py (you can look at an example here) If you want to run hypertrain for another script, you can change the program property in wandb-sweep.yaml (see below). The script must contain the description of the model and the training loop.

The Python script must also receive parameters with the same names, as specified in wandb-sweep.yaml as arguments of the command line and use them for model training/evaluation. For example, you can use command line parameters such as the argparse Python module.

There are more details below:

  1. train.py is a file that contains the model training code. It should log the metrics with W&B, for example, in our case:

wandb.log({'accuracy': 0.9})
  1. wandb-sweep.yaml has the following structure:

wandb-sweep.yaml example
  • Line 1: /../train.py is a default path to a file with model training code.

  • Line 2: a method that is used for the hyperparameters tuning. For more information, see W&B docs.

  • Lines 4, 5: the name of the metric that is supposed to be optimized. The ML engineer can change this metric according to the task.

  • Lines 6 -12: hyperparameter settings. The ML engineer should use them in the train.py file. Names, values, and ranges are changeable as well.

The name of the file wandb-sweep.yaml and the path to it may also be modified in Makefile in the root directory of your project.

Hyperparameter tuning

Now that you have set up both Neu.ro and W&B and prepared your training script, it’s time to try hyperparameter tuning. For that, you should run the following command:

make hypertrain

This starts jobs that run train.py script (or whatever name you have chosen) on the platform with different sets of hyperparameters in parallel. By default, just 3 jobs run at the same time. You can change this number by modifying N_JOBS variable in Makefile or adding it at the end of make command:

make hypertrain N_JOBS=10

To monitor the hyperparameter tuning process, follow the link which W&B provides at the beginning of the process.

If you want to stop the hyperparameter tuning, terminate all related jobs:

make kill-hypertrain-all

After that, verify that the jobs stopped (make ps), and then delete unused sweeps from the local file .wandb_sweeps.