File size: 8,283 Bytes
fc0f7bd | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 | ---
id: nevergrad_sweeper
title: Nevergrad Sweeper plugin
sidebar_label: Nevergrad Sweeper plugin
---
[](https://pypi.org/project/hydra-nevergrad-sweeper/)


[](https://pypistats.org/packages/hydra-nevergrad-sweeper)
[](https://github.com/facebookresearch/hydra/tree/master/plugins/hydra_nevergrad_sweeper/example)
[](https://github.com/facebookresearch/hydra/tree/master/plugins/hydra_nevergrad_sweeper)
[Nevergrad](https://facebookresearch.github.io/nevergrad/) is a derivative-free optimization platform proposing a library of state-of-the art algorithms for hyperparameter search. This plugin provides a mechanism for Hydra applications to use [Nevergrad](https://facebookresearch.github.io/nevergrad/) algorithms for the optimization of experiments/applications parameters.
### Installation
This plugin requires Hydra 1.0 (Release candidate)
```commandline
$ pip install hydra-nevergrad-sweeper --pre
```
### Usage
Once installed, add `hydra/sweeper=nevergrad` to your command command. Alternatively, override `hydra/sweeper` in your config:
```yaml
defaults:
- hydra/sweeper: nevergrad
```
The default configuration is [here](https://github.com/facebookresearch/hydra/blob/master/plugins/hydra_nevergrad_sweeper/hydra_plugins/hydra_nevergrad_sweeper/config.py).
## Example of training using Nevergrad hyperparameter search
We include an example of how to use this plugin. The file [`example/dummy_training.py`](plugins/hydra_nevergrad_sweeper/example/dummy_training.py) implements an example of how to perform minimization of a (dummy) function including a mixture of continuous and discrete parameters.
This application has the following configuration:
```yaml
defaults:
- hydra/sweeper: nevergrad
hydra:
sweeper:
params:
# configuration of the optimizer
optim:
# name of the Nevergrad optimizer to use. Here is a sample:
# - "OnePlusOne" extremely simple and robust, especially at low budget, but
# tends to converge early.
# - "CMA" very good algorithm, but may require a significant budget (> 120)
# - "TwoPointsDE": an algorithm good in a wide range of settings, for significant
# budgets (> 120).
# - "Shiwa" an algorithm aiming at identifying the best optimizer given your input
# definition (work in progress, it may still be ill-suited for low budget)
# find out more within nevergrad's documentation:
# https://github.com/facebookresearch/nevergrad/
optimizer: OnePlusOne
# total number of function evaluations to perform
budget: 100
# number of parallel workers for performing function evaluations
num_workers: 10
# maximize: true # comment out for maximization
# default parametrization of the search space
parametrization:
# either one or the other
db:
- mnist
- cifar
# a log-distributed positive scalar, evolving by factors of 2 on average
lr:
init: 0.02
step: 2.0
log: true
# a linearly-distributed scalar between 0 and 1
dropout:
lower: 0.0
upper: 1.0
# an integer scalar going from 4 to 16
# init and step parameters could also be provided,
# by default init is set to the middle of the range
# and step is set to a sixth of the range
batch_size:
lower: 4
upper: 16
integer: true
db: cifar
lr: 0.01
batch_size: 8
dropout: 0.6
```
The function decorated with `@hydra.main()` returns a float which we want to minimize, the minimum is 0 and reached for:
```yaml
db: mnist
lr: 0.12
dropout: 0.33
batch_size=4
```
To run hyperparameter search and look for the best parameters for this function, clone the code and run the following command in the `plugins/hydra_nevergrad_sweeper` directory:
```bash
python example/dummy_training.py -m
```
You can also override the search space parametrization:
```bash
python example/dummy_training.py -m db=mnist,cifar batch_size=4,8,16 lr=log:0.001:1 dropout=0:1
```
The initialization of the sweep and the first 5 evaluations (out of 100) look like this:
```text
[HYDRA] NevergradSweeper(optimizer=OnePlusOne, budget=100, num_workers=10) minimization
[HYDRA] with parametrization Dict(batch_size=TransitionChoice(choices=Tuple(4,8,16),position=Scalar[sigma=Log{exp=1.2}],transitions=[1. 1.]),db=Choice(choices=Tuple(mnist,cifar),weights=Array{(2,)}),dropout=Scalar{Cl(0,1)}[sigma=Log{exp=1.2}],lr=Log{exp=3.162277660168379,Cl(0.001,1)}):{'db': 'cifar', 'batch_size': 8, 'lr': 0.03162277660168379, 'dropout': 0.5}
[HYDRA] Sweep output dir: multirun/2020-03-04/17-53-29
[HYDRA] Launching 10 jobs locally
[HYDRA] #0 : db=mnist batch_size=8 lr=0.032 dropout=0.5
[__main__][INFO] - dummy_training(dropout=0.500, lr=0.032, db=mnist, batch_size=8) = 5.258
[HYDRA] #1 : db=mnist batch_size=16 lr=0.035 dropout=0.714
[__main__][INFO] - dummy_training(dropout=0.714, lr=0.035, db=mnist, batch_size=16) = 13.469
[HYDRA] #2 : db=cifar batch_size=8 lr=0.053 dropout=0.408
[__main__][INFO] - dummy_training(dropout=0.408, lr=0.053, db=cifar, batch_size=8) = 4.145
[HYDRA] #3 : db=cifar batch_size=8 lr=0.012 dropout=0.305
[__main__][INFO] - dummy_training(dropout=0.305, lr=0.012, db=cifar, batch_size=8) = 4.133
[HYDRA] #4 : db=mnist batch_size=4 lr=0.030 dropout=0.204
[__main__][INFO] - dummy_training(dropout=0.204, lr=0.030, db=mnist, batch_size=4) = 1.216
```
and the final 2 evaluations look like this:
```text
[HYDRA] #8 : db=mnist batch_size=4 lr=0.094 dropout=0.381
[__main__][INFO] - dummy_training(dropout=0.381, lr=0.094, db=mnist, batch_size=4) = 1.077
[HYDRA] #9 : db=mnist batch_size=4 lr=0.094 dropout=0.381
[__main__][INFO] - dummy_training(dropout=0.381, lr=0.094, db=mnist, batch_size=4) = 1.077
[HYDRA] Best parameters: db=mnist batch_size=4 lr=0.094 dropout=0.381
```
The run also creates an `optimization_results.yaml` file in your sweep folder with the parameters recommended by the optimizer:
```yaml
best_evaluated_result: 0.381
best_evaluated_params:
batch_size: 4
db: mnist
dropout: 0.381
lr: 0.094
name: nevergrad:
```
## Defining the parameters
The plugin can use 2 types of parameters:
### Choices
Choices are defined with **comma-separated values** in the command-line (`db=mnist,cifar` or `batch_size=4,8,12,16`) or with a list in a config file.
By default, values are processed as floats if all can be converted to it, but you can modify this behavior by adding colon-separated specifications `int` or `str` before the the list. (eg.: `batch_size=int:4,8,12,16`)
**Note:** sequences of increasing scalars are treated as a special case, easier to solve. Make sure to specify it this way when possible.
### Scalars
Scalars can be defined:
- through a commandline override with **`:`-separated values** defining a range (eg: `dropout=0:1`).
You can add specifications for log distributed values (eg.: `lr=log:0.001:1`) or integer values (eg.: `batch_size=int:4:8`)
or a combination of both (eg.: `batch_size=log:int:4:1024`)
- through a config files, with fields:
- `init`: optional initial value
- `lower` : optional lower bound
- `upper`: optional upper bound
- `log`: set to `true` for log distributed values
- `step`: optional step size for looking for better parameters. In linear mode this is an additive step, in logarithmic mode it
is multiplicative.
- `integer`: set to `true` for integers (favor floats over integers whenever possible)
Providing only `lower` and `upper` bound will set the initial value to the middle of the range, and the step to a sixth of the range.
**Note**: unbounded scalars (scalars with no upper and/or lower bounds) can only be defined through a config file.
|