This is an exact clone of the Neurips 2020 BBO challenge ended on October 15, 2020, formatted as an ever-lasting benchmark for research purposes, with a subset of the original challenge tasks.
The purpose is to evaluate black-box optimization algorithms on real-world objective functions. The problems chosen come from hyper-parameter (hp) selection/tuning in Machine Learning (ML) problems. The task submitted to the optimizer is:
maximize R(hp)
hp ∈ HP
where R(hp) = cross-validation-accuracy { dataset, ML-algorithm(hp) } and HP space includes conrinuous and discrete variables.
The participants must submit a Python class containing the optimizer, which consider R(hp) as a black-box, that is the optimizaer does NOT have access to the mathematical formula of R, all it can do is to query values of R at given points (which is time consuming).
The optimizer's class can include custom data members (e.g. storing past values of R) and must include at leaat two methods:
Hence:
S = search space = HP space
A = moves in mixed categorical and continuous space (suggest)
R = cross-validation-accuracy { dataset, ML-algorithm(hp) }
I = values of R at given points only (observe)
We provide the starting kit of the original challenge, which contains all the information needed to use this benchmark:
That's it, you can already submit your first submission! That's easy, right? However this submission is only a baseline solution. If you want to do better than this baseline, it is sufficient to modify optimizer.py. Optimizer submissions should follow the template, for a suggest-observe interface. Roughly speaking, you should modify suggest function and observe function, which are two important components of black box optimization algorithms.
We re-open the Neurips 2020 BBO challenge to make it an ever-lasting benchmark. There is only one phase: post-challenge phase, which allows to test black box algorithms on this benchmark. This particular instance of benchmark is limited to a single task (one dataset and one algorithm), specifically (GINA, MLP), where GINA is subset of MNIST and MLP is a fully connected multi-layer Perceptron trained with stochastic gradient descent.
The score by which the optimizer is evaluated id the cross-valiation accuracy.
This benchmark has NO prizes and is just for educational purposes.
The original terms of the BBO challenge terms and conditions do NOT apply.
Download | Size (mb) | Phase |
---|---|---|
Starting Kit | 0.548 | #1 Post-challenge |
Start: Jan. 24, 2021, midnight
Description: Only one phase: post-challenge phase
Never
You must be logged in to participate in competitions.
Sign In# | Username | Score |
---|---|---|
1 | guyon | 81.60 |