File: interruptible-optimization.py

package info (click to toggle)
scikit-optimize 0.10.2-4
  • links: PTS, VCS
  • area: main
  • in suites: forky, sid, trixie
  • size: 7,672 kB
  • sloc: python: 10,659; javascript: 438; makefile: 136; sh: 6
file content (132 lines) | stat: -rw-r--r-- 4,683 bytes parent folder | download | duplicates (2)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
"""
================================================
Interruptible optimization runs with checkpoints
================================================

Christian Schell, Mai 2018
Reformatted by Holger Nahrstaedt 2020

.. currentmodule:: skopt

Problem statement
=================

Optimization runs can take a very long time and even run for multiple days.
If for some reason the process has to be interrupted results are irreversibly
lost, and the routine has to start over from the beginning.

With the help of the :class:`callbacks.CheckpointSaver` callback the optimizer's current state
can be saved after each iteration, allowing to restart from that point at any
time.

This is useful, for example,

* if you don't know how long the process will take and cannot hog computational resources forever
* if there might be system failures due to shaky infrastructure (or colleagues...)
* if you want to adjust some parameters and continue with the already obtained results

"""

print(__doc__)
import numpy as np

np.random.seed(777)

#############################################################################
# Simple example
# ==============
#
# We will use pretty much the same optimization problem as in the
# :ref:`sphx_glr_auto_examples_bayesian-optimization.py`
# notebook. Additionally we will instantiate the :class:`callbacks.CheckpointSaver`
# and pass it to the minimizer:

from skopt import gp_minimize
from skopt.callbacks import CheckpointSaver

noise_level = 0.1


def obj_fun(x, noise_level=noise_level):
    return np.sin(5 * x[0]) * (1 - np.tanh(x[0] ** 2)) + np.random.randn() * noise_level


checkpoint_saver = CheckpointSaver("./checkpoint.pkl", compress=9) # kwargs passed to `skopt.dump`

gp_minimize(
    obj_fun,  # the function to minimize
    [(-20.0, 20.0)],  # the bounds on each dimension of x
    x0=[-20.0],  # the starting point
    acq_func="LCB",  # the acquisition function (optional)
    n_calls=10,  # number of evaluations of f including at x0
    n_random_starts=3,  # the number of random initial points
    callback=[checkpoint_saver],
    # a list of callbacks including the checkpoint saver
    random_state=777,
)

#############################################################################
# Now let's assume this did not finish at once but took some long time: you
# started this on Friday night, went out for the weekend and now, Monday
# morning, you're eager to see the results. However, instead of the
# notebook server you only see a blank page and your colleague Garry
# tells you that he had had an update scheduled for Sunday noon – who
# doesn't like updates?
#
# :class:`gp_minimize` did not finish, and there is no `res` variable with the
# actual results!
#
# Restoring the last checkpoint
# =============================
#
# Luckily we employed the :class:`callbacks.CheckpointSaver` and can now restore the latest
# result with :class:`skopt.load`
# (see :ref:`sphx_glr_auto_examples_store-and-load-results.py` for more
# information on that)

from skopt import load

res = load('./checkpoint.pkl')

res.fun

#############################################################################
# Continue the search
# ===================
#
# The previous results can then be used to continue the optimization process:

x0 = res.x_iters
y0 = res.func_vals

# To ensure that the base estimator is loaded properly and that the next 
# parameters are new ones:
base_estimator = res.specs['args']['base_estimator']
random_state = res.random_state

gp_minimize(
    obj_fun,  # the function to minimize
    [(-20.0, 20.0)],  # the bounds on each dimension of x
    base_estimator=base_estimator, # warm-started base-estimator from checkpoint
    x0=x0,  # already examined values for x
    y0=y0,  # observed values for x0
    acq_func="LCB",  # the acquisition function (optional)
    n_calls=10,  # number of evaluations of f including at x0
    n_random_starts=3,  # the number of random initialization points
    callback=[checkpoint_saver],
    random_state=random_state,
)

#############################################################################
# Possible problems
# =================
#
# * **changes in search space:** You can use this technique to interrupt
#   the search, tune the search space and continue the optimization. Note
#   that the optimizers will complain if `x0` contains parameter values not
#   covered by the dimension definitions, so in many cases shrinking the
#   search space will not work without deleting the offending runs from
#   `x0` and `y0`.
# * see :ref:`sphx_glr_auto_examples_store-and-load-results.py`
#
# for more information on how the results get saved and possible caveats