**r-bloggers on Machine Learning in R**, and kindly contributed to R-bloggers)

With the release of the new version of mlrMBO we added some minor fixes and added a practical feature called *Human-in-the-loop MBO*.

It enables you to sequentially

- visualize the state of the surrogate model,
- obtain the suggested parameter configuration for the next iteration and
- update the surrogate model with arbitrary evaluations.

In the following we will demonstrate this feature on a simple example.

First we need an objective function we want to optimize.

For this post a simple function will suffice but note that this function could also be an external process as in this mode **mlrMBO** does not need to access the objective function as you will only have to pass the results of the function to **mlrMBO**.

```
library(mlrMBO)
library(ggplot2)
set.seed(1)
fun = function(x) {
x^2 + sin(2 * pi * x) * cos(0.3 * pi * x)
}
```

However we still need to define the our search space.

In this case we look for a real valued value between -3 and 3.

For more hints about how to define ParamSets you can look here or in the help of ParamHelpers.

```
ps = makeParamSet(
makeNumericParam("x", lower = -3, upper = 3)
)
```

We also need some initial evaluations to start the optimization.

The design has to be passed as a `data.frame`

with one column for each dimension of the search space and one column `y`

for the outcomes of the objective function.

```
des = generateDesign(n = 3, par.set = ps)
des$y = apply(des, 1, fun)
des
## x y
## 1 -1.1835844 0.9988801
## 2 -0.5966361 0.8386779
## 3 2.7967794 8.6592973
```

With these values we can initialize our sequential MBO object.

```
ctrl = makeMBOControl()
ctrl = setMBOControlInfill(ctrl, crit = crit.ei)
opt.state = initSMBO(
par.set = ps,
design = des,
control = ctrl,
minimize = TRUE,
noisy = FALSE)
```

The `opt.state`

now contains all necessary information for the optimization.

We can even plot it to see how the Gaussian process models the objective function.

`plot(opt.state)`

In the first panel the *expected improvement* (\(EI = E(y_{min}-\hat{y})\)) (see Jones et.al.) is plotted over the search space.

The maximum of the *EI* indicates the point that we should evaluate next.

The second panel shows the mean prediction of the surrogate model, which is the Gaussian regression model aka *Kriging* in this example.

The third panel shows the uncertainty prediction of the surrogate.

We can see, that the *EI* is high at points, where the mean prediction is low and/or the uncertainty is high.

To obtain the specific configuration suggested by mlrMBO for the next evaluation of the objective we can run:

```
prop = proposePoints(opt.state)
prop
## $prop.points
## x
## 969 -2.999979
##
## $propose.time
## [1] 0.092
##
## $prop.type
## [1] "infill_ei"
##
## $crit.vals
## [,1]
## [1,] -0.3733677
##
## $crit.components
## se mean
## 1 2.8899 3.031364
##
## $errors.model
## [1] NA
##
## attr(,"class")
## [1] "Proposal" "list"
```

We will execute our objective function with the suggested value for `x`

and feed it back to mlrMBO:

```
y = fun(prop$prop.points$x)
y
## [1] 8.999752
updateSMBO(opt.state, x = prop$prop.points, y = y)
```

The nice thing about the *human-in-the-loop* mode is, that you don’t have to stick to the suggestion.

In other words we can feed the model with values without receiving a proposal.

Let’s assume we have an expert who tells us to evaluate the values \(x=-1\) and \(x=1\) we can easily do so:

```
custom.prop = data.frame(x = c(-1,1))
ys = apply(custom.prop, 1, fun)
updateSMBO(opt.state, x = custom.prop, y = as.list(ys))
plot(opt.state, scale.panels = TRUE)
```

We can also automate the process easily:

```
replicate(3, {
prop = proposePoints(opt.state)
y = fun(prop$prop.points$x)
updateSMBO(opt.state, x = prop$prop.points, y = y)
})
```

*Note:* We suggest to use the normal mlrMBO if you are only doing this as mlrMBO has more advanced logging, termination and handling of errors etc.

Let’s see how the surrogate models the true objective function after having seen seven configurations:

`plot(opt.state, scale.panels = TRUE)`

You can convert the `opt.state`

object from this run to a normal mlrMBO result object like this:

```
res = finalizeSMBO(opt.state)
res
## Recommended parameters:
## x=-0.22
## Objective: y = -0.913
##
## Optimization path
## 3 + 6 entries in total, displaying last 10 (or less):
## x y dob eol error.message exec.time ei
## 1 -1.1835844 0.9988801 0 NA
``` NA NA
## 2 -0.5966361 0.8386779 0 NA NA NA
## 3 2.7967794 8.6592973 0 NA NA NA
## 4 -2.9999793 8.9997519 4 NA NA -0.3733677
## 5 -1.0000000 1.0000000 5 NA NA -0.3136111
## 6 1.0000000 1.0000000 6 NA NA -0.1366630
## 7 0.3010828 1.0016337 7 NA NA -0.7750916
## 8 -0.2197165 -0.9126980 8 NA NA -0.1569065
## 9 -0.1090728 -0.6176863 9 NA NA -0.1064289
## error.model train.time prop.type propose.time se mean
## 1 NA initdesign NA NA NA
## 2 NA initdesign NA NA NA
## 3 NA initdesign NA NA NA
## 4 0 manual NA 2.8899005 3.0313640
## 5 0 manual NA 0.5709559 0.6836938
## 6 NA NA 3.3577897 5.3791930
## 7 0 manual NA 1.2337881 0.3493416
## 8 0 manual NA 0.4513106 0.8870228
## 9 0 manual NA 0.3621550 -0.8288961

*Note:* You can always run the *human-in-the-loop MBO* on `res$final.opt.state`

.

For the curious, let’s see how our original function actually looks like and which points we evaluated during our optimization:

```
plot(fun, -3, 3)
points(x = getOptPathX(res$opt.path)$x, y = getOptPathY(res$opt.path))
```

We can see, that we got pretty close to the global optimum and that the surrogate in the previous plot models the objective quite accurate.

For more in-depth information look at the Vignette for Human-in-the-loop MBO and check out the other topics of our mlrMBO page.

**leave a comment**for the author, please follow the link and comment on their blog:

**r-bloggers on Machine Learning in R**.

R-bloggers.com offers

**daily e-mail updates**about R news and tutorials on topics such as: Data science, Big Data, R jobs, visualization (ggplot2, Boxplots, maps, animation), programming (RStudio, Sweave, LaTeX, SQL, Eclipse, git, hadoop, Web Scraping) statistics (regression, PCA, time series, trading) and more...