Site icon R-bloggers

One Recipe Step to Rule Them All

[This article was first published on Posts on pRopaganda by smaakagen, and kindly contributed to R-bloggers]. (You can report issue about the content on this page here)
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.

In this post I will demonstrate, how my new R package customsteps can be used to create recipe steps, that apply custom transformations to a data set.

Note, you should already be fairly familiar with the recipes package before you continue reading this post or give customsteps a spin!

Recommended music for this reading session:

Introducing the customsteps package

Along with the recipes package distribution comes a number of pre-specified steps, that enables the user to manipulate data sets in various ways. The resulting data sets (/design matrices) can then be used as inputs into statistical or machine learning models.

If you want to apply a specific transformation to your data set, that is not supported by the pre-specified steps, you have two options. You can write an entire custom recipe step from scratch. This however takes quite a bit of work and code. An alternative – and sometimes better – approach is to apply the customsteps package, that I have just released on CRAN.

# install.packages("customsteps")
library(customsteps)

Customizable Higher-Order Steps

customsteps contains a set of customizable higher-order recipe step functions, that create specifications of recipe steps, that will transform or filter the data in accordance with custom input functions.

Let me just remind you of the definition of higher-order functions:

In mathematics and computer science, a higher-order function is a function that does at least one of the following: 1. takes one or more functions as arguments, 2. returns a function as its result.

Next, I will present an example of how to use the customsteps package in order to create a recipe step, that will apply a custom transformation to a data set.

Use Case: Centering and Scaling Numeric Data

Assume, that I want to transform a variable \({\mathbf{x}}\) like this:

  1. Center \({\mathbf{x}}\) around an arbitrary number \(\alpha\).
  2. Scale the transformed variable, such that its standard deviation equals an arbitrary number \(\beta\).

The transformed variable \(\hat{\mathbf{x}}\) can then be derived as (try to do it yourself):

\(\hat{\mathbf{x}} = \alpha + (\mathbf{x} – \bar{\mathbf{x}})\frac{\beta}{s_\mathbf{x}}\)

where \(\bar{\mathbf{x}}\) is the mean of \(\mathbf{x}\), and \(s_\mathbf{x}\) is the standard deviation of \({\mathbf{x}}\).

Note that centering \({\mathbf{x}}\) around 0 and scaling it in order to arrive at a standard deviation of 1 is just a special case of the above transformation with parameters \(\alpha = 0, \beta = 1\).

Write the prep helper function

First, I need to write a function, that estimates the relevant statistical parameters from an initial data set. I call this function the prep helper function.

Obviously, the above transformation requires the mean \(\bar{\mathbf{x}}\) and standard deviation \(s_\mathbf{x}\) to be learned from the initial data set. Therefore I define a function compute_means_sd, that estimates the two parameters for (any arbitrary number of) numeric variables.

By convention the prep helper function must take the argument x: the subset of selected variables from the initial data set.

library(purrr)

compute_means_sd <- function(x) {
  
  map(.x = x, ~ list(mean = mean(.x), sd = sd(.x)))

}

Let us see the function in action. I will apply it to a subset of the famous mtcars data set.

library(dplyr)

# divide 'mtcars' into two data sets.
cars_initial <- mtcars[1:16, ]
cars_new <- mtcars[17:nrow(mtcars), ]

# learn parameters from initial data set.
params <- cars_initial %>%
  select(mpg, disp) %>%
  compute_means_sd(.)

# display parameters. 
as.data.frame(params)
#>   mpg.mean  mpg.sd disp.mean disp.sd
#> 1     18.2 4.14761  250.8187 113.372

It works like a charm. Great, we are halfway there!

Write the bake helper function

Second, I have to specify a bake helper function, that defines how to apply the transformation to a new data set using the parameters estimated from the intial data set.

By convention the bake helper function must take the following arguments:

  • x: the new data set, that the step will be applied to.
  • prep_output: the output from the prep helper function containing any parameters estimated from the initial data set.

I define the function center_scale, that will serve as my bake helper function. It will center and scale variables of a new data set.

center_scale <- function(x, prep_output, alpha, beta) {

  # extract only the relevant variables from the new data set.
  new_data <- select(x, names(prep_output))

  # apply transformation to each of these variables.
  # variables are centered around 'alpha' and scaled to have a standard 
  # deviation of 'beta'.
  map2(.x = new_data,
       .y = prep_output,
       ~ alpha + (.x - .y$mean) * beta / .y$sd)
  
}

My first (sanity) check of the function is to apply it to the initial data set, that was used for estimation of the means and standard deviations.

library(tibble)
# center and scale variables of new data set to have a mean of zero
# and a standard deviation of one.
cars_initial_transformed <- center_scale(x = cars_initial, 
                                         prep_output = params,
                                         alpha = 0, 
                                         beta = 1)

# display transformed variables.
cars_initial_transformed %>%
  compute_means_sd(.) %>%
  as.data.frame(.)
#>       mpg.mean mpg.sd    disp.mean disp.sd
#> 1 1.731877e-16      1 7.199102e-17       1

Results are correct within computational precision.

Also, I will just check the function out on the other subset of mtcars.

# center and scale variables of new data set to have a mean of zero
# and a standard deviation of one.
cars_new_transformed <- center_scale(x = cars_new, 
                                     prep_output = params,
                                     alpha = 0, 
                                     beta = 1)

# display transformed variables.
cars_new_transformed %>%
  as.tibble(.) %>%
  head(.)
#> # A tibble: 6 x 2
#>      mpg   disp
#>    <dbl>  <dbl>
#> 1 -0.844  1.67 
#> 2  3.42  -1.52 
#> 3  2.94  -1.54 
#> 4  3.79  -1.59 
#> 5  0.796 -1.15 
#> 6 -0.651  0.593

Looks right! All that is left now is to put the pieces together into my new very own custom recipe step.

Putting the pieces together

The function step_custom_transformation takes prep and bake helper functions as inputs and turns them into a complete recipe step, that can be used out of the box.

I create the specification of the recipe step from the new functions compute_means_sd and center_scale by invoking step_custom_transformation.

library(recipes)
rec <- recipe(cars_initial) %>%
  step_custom_transformation(mpg, disp,
                             prep_function = compute_means_sd,
                             bake_function = center_scale,
                             bake_options = list(alpha = 0, beta = 1),
                             bake_how = "replace")

And that is all there is to it! Easy.

Note, by setting ‘bake_options’ to “replace”, the selected terms will be replaced with the transformed variables, when the recipe is baked.

I will just check, that the recipe works as expected. First I will prep(/train) the recipe.

# prep recipe.
rec <- prep(rec)

# print recipe.
rec
#> Data Recipe
#> 
#> Inputs:
#> 
#>   11 variables (no declared roles)
#> 
#> Training data contained 16 data points and no missing data.
#> 
#> Operations:
#> 
#> The following variables are used for computing transformations
#>  and will be dropped afterwards:
#>  mpg, disp

I will go right ahead and bake the new recipe.

# bake recipe.
cars_baked <- rec %>%
  bake(cars_new) %>%
  select(mpg, disp)
  
# display results.
cars_baked %>%
  head(.)
#> # A tibble: 6 x 2
#>      mpg   disp
#>    <dbl>  <dbl>
#> 1 -0.844  1.67 
#> 2  3.42  -1.52 
#> 3  2.94  -1.54 
#> 4  3.79  -1.59 
#> 5  0.796 -1.15 
#> 6 -0.651  0.593

Results are as expected (same as before). Great succes!

You should now be able to create your very own recipe steps to do (almost) whatever transformation you want to your data.

Conclusions

  • Customized recipe steps can be created with a minimum of effort and code using my new R package customsteps.
  • The customsteps step functions are higher-order functions, that create specificiations of recipe steps from custom input functions.

Please let me hear from you, if you have any feedback on the package.

Best, smaakagen

To leave a comment for the author, please follow the link and comment on their blog: Posts on pRopaganda by smaakagen.

R-bloggers.com offers daily e-mail updates about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.