A tensor is essentially a multidimensional array:
 a tensor of order one is a vector, which simply is a column of numbers,
 a tensor of order two is a matrix, which is basically numbers arranged in a rectangle,
 a tensor of order three looks like numbers arranged in rectangular box (or a cube, if all modes have the same dimension),
 an nth order (or nway) tensor looks like numbers arranged in… an hyperrectangle or something like that… you get the idea…
In many applications, data naturally form an nway tensor with n > 2, rather than a “tidy” table.
So, what’s a tensor decomposition?
Well, there are several types of tensor decomposition, but in this blog post I will introduce only the CANDECOMP/PARAFAC decomposition. Following Kolda & Bader (2009) I will refer to it as CP decomposition. But before spelling it out in mathematical terms, let’s start with a simple toy example using the R language.
Rank1 approximation to a 3way tensor (toy example)
Assume that we observe spatiotemporal^{1} data that, apart from random noise, correspond to the following three classes:
As you can see from the above figures, the essential difference between the three cases is that they behave differently with respect to time. The spatial component is just a Gaussian curve, while the temporal component is piecewise constant with a sudden jump at time 50, which differs in magnitude and in direction between the three classes.
Now assume we have collected samples that correspond to the three classes above (with some added noise). But we don’t know which sample falls in what class, how many classes there are, and how they differ :confused:. In an unsupervised fashion we want to learn the different classes and their differences with respect to time and space.
We can arrange the samples in a 3way tensor, samplebyspacebytime. For simplicity, however, assume that the samples are already grouped according to their class within the tensor (but the algorithm doesn’t know that!), so that the resulting tensor looks like this:
In R the above tensor (let’s call it X
) can be generated with the following lines of code:
space_index < seq(1, 1, l = 100)
bell_curve < dnorm(space_index, mean = 0, sd = 0.5)
case1 < matrix(rep(bell_curve, 10), 100, 100)
case2 < matrix(rep(bell_curve, 10), 100, 100)
case3 < matrix(rep(bell_curve, 10), 100, 100)
case2[ , 51:100] < case2[ , 51:100] + 0.1
case3[ , 51:100] < case3[ , 51:100]  0.1
X < array(NA, dim = c(90, 100, 100))
for(i in 1:30) {
X[i, , ] < case1 + matrix(rnorm(10000, sd = 0.1), 100, 100)
X[i+30, , ] < case2 + matrix(rnorm(10000, sd = 0.1), 100, 100)
X[i+60, , ] < case3 + matrix(rnorm(10000, sd = 0.1), 100, 100)
}
Using the excellent R package rTensor
we obtain the CP decomposition with one component per mode of the tensor:
library(rTensor)
cp_decomp < cp(as.tensor(X), num_components = 1)
str(cp_decomp$U)
# List of 3
# $ : num [1:90, 1] 0.0111 0.0111 0.0111 0.0111 0.0112 ...
# $ : num [1:100, 1] 0.00233 0.00251 0.00271 0.00292 0.00314 ...
# $ : num [1:100, 1] 0.00996 0.00994 0.00996 0.00993 0.00997 ...
# NULL
Visualizing the three components, we get the following figures:
We can clearly see that
 The samplespecific component has correctly separated the samples into three groups (in fact,
1:30
correspond to case 1,31:60
correspond to case 2,61:90
correspond to case 3).  The spatial component is bellshaped, just as the input data with respect to the spatial dimension.
 The temporal component has correctly picked up a change at time 50, which is where the three classes differ.
So, what did just happen? :astonished:
CP decomposition (quick summary of the math behind it)
The CP decomposition factorizes a tensor into a sum of outer products of vectors. For example, for a 3way tensor , the CP decomposition can be written as
where and , , are vectors of appropriate dimensions, and where the notation “” denotes the outer product for tensors, i.e.,
In case of the previous toy example we have that . Thus, the corresponding CP decomposition has the following form:
I hope that explains, why the components u
, v
, and w
in the toy example look the way they do! :bowtie:
Now, how do you solve for the components , , ()? You need to solve the following optimization problem:
where is the Frobenius norm.
The simplest way to do it is via an alternating least squares approach, where we would regard certain components as fixed while solving for others, and then iterate while alternating the components regarded as fixed… For much more rigour and detail see Kolda & Bader (2009) Tensor Decompositions and Applications.
A higherrank approximation via CP (toy example cont.)
Since in the previous toy example, there are no differentiating features between the three classes, apart from a jump in the temporal component, it makes perfect sense to set in CP. In order to try out CP with more than one component per mode, I generated data with a more complex structure, and with further differentiation between the three groups with respect to their temporal and spatial makeup:
 The three classes still have a bellshaped component in the spatial mode, but now each class has a different mean.
 In the temporal mode the data is shaped like a sine wave, with different scaling per class.
 As before, there is a sudden jump in the temporal mode at time 50.
The three classes in the resulting dataset have the following means.
As before, we generate a tensor X
of dimensions 90 × 100 × 100, with 30 samples per class obscured with random noise.
We use a CP decomposition in order to obtain a rank3 approximation to that tensor:
cp_decomp < cp(as.tensor(X), num_components = 3, max_iter = 100)
Here, we increase max_iter
to 100, in order to ensure convergence, as can be checked with the conv
attribute:
cp_decomp$conv
# [1] TRUE
Since we set num_components = 3
, the solution now has three components per mode, organized in a threecolumn matrix for each mode:
str(cp_decomp$U)
# List of 3
# $ : num [1:90, 1:3] 0.00131 0.00137 0.00141 0.0014 0.00135 ...
# $ : num [1:100, 1:3] 0.000926 0.001345 0.001799 0.002228 0.002755 ...
# $ : num [1:100, 1:3] 0.00969 0.0097 0.00974 0.0097 0.00971 ...
# NULL
And we can even check the percentage of the Frobenius norm of explained by the rank3 approximation :
cp_decomp$norm_percent
# [1] 83.13865
83% isn’t too bad!
:grinning: Let’s look at a visualization of the obtained components!
Indeed, we observe that,
 the samplespecific components clearly distinguish between the three groups of samples (
1:30
,31:60
, and61:90
),  the spatial components clearly picks up the Gaussian shapes with the three different means (at 0.5, 0, and 0.5),
 the temporal components clearly show the sudden jump at time 50, as well as a sine wave.
That’s it for now. In the next couple of weeks I am planning to write a couple blog posts on other types of tensor decompositions and tensor regression methods, as I am learning about them.

The two data modes can correspond to many types of measurements, other than space and time. Here, I use space and time for example purposes only because those are very familiar concepts. I am neither suggesting that specifically spatiotemporal data should be analyzed in this way, nor that tensor decomposition is generally a good approach for spatiotemporal data (I actually have no idea). ↩
Rbloggers.com offers daily email updates about R news and tutorials on topics such as: Data science, Big Data, R jobs, visualization (ggplot2, Boxplots, maps, animation), programming (RStudio, Sweave, LaTeX, SQL, Eclipse, git, hadoop, Web Scraping) statistics (regression, PCA, time series, trading) and more...