# A bit of benchmarking with string distances

September 7, 2013
By

(This article was first published on Mark van der Loo, and kindly contributed to R-bloggers)

After my last post about the stringdist package, Zachary Mayer pointed out to me that the implementation of the Levenshtein and Jaro-Winkler distances implemented in the RecordLinkage package are about two-three times faster. His benchmark compares randomly generated character strings of 5-25 characters, which probably covers many use cases
involving natural language. If your language happens to be denoted in single-byte encoding that is, but more on that below. Here's Zachary's benchmark (rerun by myself).

 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15  library(rbenchmark) library(RecordLinkage) library(stringdist)   > x <- sapply(sample(5:25,1e5,replace=TRUE), function(n) paste(sample(letters,n,replace=TRUE),collapse="")) > y <- sapply(sample(5:25,1e5,replace=TRUE), function(n) paste(sample(letters,n,replace=TRUE),collapse=""))   > benchmark( + d1 <- stringdist(x,y,method='lv') + , d2 <- levenshteinDist(x,y) + , replications=10 + ) test replications elapsed relative user.self 1 d1 <- stringdist(x, y, method = "lv") 10 4.939 1.876 4.924 2 d2 <- levenshteinDist(x, y) 10 2.633 1.000 2.632

Auch, indeed.

As we all know, premature optimization is the root of all evil so I felt it was time for some mature optimization.

Checking the differences between stringdist and RecordLinkage I found that RecordLinkage uses C code that works directly on the char representation of strings while I designed stringdist to make sure that multibyte characters are treated as a single character. This is done by converting them to integers before feeding them to my underlying C routines (using native R functionality). Hence, there is some conversion overhead that is linear in string length while the time spent on computing the Levenshtein distance grows with the product of the length of strings compared (say, quadratically). This is easily checked by benchmarking over longer strings.

 1 2 3 4 5 6 7 8 9 10 11 12  > x <- sapply(sample(25:250,1e5,replace=TRUE), function(n) paste(sample(letters,n,replace=TRUE),collapse="")) > y <- sapply(sample(25:250,1e5,replace=TRUE), function(n) paste(sample(letters,n,replace=TRUE),collapse="")) > > > benchmark( + d1 <- stringdist(x,y,method='lv') + , d2 <- levenshteinDist(x,y) + , replications=10 + ) test replications elapsed relative user.self 1 d1 <- stringdist(x, y, method = "lv") 10 111.668 1.258 111.584 2 d2 <- levenshteinDist(x, y) 10 88.784 1.000 88.716

And indeed, the relative difference decreased from a factor of 1.8 to 1.25. So, I rolled up my sleeves and built in a useBytes option that allows one to skip the conversion step. Here's the result on Zachary's original benchmark.

 1 2 3 4 5 6 7 8 9 10 11 12 13 14  > x <- sapply(sample(5:25,1e5,replace=TRUE), function(n) paste(sample(letters,n,replace=TRUE),collapse="")) > y <- sapply(sample(5:25,1e5,replace=TRUE), function(n) paste(sample(letters,n,replace=TRUE),collapse="")) > > > benchmark( + d1 <- stringdist(x,y,method='lv') + , d2 <- stringdist(x,y,method='lv',useBytes=TRUE) + , d3 <- levenshteinDist(x,y) + , replications=10 + ) test replications elapsed 1 d1 <- stringdist(x, y, method = "lv") 10 4.806 2 d2 <- stringdist(x, y, method = "lv", useBytes = TRUE) 10 1.463 3 d3 <- levenshteinDist(x, y) 10 2.942

Yowza! not only did I get rid of some overhead, stringdist is also about twice as fast as RecordLinkage in this benchmark. I haven't really profiled this but the main reason is probably that stringdist has a better memory allocation strategy[*]

We can also check this for the jaro-winkler distance, which is also implemented by RecordLinkage.

 1 2 3  1 d1 <- stringdist(x, y, method = "jw") 10 4.526 2 d2 <- stringdist(x, y, method = "jw", useBytes = TRUE) 10 0.513 3 d3 <- jarowinkler(x, y) 10 1.303

Again, stringdist is a factor of two faster.

But, there's always a but. To be fair we need to do the benchmark on longer strings as well, and that's where RecordLinkage comes back with a vengeance.

 1 2 3 4 5 6 7 8 9 10 11 12 13 14  > x <- sapply(sample(50:250,1e4,replace=TRUE), function(n) paste(sample(letters,n,replace=TRUE),collapse="")) > y <- sapply(sample(50:250,1e4,replace=TRUE), function(n) paste(sample(letters,n,replace=TRUE),collapse="")) > > > benchmark( + d1 <- stringdist(x,y,method='lv') + , d2 <- stringdist(x,y,method='lv',useBytes=TRUE) + , d3 <- levenshteinDist(x,y) + , replications=10 + ) test replications elapsed 1 d1 <- stringdist(x, y, method = "lv") 10 12.557 2 d2 <- stringdist(x, y, method = "lv", useBytes = TRUE) 10 11.972 3 d3 <- levenshteinDist(x, y) 10 10.091

Here, stringdist is 10-20 percent slower again. Again, I have not seriously profiled this but there is still some overhead in stringdist that's not in RecordLinkage. stringdist needs to select the algorithm, perform some more argument checking, and I still have to do a (very simple) conversion from char to int. Oh well, you can't win 'em all.

The main point for users is to decide whether the useBytes option is appropriate for the problem their trying to solve. The problem of working directly on the bytes in a character string is that for non-ASCII characters, the results depend on the character encoding scheme (and therefore most likely on the OS you're using.). Here's an example.

 1 2 3 4 5 6 7 8  # the function from RecordLinkage > levenshteinDist("ü","u") [1] 2 # stringdist > stringdist("ü","u",method='lv') [1] 1 > stringdist("ü","u",method='lv',useBytes=TRUE) [2] 2

On my system (linux) strings are encoded in utf8 where, u-umlaut is a 2-byte character[**]. Since RecordLinkage treats strings byte-by-byte it 'sees' two characters. It needs to substitute one and delete another yielding a distance value of 2. stringdist sees u-umlaut as a single character, but this may be turned off with the useBytes switch.

The bottom line is that it depends on the application which solution is best for your purpose. Here are some quick rules of thumb:

- Comparing strings with characters possibly outside of the ASCII range, with results that needs to be portable: stringdist
- Comparing single-byte encoded strings (like ASCII): for strings upto ~100 characters: stringdist with useBytes=TRUE. Otherwise RecordLinkage
- Distances other than Levenshtein or Jaro-Winkler: stringdist.

The latest version of stringdist (0.7.0) was accepted on CRAN yesterday (06.09.2013) and comes with the useBytes option in all functions. Also read the NEWS for bugfixes and new features.

The latest version always comes fully packed with the newest, most advanced bugs that offer game-changing insights[***] into the cranial processes of the programmer, so please don't forget to drop me a line when you find one.

[*]By the way, you can do really cool things with oprofile under linux if you want to profile C-code linked to R, but that's for another blog.
[**] well there are several ways to represent u-umlaut in utf8 but let's not get too technical.
[***] Buzzword bingo!

R-bloggers.com offers daily e-mail updates about R news and tutorials on topics such as: visualization (ggplot2, Boxplots, maps, animation), programming (RStudio, Sweave, LaTeX, SQL, Eclipse, git, hadoop, Web Scraping) statistics (regression, PCA, time series, trading) and more...