Necessity is the mother of invention and I had the opportunity today to take an R package from development to CRAN in less than the span of 24 hours.
Despite being on vacation, I answered an R question on StackOverflow pertaining to the use of
decode_short_url from the
twitteR::decode_short_url function uses the LongURL service, which has a dirt simple API to both get a list of all long URL domains it knows about and expand a shortened URL (if it can). I wrapped this API into a small R package right after I answered the SO question and proceded to add the finishing touches to be able to, then, submit it to CRAN. Attempt #1 failed, but I amended the
DESCRIPTION file after very helpful suggestion from the volunteer CRAN maintainers and it’s now on CRAN. That makes CRAN in Par 2!
As Jay pointed out to me later in the day, the attention to detail by the CRAN Guardians is one of the things that helps maintain super-high quality in the R community and truly sets us apart from those “other” data science languages (I might have taken some liberties with Jay’s original quote to me).
Why do we need longurl?
I’ll point readers to a paper—Two Years of Short URLs Internet Measurement: Security Threats and Countermeasures [PDF]—by Maggi, Frossi, Zanero, Stringhini, Stone-Gross, Kruegel & Vigna where the authors look at the potential (and actual) evil behind short URLs. Many things can hide there, from malware to phishing sites and knowing both the short and full URL can help defenders stop attacks before they are fully successful.
How to use longurl
I took a sampling of
t.co referer domains from some
datadrivensecurity.info weblogs (you can grab that here) to show how to use this and other packages to parse weblogs, expand URLs and extract various bits of info from them, soley with R. I’ll be using packages from myself, Oliver Keyes, Jay Jacobs & Hadley Wickham to accomplish this task.
All of the packages can be
install.packages from CRAN, except for
webtools. For that, you can just
Let’s get package loading out of the way:
library(webtools) library(dplyr) library(stringr) library(longurl) library(urltools)
Oliver made it super-easy to read in web logs. I use the “combined” common log format (CLF) on the blog’s web server, which can be parsed in one line of R:
log <- read_combined("web.log", has_header=FALSE) glimpse(log) ## Observations: 484 ## Variables: ## $ ip_address (chr) "18.104.22.168", "22.214.171.124", "36.80.104... ## $ remote_user_ident (chr) NA, NA, NA, NA, NA, NA, NA, NA, NA, NA, NA, ... ## $ local_user_ident (chr) NA, NA, NA, NA, NA, NA, NA, NA, NA, NA, NA, ... ## $ timestamp (time) 2015-06-14 02:07:49, 2015-06-13 01:35:00, 2... ## $ request (chr) "GET /blog/posts/2014/Dec/ponemon/ HTTP/1.1"... ## $ status_code (int) 200, 200, 200, 200, 301, 200, 200, 301, 200,... ## $ bytes_sent (int) 10036, 6404, 22120, 6283, 185, 6132, 23667, ... ## $ referer (chr) "http://t.co/G4XiI9USB3", "http://t.co/j9Rmm... ## $ user_agent (chr) "Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWe...
To get some short URLs to expand we only need focus the
referer for this example, so let’s:
- remove all the query strings (this is just an example, after all)
- only work on the unique ones
- expand them
Here’s the code:
log %>% mutate(referer=str_replace(referer, "\?.*$", "")) %>% distinct(referer) %>% .$referer %>% expand_urls(check=FALSE, warn=FALSE) -> referers glimpse(referers) ## Observations: 61 ## Variables: ## $ orig_url (chr) "http://t.co/G4XiI9USB3", "http://t.co/j9RmmOY9Kr", "http:... ## $ expanded_url (chr) "http://datadrivensecurity.info/blog/posts/2014/Dec/ponemo...
Now that we have a nice set of expanded URLs, we can parse them into their components:
parsed_refs <- url_parse(referers$expanded_url)
We went from 484 potential URLs to shorten to 61 (after de-duping).
Please be kind to the LongURL service and also note that parsing huge lists of URLs can take a while, especially if you turn on validity checking. You’ll at least get a free progress bar when using an interactive session (unless you disable it).
glimpse(parsed_refs) ## Observations: 61 ## Variables: ## $ scheme (chr) "http", "http", "http", "http", "http", "http", "http", "http... ## $ domain (chr) "datadrivensecurity.info", "datadrivensecurity.info", "datadr... ## $ port (chr) "", "", "", "", "", "", "", "", "", "", "", "", "", "", "", "... ## $ path (chr) "blog/posts/2014/dec/ponemon/", "blog/", "podcast/data-driven... ## $ parameter (chr) "", "", "", "ie=utf8&qid=1374598875&sr=8-1&keywords=best+secu... ## $ fragment (chr) "", "", "", "", "", "", "gsc.tab=0&gsc.q=hosting", "", "", ""...
With parsed URLs in hand we can can proceed with any other bits of analysis, such as seeing the top domains (which is, unsurprisingly, this very blog):
sort(table(parsed_refs$domain)) ## de.buyvip.com iosappstar.blogspot.co.at serv.adwingate.com ## 1 1 1 ## sony.attributed.to sports.bwin.com www.amazon.com ## 1 1 1 ## www.godaddy.com www.google.com dds.ec ## 1 1 2 ## www.netnews.at datadrivensecurity.info ## 2 49
This was (obviously) a trivial example to get you started on using some of these truly helpful packages when doing URL/domain analysis in R. These URL/domain components can further be used to develop features for machine learning pipelines, metrics/reports or even forensic investigations. If you have other helpful R packages for the cybersecurity domain or use
longurl or any of the other packages in an interesting way, drop a note in the comments.