Using Google BigQuery with R

[This article was first published on julianhi's Blog » R Tutorials, and kindly contributed to R-bloggers]. (You can report issue about the content on this page here)
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.

BigQuery header

Hey everybody,

this tutorial is about combining two great and powerful tools: R and Google BigQuery.

Big Query:

So but what is BigQuery? I think Google describes it very well on the developers page:
“Querying massive datasets can be time consuming and expensive without the right hardware and infrastructure. Google BigQuery solves this problem by enabling super-fast, SQL-like queries against append-only tables, using the processing power of Google’s infrastructure. Simply move your data into BigQuery and let us handle the hard work. You can control access to both the project and your data based on your business needs, such as giving others the ability to view or query your data.” (source:
 And BigQuery is fast. BigQuery is extremely fast but you will see that later when we query some sample data.

Getting Ready

But before we can enjoy the speed we need to do some work. Maybe “work” is the wrong way as using BigQuery is as simple as possible. First we need to create a project for our test in the Google Developers Console

Project BigQuery

You can name  it however you want to and so you can do with the ProjectID. It might shouldn´t be too complicated as we need it later.  In the next step we have to activate the BigQuery API for our Project what we can do when we click on “APIs” on the left-hand site.

API sidebar

Let´s bring it to R

And the R part is as easy as the Google part because the awesome Hadley Wickham wrote a nice and easy package called “bigrquery“.

Start R and type in:


Ok now we installed the necessary packages and can do our first query.


project <- "fleet-standard-515" # put your projectID here

sql <- 'SELECT title,contributor_username,comment FROM[publicdata:samples.wikipedia] WHERE title CONTAINS "beer" LIMIT 100;'

data <- query_exec("publicdata", "samples", sql, billing = project)

Ok what happens here?

First we set our ProjectID we chose when we created the project in the Google Developer Console.

In the next step we set our SQL query which we execute in the last row and save the result in the variable “data”.

In this case I query a sample dataset containing a list of wikipedia entries.

This dataset has 313,797,035 rows and a total size 35,7 GB! And you can query it in under 2 seconds!

This is a good example for showing the speed of BigQuery. We want the fields “title, contributor_username, comment” from every row in the dataset where the title contains “beer”. We then limit the output to 100 entries because otherwise in this example we would receive something about 360,00 entries with the word beer in the title (a huge amount).

A tab in browser will then open where you have to login into your Google account and give your R instance the permission to access your project. After the authentication you can see the result of your query in R.

data table

BigQuery speed

To leave a comment for the author, please follow the link and comment on their blog: julianhi's Blog » R Tutorials. offers daily e-mail updates about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.

Never miss an update!
Subscribe to R-bloggers to receive
e-mails with the latest R posts.
(You will not see this message again.)

Click here to close (This popup will not appear again)