# Beyond Basic R – Data Munging

**The USGS OWI blog**, and kindly contributed to R-bloggers]. (You can report issue about the content on this page here)

Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.

## What we couldn’t cover

In the data cleaning portion of our Intro to R class, we cover a variety of common data manipulation tasks. Most of these were achieved using the package `dplyr`

, including removing or retaining certain columns (`select`

), filtering out rows by column condition (`filter`

), creating new columns (`mutate`

), renaming columns (`rename`

), grouping data by categorical variables (`group_by`

), and summarizing data (`summarize`

).

But often, data users need to do more complex manipulation of their data, like changing the shape of the data or creating a new column conditional on values in another column. In this post, we will use USGS stream data to demonstrate data munging techniques that are beyond the basics taught in our Intro to R course. If you’re following along and see a new function and want to learn more, a reminder that you can see the help file for that function by executing a `?`

followed by the function name in your console to see the help file and examples.

## What we want to achieve

In this example, a data user is interested in how nutrient concentrations are related to discharge in a stream within a mixed use watershed of urban and agricultural land use. The site is located in Dane County near Madison, WI, and we’re focusing on daily values from a 20-year period from 1998 to 2017. We’ll use these data to demonstrate a variety of data munging techniques that are beyond the basics.

## How to get there

### Get the data

The first step falls within what we’ve already learned in the Intro to R class: pull data from NWIS using the USGS package `dataRetrieval`

, select the columns we want to keep, and rename columns.

# load libraries library(dplyr) library(dataRetrieval) library(lubridate) library(tidyr) library(ggplot2) library(viridis) # Get data for the Yahara River at Windsor in Dane County, Wisconsin yahara_no <- '05427718' # define parameters of interest, and get those parameter names params <- c('00060', '00671', '80154', '00665') # get daily values from NWIS yahara_dat <- readNWISdv(siteNumbers = yahara_no, parameterCd = params, startDate = '1997-10-01', endDate = '2017-09-30') # rename columns using renameNWISColumns from package dataRetrieval yahara_dat <- renameNWISColumns(yahara_dat, p00665 = "TP_mgL", p00671 = "Orthophosphate_mgL", p80154 = "SS_mgL")

You may be seeing the pipe (`%>%`

) for the first time in the above code snippet. The pipe comes from the package `magrittr`

and is used in `dplyr`

to connect commands. `dplyr`

functions generally take arguments in the same order: the dataframe you want to manipulate, and what you want to do it. A pipe upstream of a command “pipes in” the results of the upstream function; that is, you no longer have to specify the dataframe you want to manipulate. It’s an efficient way to code if you don’t care to create intermediate data objects. You can read more about the history and functionality of the pipe in R from DataCamp’s tutorial on pipes.

### Match strings

Notice that in our new dataset, we have retained code columns, or those that end in “_cd”. How can we quickly identify which columns contain that string? Here, we’ll introduce you to pattern matching functions (including `grep`

and `gsub`

) which have wide application to data munging. First, we’ll use grep to return the index of the string or the string itself that contains “_cd”.

# use function `grep` to identify which columns are code columns yahara_names <- names(yahara_dat) grep('_cd', yahara_names) # returns the index of the match ## [1] 1 5 7 9 11 grep('_cd', yahara_names, value = TRUE) # returns the matched elements themselves ## [1] "agency_cd" "Flow_cd" "TP_mgL_cd" ## [4] "Orthophosphate_mgL_cd" "SS_mgL_cd" # change the code column names to be more explicit about what they contain # using function gsub gsub('_cd', '_code', yahara_names) ## [1] "agency_code" "site_no" ## [3] "Date" "Flow" ## [5] "Flow_code" "TP_mgL" ## [7] "TP_mgL_code" "Orthophosphate_mgL" ## [9] "Orthophosphate_mgL_code" "SS_mgL" ## [11] "SS_mgL_code"

The package `dplyr`

also has some handy pattern matching functions built in. For example, we can use the function `contains`

in `select`

to exclude those columns with the “_cd” string.

yahara_dat <- select(yahara_dat, -contains('_cd')) head(yahara_dat) ## site_no Date Flow TP_mgL Orthophosphate_mgL SS_mgL ## 1 05427718 1997-10-01 11 0.05 NA 18.8 ## 2 05427718 1997-10-02 11 0.05 NA 18.6 ## 3 05427718 1997-10-03 12 0.05 NA 18.5 ## 4 05427718 1997-10-04 11 0.04 NA 18.5 ## 5 05427718 1997-10-05 11 0.04 NA 18.4 ## 6 05427718 1997-10-06 11 0.04 NA 18.4

These functions use regular expression to evaluate matches, and can be used to find much more complex patterns. We’ve only scratched the surface here on what can be done with pattern matching. Further reading on the topic of regular expression in R that we recommend include RStudio’s regular expression cheat sheet and the “strings” section from the R for Data Science book.

### Create new variables to summarize by water year

First, let’s compare discharge across water years. The current data frame only has dates, so we’ll need to create a water year and a day of water year variable to use in our comparisons. We can use the handy `addWaterYear`

function from `dataRetrieval`

to do so. Next, we’ll calculate cumulative discharge by day for each year using `group_by`

and `mutate`

.

# add water year variable "waterYear" to our dataframe yahara_dat <- addWaterYear(yahara_dat) # calculate cumulative discharge for each year by first grouping by water year, # and then using the "cumsum" function. Add day of water year for plotting purposes. # These steps will build a new dataframe, with the existing information in yahara_dat # but with two additional columns. cumulative_dat <- group_by(yahara_dat, waterYear) %>% mutate(cumulative_dis = cumsum(Flow), wy_doy = seq(1:n())) # visually compare cumulative discharge across years ggplot(cumulative_dat, aes(x = wy_doy, y = cumulative_dis, group = waterYear)) + geom_line(aes(color = waterYear)) + scale_color_viridis_c() + scale_x_continuous(breaks = c(1, 93, 184, 275), labels = c("Oct 1", "Jan 1", "Apr 1", "July 1")) + theme_bw() + labs(color = "Water Year", x = "", y = "Cumulative Discharge")

### Create categorical variables from continuous data

There is variability year-to-year in cumulative discharge and the timing of large discharge events. As we look at the nutrient ~ discharge relationships, we may want to keep track of which years were wet versus dry. That is, we want to categorize years based on the total annual discharge. Below, we show several ways to create a categorical variable from continuous data.

# sum discharge by water year, also check to be sure each year has # data for every day by counting observations in a year year_sums <- yahara_dat %>% filter(!is.na(Flow)) %>% group_by(waterYear) %>% summarize(yearly_discharge = sum(Flow, na.rm = TRUE), ndays = n())

First, we’ll use an `ifelse`

statement to create a categorical variable based on TRUE or FALSE conditions.

# use ifelse to create a categorical variable that has two conditions median_yearly_discharge <- median(year_sums$yearly_discharge) year_sums <- mutate(year_sums, discharge_high_low = ifelse(yearly_discharge <= median_yearly_discharge, 'low', 'high')) # we should get equal group sizes of "low" and "high" summary(as.factor(year_sums$discharge_high_low)) ## high low ## 10 10

Now suppose you want to create a variable with three categories: dry, normal, and wet. We’ll use the `cut`

function to do this so we can give multiple numeric break points to create the categories. We’ll arbitrarily use the 25th and 75th percentiles to define our cutoffs values.

cut_vals <- quantile(year_sums$yearly_discharge, probs = c(0.25, 0.75), na.rm = TRUE) year_sums <- mutate(year_sums, discharge_dry_normal_wet = cut(yearly_discharge, breaks = c(-Inf, cut_vals, Inf), labels = c("dry", "normal", "wet"))) # we expect roughly twice as many "normal" years as "wet" or "dry" years summary(as.factor(year_sums$discharge_dry_normal_wet)) ## dry normal wet ## 5 10 5

In some instances, you may want to create categorical variables based on a more complex set of rules. For example, if there was an important management action in this watershed that occurred in the year 2000, you may want your low/high discharge categories to be further classified as before/after. We’ll use the function `case_when`

to use multiple rules to define our categories.

year_sums <- year_sums %>% mutate(discharge_before_after = case_when( yearly_discharge > median_yearly_discharge & waterYear > 2000 ~ "High Discharge, After", yearly_discharge > median_yearly_discharge & waterYear <= 2000 ~ "High Discharge, Before", yearly_discharge <= median_yearly_discharge & waterYear <= 2000 ~ "Low Discharge, Before", yearly_discharge <= median_yearly_discharge & waterYear > 2000 ~ "Low Discharge, After" )) # check how many years of data fall into each category summary(as.factor(year_sums$discharge_before_after)) ## High Discharge, After Low Discharge, After Low Discharge, Before ## 10 7 3

### Gather or stack variables in long data frame using `gather`

We can plot the variables through time to see major discharge and concentration events. With our current data structure, we could easily create individual plots of all constituents and discharge through time - e.g., this plot of TP through time:

ggplot(cumulative_dat, aes(x = Date, y = TP_mgL)) + geom_point() + labs(x = "Date", y = 'TP (mg/L)') + theme_bw()

But what if we wanted to stack this figure with a discharge through time plot, where the panels are aligned by date and share x-axis labels? We can do this by faceting in `ggplot`

, but our current data structure is wide (each variable is in its own column), and we need a single “variable” column with an associated “value” that is presented in long format. To get our data in long format, we will `gather`

(from the package `tidyr`

) the nutrient and discharge columns.

For the `gather`

function, we need to provide a key name which will become the column where the measured parameter names are stored. We also provide a value name, which will become the column where the measured values are stored. Finally, we tell the gather function which columns we do (by listing) or do not (by list with a negative sign) want to gather by adding a negative sign to those column names.

yahara_long <- cumulative_dat %>% gather(key = measured_parameter, value = value, Flow, TP_mgL, Orthophosphate_mgL, SS_mgL) head(yahara_long) ## # A tibble: 6 x 7 ## # Groups: waterYear [1] ## site_no Date waterYear cumulative_dis wy_doy measured_parameter ## <chr> <date> <dbl> <dbl> <int> <chr> ## 1 05427718 1997-10-01 1998 11 1 Flow ## 2 05427718 1997-10-02 1998 22 2 Flow ## 3 05427718 1997-10-03 1998 34 3 Flow ## 4 05427718 1997-10-04 1998 45 4 Flow ## 5 05427718 1997-10-05 1998 56 5 Flow ## 6 05427718 1997-10-06 1998 67 6 Flow ## # ... with 1 more variable: value <dbl>

Now, we can use `ggplot`

to plot all values against time and `facet_wrap`

by the “variable” column to create panels of data.

ggplot(yahara_long, aes(x = Date, y = value)) + geom_point() + facet_wrap(~measured_parameter, ncol = 1, scales = 'free_y') + labs(x = "Date", y = '') + theme_bw()

If you’re following along and executing the code, you may see a warning after the last `ggplot`

command, stating you’ve removed rows due to missing data. Not all variables had the same temporal coverage, and are therefore “missing” during certain time periods. The plot still should have generated, and the message serves as an “FYI” to what was going on behind the scenes.

### Spreading data to pair observations using `spread`

We can see that large discharge events are generally associated with increases in concentration, but we suspect that dissolved and particulate nutrients may behave differently during large events. Dissolved P is of particular concern in the basin as the downstream lakes are eutrophic, and so we are interested in the relative amount of total phosphorus that is entering the systems as dissolved P, and how that ratio is affected by the magnitude of the runoff event. We can also evaluate these relationships in the context of wet and dry years, which we determined in the steps above.

The first step is to create a TP:OrthoP variable, but in our long dataset, our two P variables are no longer paired by date (i.e., row). Let’s spread our dataset back out into wide format using the `spread`

function from package `tidyr`

.

yahara_wide <- yahara_long %>% spread(key = measured_parameter, value = value) %>% mutate(TP_OrthoP = TP_mgL/Orthophosphate_mgL) # again, gather the phosphorus variables # merge with our wet/dry variable of interest p_dat <- ungroup(yahara_wide) %>% select(Date, TP_mgL, Orthophosphate_mgL, TP_OrthoP, Flow, waterYear) %>% gather(key = P_variables, value = value, -Date, -Flow, -waterYear) %>% left_join(select(year_sums, discharge_dry_normal_wet, waterYear)) # visualize phosphorus ~ discharge relationships and color code # by wet/dry year ggplot(p_dat, aes(x = Flow, y = value)) + geom_point(aes(color = discharge_dry_normal_wet), alpha = 0.3) + facet_wrap(~P_variables, ncol = 1, scales = 'free_y') + geom_smooth(aes(group = discharge_dry_normal_wet, color = discharge_dry_normal_wet), alpha = 0.5, method = 'lm') + scale_y_log10() + scale_x_log10() + annotation_logticks() + theme_bw() + labs(x = "Discharge (cfs)", color = "Annual Hydrologic \nCondition")

### Operations by row using `rowwise`

In some instances, you may want make a calculation or manipulate data by row. Many functions, such as `min`

and `max`

, operate on the column. For example, there is an another USGS site on the Yahara, and you want to know which site has greater TP concentrations on a daily scale. Below, we’ll grab the additional data, merge it to the original site data, and calculate which site has the maximum TP value each day. We’ll use the function `which.max`

, which returns the index of the maximum value. We’ll use that index to return the name of the site with the highest value.

# set USGS site and pull data yahara_site2 <- '05427850' yahara_dat2 <- readNWISdv(siteNumbers = yahara_site2, parameterCd = '00665', startDate = '1997-10-01', endDate = '2017-09-30') %>% select(Date, TP_Yah_Hwy113 = X_00665_00003) # merge the two Yahara TP datasets yahara_combo <- left_join(yahara_dat2, select(yahara_dat, TP_Yah_Windsor = TP_mgL, Date)) # find which site has has the maximum value by row yahara_highest <- yahara_combo %>% rowwise() %>% mutate(max_tp_site = c('Yahara_at_Hwy113', 'Yahara_at_Windsor')[which.max(c(TP_Yah_Hwy113, TP_Yah_Windsor))]) summary(as.factor(yahara_highest$max_tp_site)) ## Yahara_at_Hwy113 Yahara_at_Windsor ## 1527 1759

### Happy data munging!

We hope these examples can help you navigate common data munging tasks in R. There are certainly many more tricks, tips, and functions that can be used for data munging, but we hope that our Intro to R material on data cleaning, along with this post, have given you a jump start.

**leave a comment**for the author, please follow the link and comment on their blog:

**The USGS OWI blog**.

R-bloggers.com offers

**daily e-mail updates**about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.

Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.