416 search results for "hadoop"

How Orbitz uses Hadoop and R to optimize hotel search

December 21, 2010
By
How Orbitz uses Hadoop and R to optimize hotel search

Positional bias — the tendency for users to preferentially select results in the first few positions of a search — is a big issue for all kinds of search engines. But for online travel site Orbitz the stakes are higher than for a traditional Web search engine: if a customer chooses the first-listed hotel in a search for accommodations,...

Read more »

What’s Next for Revolution R and Hadoop?

November 30, 2010
By

It's been a busy fall season for the team at Revolution Analytics. Over the past few months, we've announced major product enhancements for Revolution R -- RevoScaleR, for tackling big data sets, and RevoDeployR, for embedding Revolution R into wider applications. We've continued to add to our growing customer base at an aggressive rate and we've been busy crisscrossing...

Read more »

Using R and Hadoop to analyze VOIP data

November 8, 2010
By

Last month, the newest member of Revolution's engineering team, Saptarshi Guha, gave a presentation at Hadoop World 2010 on using R and Hadoop to analyze 1.3 billion voice-over-IP packets to identify calls and measure call quality. Saptarshi, of course, is the author of RHIPE, which lets R programmers write map-reduce algorithms in the Hadoop framework without needing to learn...

Read more »

Saptarshi Guha on Hadoop, R

September 20, 2010
By

Saptarshi Guha (author of the Rhipe package) joins the likes of Ebay, Yahoo, Twitter and Facebook and as one of just 37 presenters at the Hadoop World conference. (Revolution Analytics is proud to sponsor Saptarshi's presence at this event, which take place in New York on October 12.) He'll be talking about using R and Hadoop to analyze Voice-over-IP...

Read more »

My Experience at Hadoop Summit 2010 #hadoopsummit

June 30, 2010
By
My Experience at Hadoop Summit 2010 #hadoopsummit

This week I had the opportunity the trek up north to Silicon Valley to attend Yahoo’s Hadoop Summit 2010. I love Silicon Valley. The few times I’ve been there the weather was perfect (often warmer than LA), little to no traffic, no road rage and people overall seem friendly and happy. Not to mention there are so many trees...

Read more »

You can Hadoop it! It’s elastic! Boogie woogie woog-ie!

February 16, 2010
By
You can Hadoop it! It’s elastic! Boogie woogie woog-ie!

I just came back from the future and let me be the first to tell you this: Learn some Chinese. And more than just cào nǐ niáng (肏你娘) which your friend in grad school told you means “Live happy with many blessings”. Trust me, I’ve been hanging with Madam Wu and she told me

Read more »

Streaming Hadoop Data Into R Scripts

March 23, 2009
By
Streaming Hadoop Data Into R Scripts

Along the lines of Mongo Measurement Requires Mongo Management, the HadoopStreaming package on CRAN provides utilities for applying R scripts to Hadoop streaming. Hadoop is used on Amazon's EC2.

Read more »

In case you missed it: July 2015 roundup

August 10, 2015
By

In case you missed them, here are some articles from July of particular interest to R users. An alternative to stacked bar charts with the streamgraphs package. Joseph Rickert shares his process for creating the monthly new and updated packages "spotlight" feature on MRAN. Using R to analyze data from its API reveals R to be the 8th most...

Read more »

partools: a Sensible R Package for Large Data Sets

August 5, 2015
By
partools: a Sensible R Package for Large Data Sets

As I mentioned recently, the new, greatly extended version of my partools package is now on CRAN. (The current version on CRAN is 1.1.3, whereas at the time of my previous announcement it was only 1.1.1. Note that Unix is NOT required.) It is my contention that for most R users who work with large … Continue reading...

Read more »

Getting Started: Adobe Analytics Clickstream Data Feed

August 4, 2015
By
Getting Started: Adobe Analytics Clickstream Data Feed

“Well, first you need a TMS and a three-tiered data layer, then some jQuery with a node backend to inject customer data into the page asynchronously if you want to avoid cookie-based limitations with cross-domain tracking and be Internet Explorer 4 compatible…” Blah Blah Blah. There’s a whole cottage industry around jargon-ing each other to

Read more »