Sessions at Strange Loop 2011 about Big Data in Lindbergh

Your current filters are…


Tuesday 20th September 2011

  • Distributed Data Analysis with Hadoop and R

    by Jonathan Seidman and Ramesh Venkataramaiah

    The R programming language has become a standard environment for statistical computing, but out of the box R is restricted to analysis on data sets that fit in memory. Hadoop has become a popular platform for storing and analyzing data sets that are too large to fit on a single machine. Not surprisingly, there’s significant interest in bringing these two platforms together to perform sophisticated analysis on data that’s too large to fit in memory on a single machine. Although there are several systems being developed by the R community to support this such as Ricardo and RHIPE, as well as newer interfaces such as Segue and Hadoop InteractiVE, there’s still considerable confusion as to how to effectively use these two systems together. This talk will provide a survey of available R/Hadoop interfaces and use an example use case to provide a comparison between systems. We’ll also discuss problems that are a good fit for distributed analysis with R, and those that aren’t.

    At 9:30am to 10:20am, Tuesday 20th September

    In Lindbergh, Hilton St. Louis at the Ballpark

    Coverage slide deck

Schedule incomplete?

Add a new session

Filter by Day

Filter by coverage

Filter by Topic

Filter by Venue

Filter by Space