Skip to content

pcarbo/R-survival-large-scale

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

45 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

A survival guide to large-scale data analysis in R

Materials for tutorial, "A survival guide to large-scale data analysis in R."

Overview

Large-scale data analysis in R is like the "super G" events in the Winter Olympics---it is about pushing the physical limits of your computer (or compute cluster). My first aim is to show you some techniques to push your R computing further. My second aim is to help you make more effective use of the most precious commodity in computing---memory---and to demonstrate how R sometimes makes poor use of it. This presentation is intended to be hands-on---bring your laptop, and we will work through the examples together. This git repository contains the source code for running the demos.

Other information

Credits

These materials were developed by Peter Carbonetto at the University of Chicago. Thank you to Matthew Stephens for his support and guidance. Also thanks to Gao Wang for sharing the Python script to profile memory usage, to David Gerard for sharing his code that ultimately improved several of the examples, and to John Blischak, John Novembre and Stefano Allesia for providing great examples to learn from.

About

Materials for tutorial, "A survival guide to large-scale data analysis in R."

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published