Build your first MapReduce with Hadoop and Ruby

Build your first MapReduce with Hadoop and Ruby

Supplementary code I used in the "Build your first MapReduce with Hadoop and Ruby" at the BRUG March meet-up.

The Code for the live demo resides here: https://github.com/swanandp/MapReduce_with_hadoop

A9e271fb1622f8dbb6d652993f5a23a7?s=128

Swanand Pagnis

March 16, 2013
Tweet

Transcript

  1. Ohai Hadoop! Build your first MapReduce with Hadoop & Ruby

  2. Tweet@_swanand GitHub@swanandp StackOverflow@18678 Work@Kaverisoft Make { DispatchTrack } mailto:swanand@pagnis. in

    Who am I? Ruby, Coffeescript, Java, Rails, Sinatra, Android, TextMate, Emacs, Minitest, MySQL, Cassandra, Hadoop, Mountain Lion, Curl, Zsh, GMail, Solarized, Oscar Wilde, Robert Jordan, Quentin Tarantino, Charlize Theron
  3. • MapReduce! Wait, what? • Enter the Hadoop. *gong* •

    Convention over Configuration? You wish. • Instant Gratification. Now you're talkin' • Further Reading. Go forth and read! Tell 'em what you're going to tell 'em
  4. MapReduce! Wait, what? • Map: Given a set of values

    (or key-values), output another set of values (or key-values) • [K1, V1] -> map -> [K2, V2] • Map each value into a new value
  5. MapReduce! Wait, what? • Reduce: Given a set of values

    for a key, come up with a summarized version • K1[V1, V2 ... Vn] -> reduce -> K1[Y] • Reduce given values into 1 value
  6. MapReduce! Wait, what?

  7. MapReduce! Um.. hmm.. Q: What is the single biggest takeaway

    from mapping? A: Map operation is stateless i.e. one iteration doesn't depend on previous iteration. Q: What is the single biggest takeaway from reducing? A: Reduce represents an operation for a particular key.
  8. Enter the Hadoop. *gong* "The really interesting thing I want

    you to notice, here, is that as soon as you think of map and reduce as functions that everybody can use, and they use them, you only have to get one supergenius to write the hard code to run map and reduce on a global massively parallel array of computers, and all the old code that used to work fine when you just ran a loop still works only it's a zillion times faster which means it can be used to tackle huge problems in an instant." - Joel Spolsky
  9. MapReduce! Oh, yeah! 1. Convert raw data into readable format

    2. Iterate over data chunks, convert each chunk into meaningful key, value pairs 3. Do this for all your data using massive parallelization 4. Group all the keys and their respective values 5. Take values for a key and convert into desired meaningful format 6. Step 2 is called mapper 7. Step 5 is called reducer
  10. Enter the Hadoop. *gong* Same process has now become: 1.

    Put data into Hadoop 2. Define your mapper 3. Define your reducer 4. Run your jobs 5. Read processed data from Hadoop Other advantages: • Encapsulations over common problems like large files, process management, disk / node failure
  11. Top Level Descriptor job has_many tasks HDFS Boss core-site.xml HDFS

    Slaves slaves MapReduce Boss mapred-site.xml MapReduce Slave mapred-site.xml User's window into Hadoop, through the command hadoop Convention over Configuration? You wish. Job Task NameNode DataNode JobTracker TaskTracker Client
  12. Convention over Configuration? You wish. • Configuration in XML &

    Shell scripts. Yuck! • Respite: ◦ Option for specifying a configuration directory ◦ Shell script configuration is mostly ENV variables • Which means: ◦ Configuration can be written in YML or JSON or Ruby and exported in XML ◦ ENV variables can be set using rake, thor or just plain Ruby • Caveats: ◦ No standard wrapper to do this (Go write one!)
  13. Convention over Configuration? You wish. • Default mappers and reducers

    are defined in Java • Other languages supported using Streaming API • Streaming API makes use of STDIN and STDOUT to read and output data and executable binaries for processing • Caveats ◦ No dependency management, we are on our own
  14. Instant Gratification. Now you're talkin' GOAL: 1. Take a couple

    of books in txt format 2. Find out the total usage of each character in the english alphabet. 3. Establish that e is the most used. 4. Why this example? a. Perfect use case for MapReduce. b. Algorithm is simple. c. Results are simple to analyze. d. Txt formatted books are easily available in Project Gutenberg.
  15. • Official Documentation • Wiki: http://wiki.apache.org/hadoop/ • Hadoop examples that

    ship with Hadoop • http://www.bigfastblog.com/map-reduce- with-ruby-using-hadoop • http://www.youtube.com/watch? v=d2xeNpfzsYI Further Reading and Watching
  16. Questions?

  17. Thank you!