The Introduction to Hadoop for Developers training course is designed to demonstrate the fundamentals of setting up a Hadoop cluster, as well as the "soup" of related technologies like Hive, Pig and Oozie.
The course begins with an examination of how to access the Hadoop file system and write MapReduce jobs using Java, Pig, and Hive Oozie. Next, the course discusses examples of real world Map Reduce jobs and how Hadoop has solved real world data-intensive processing problems. The course concludes by exploring the different modes in which Hadoop can be run to support massive amounts of data, as well as students' MapReduce jobs during development.
Prerequisites: Basic Java knowledge is expected, and experience with Eclipse is a plus.
Purpose
|
Learn how to write MapReduce programs using Java. |
Audience
|
System adminstrators, developers, and DevOps engineers creating Big Data solutions using Hadoop. |
Role
| Software Developer - System Administrator |
Skill Level
| Intermediate |
Style
| Hack-a-thon - Learning Spikes - Workshops |
Duration
| 4 Days |
Related Technologies
| Java | Hadoop | Apache |
Productivity Objectives
- Discover the Hadoop Distributed File System (HDFS)
- Interpret general Hadoop Cluster/HDFS administration
- Explain MapReduce
- Define how to write a MapReduce job with Java, Pig, and Hive
- Differentiate how the different Hadoop technologies inter-operate to provide a cohesive big data solution
- Demonstrate basic management of a Hadoop cluster
- Give examples of how to perform basic unit testing of MapReduce jobs
- Distinguish how Message Passing Interface (MPI) and High Performance Computing (HPC) intersect with Hadoop