Learn all the skills and techniques you'll need to deploy each key component of a Hadoop platform in 24 lessons of an hour or less. Jeffrey Aven shows how to put a platform together in your local environment or in the cloud, building a fully functional Hadoop cluster and using it with real programs and datasets.
The book has step-by-step instructions to walk you through common questions, issues, and tasks; Q-and-As, Quizzes, and Exercises build and test your knowledge; "Did You Know?" tips; and "Watch Out!" alerts help you avoid pitfalls.
Author: Jeffrey Aven
Date: April 2017
Audience: Big data developers
Category: Data Science
Understanding Hadoop and the Hadoop Distributed File System (HDFS)
Importing data into Hadoop, and process it there
Mastering basic MapReduce Java programming, and using advanced MapReduce API concepts
Making the most of Apache Pig and Apache Hive
Implementing and administering YARN
Taking advantage of the full Hadoop ecosystem
Managing Hadoop clusters with Apache Ambari
Working with the Hadoop User Environment (HUE)
Scaling, securing, and troubleshooting Hadoop environments
Integrating Hadoop into the enterprise
Deploying Hadoop in the cloud
Getting started with Apache Spark
Also see Reading Your Way Into Big Data, an article on Programmer's Bookshelf in which Ian Stirk provides a roadmap of the reading required to take you from novice to competent in areas relating to data science.
Follow @bookwatchiprog on Twitter or subscribe to our Books RSS feed for each day's new addition to Book Watch and for new reviews.
To have new titles included in Book Watch contact BookWatch@i-programmer.info