對象:Decision makers, IT managers, technical managers, software developers, data analysts who want to learn latest new technologies in handling massive and ever-growing data assets.
內容:
Installing Hadoop and HDFS operations
-Install from CDH
-Format the NameNode
-Start and stop Hadoop daemons
-List the Hadoop daemons
-Test the Hadoop installation
Running a MapReduce Job in pseudo-distributed mode
-Upload sample data to HDFS
-Run a MapReduce job
-View the output from the MapReduce job
-View the output from the Hadoop console
Running a MapReduce Job in a clustered mode
-Configure the IP addresses and host names
-Configure the core site properties
-Configure the HDFS properties
-Configure the MapReduce properties
-Creating required Hadoop directories
-Setting the ownership for Hadoop directories
-Start the daemons for master node
-Start the daemons for slave node
-Test the Hadoop installation
Managing Jobs
-List running jobs on the cluster
-Kill running jobs
Breaking the Cluster
-Causing failures on Data Node
-Verifying the cluster
Self-Healing Features of a Cluster
-Causing failures on Name Node
-Verifying the cluster