Cloudera Administration Study Guide

Download as txt, pdf, or txt
Download as txt, pdf, or txt
You are on page 1of 3
At a glance
Powered by AI
The document outlines the sections and objectives covered in the Cloudera Certified Associate exams CCA-410, CCA-500, and CCA-505.

The major sections covered in the CCA-410 exam are HDFS, MapReduce, Hadoop Cluster Planning, Hadoop Cluster Installation and Administration, Resource Management, and Monitoring and Logging.

Important considerations for hardware configuration in Hadoop cluster planning include choosing appropriate operating systems, hardware sizing based on workload patterns, disk configuration and sizing, and network topology design.

CCA410 Exam Sections and Blueprint

1. HDFS (38%)
Describe the function of all Hadoop Daemons
Describe the normal operation of an Apache Hadoop cluster, both in data storage
and in data processing.
Identify current features of computing systems that motivate a system like Apach
e Hadoop.
Classify major goals of HDFS Design
Given a scenario, identify appropriate use case for HDFS Federation
Identify components and daemon of an HDFS HA-Quorum cluster
Analyze the role of HDFS security (Kerberos)
Determine the best data serialization choice for a given scenario
Describe file read and write paths
Identify the commands to manipulate files in the Hadoop File System Shell
2. MapReduce (10%)
Understand how to deploy MapReduce MapReduce v1 (MRv1)
Understand how to deploy MapReduce v2 (MRv2 / YARN)
Understand basic design strategy for MapReduce v2 (MRv2)
3. Hadoop Cluster Planning (12%)
Principal points to consider in choosing the hardware and operating systems to h
ost an Apache Hadoop cluster.
Analyze the choices in selecting an OS
Understand kernel tuning and disk swapping
Given a scenario and workload pattern, identify a hardware configuration appropr
iate to the scenario
Cluster sizing: given a scenario and frequency of execution, identify the specif
ics for the workload, including CPU, memory, storage, disk I/O
Disk Sizing and Configuration, including JBOD versus RAID, SANs, virtualization,
and disk sizing requirements in a cluster
Network Topologies: understand network usage in Hadoop (for both HDFS and MapRed
uce) and propose or identify key network design components for a given scenario
4. Hadoop Cluster Installation and Administration (17%)
Given a scenario, identify how the cluster will handle disk and machine failures
.
Analyze a logging configuration and logging configuration file format.
Understand the basics of Hadoop metrics and cluster health monitoring.
Identify the function and purpose of available tools for cluster monitoring.
Identify the function and purpose of available tools for managing the Apache Had
oop file system.
5. Resource Management (6%)
Understand the overall design goals of each of Hadoop schedulers.
Given a scenario, determine how the FIFO Scheduler allocates cluster resources.
Given a scenario, determine how the Fair Scheduler allocates cluster resources.
Given a scenario, determine how the Capacity Scheduler allocates cluster resourc
es.
6. Monitoring and Logging (12%)
Understand the functions and features of Hadoops metric collection abilities
Analyze the NameNode and JobTracker Web UIs
Interpret a log4j configuration
Understand how to monitor the Hadoop Daemons
Identify and monitor CPU usage on master nodes
Describe how to monitor swap and memory allocation on all nodes
Identify how to view and manage Hadoops log files
Interpret a log file
7. The Hadoop Ecosystem (5%)
Understand Ecosystem projects and what you need to do to deploy them on a cluste
r.
CCA500 and 505 Exam Sections and Blueprint
Notes: Hadoop ecosystem items are no longer treated separately as their own sect
ion and are integrated throughout the exam. Both CCA500 and CCA505 share the same
proportion of items per section.
1. HDFS (17%)
Describe the function of HDFS Daemons
Describe the normal operation of an Apache Hadoop cluster, both in data storage
and in data processing.
Identify current features of computing systems that motivate a system like Apach
e Hadoop.
Classify major goals of HDFS Design
Given a scenario, identify appropriate use case for HDFS Federation
Identify components and daemon of an HDFS HA-Quorum cluster
Analyze the role of HDFS security (Kerberos)
Determine the best data serialization choice for a given scenario
Describe file read and write paths
Identify the commands to manipulate files in the Hadoop File System Shell
2. YARN and MapReduce version 2 (MRv2) (17%)
Understand how upgrading a cluster from Hadoop 1 to Hadoop 2 affects cluster set
tings
Understand how to deploy MapReduce v2 (MRv2 / YARN), including all YARN daemons
Understand basic design strategy for MapReduce v2 (MRv2)
Determine how YARN handles resource allocations
Identify the workflow of MapReduce job running on YARN
Determine which files you must change and how in order to migrate a cluster from
MapReduce version 1 (MRv1) to MapReduce version 2 (MRv2) running on YARN.
3. Hadoop Cluster Planning (16%)
Principal points to consider in choosing the hardware and operating systems to h
ost an Apache Hadoop cluster.
Analyze the choices in selecting an OS
Understand kernel tuning and disk swapping
Given a scenario and workload pattern, identify a hardware configuration appropr
iate to the scenario
Given a scenario, determine the ecosystem components your cluster needs to run i
n order to fulfill the SLA
Cluster sizing: given a scenario and frequency of execution, identify the specif
ics for the workload, including CPU, memory, storage, disk I/O
Disk Sizing and Configuration, including JBOD versus RAID, SANs, virtualization,
and disk sizing requirements in a cluster
Network Topologies: understand network usage in Hadoop (for both HDFS and MapRed
uce) and propose or identify key network design components for a given scenario
4. Hadoop Cluster Installation and Administration (25%)
Given a scenario, identify how the cluster will handle disk and machine failures
Analyze a logging configuration and logging configuration file format
Understand the basics of Hadoop metrics and cluster health monitoring
Identify the function and purpose of available tools for cluster monitoring
Be able to install all the ecoystme components in CDH 5, including (but not limi
ted to): Impala, Flume, Oozie, Hue, Cloudera Manager, Sqoop, Hive, and Pig
Identify the function and purpose of available tools for managing the Apache Had
oop file system
5. Resource Management (10%)
Understand the overall design goals of each of Hadoop schedulers
Given a scenario, determine how the FIFO Scheduler allocates cluster resources
Given a scenario, determine how the Fair Scheduler allocates cluster resources u
nder YARN
Given a scenario, determine how the Capacity Scheduler allocates cluster resourc
es
6. Monitoring and Logging (15%)
Understand the functions and features of Hadoops metric collection abilities
Analyze the NameNode and JobTracker Web UIs
Understand how to monitor cluster Daemons
Identify and monitor CPU usage on master nodes
Describe how to monitor swap and memory allocation on all nodes
Identify how to view and manage Hadoops log files
Interpret a log file
Disclaimer: These exam preparation pages are intended to provide information abo
ut the objectives covered by each exam, related resources, and recommended readi
ng and courses. The material contained within these pages is not intended to gua
rantee a passing score on any exam. Cloudera recommends that a candidate thoroug
hly understand the objectives for each exam and utilize the resources and traini
ng courses recommended on these pages to gain a thorough understand of the domai
n of knowledge related to the role the exam evaluates.

You might also like