CCA-500 : Cloudera Certified Administrator for Apache Hadoop (CCAH)

Exam CCA-500 | Cloudera Certification

ExamCollection Verified – Instant Download. Download Exam CCA-500 PDF files free.

Here you can download free practice tests for such certifications as MCSAMCSE, MCDBA, MCSD, A+,Network+, Security+, CCNA, CCNP, IBM, HP, VMware  and so on. All tests on this site have been converted with VCE Exam Simulator.

All latest VCE files will be convertedFiles are converted in 24 to 36 hours. VCE to PDF converter FREE

CCA-500 PREMIUM VCE file + PDF File LATEST : $20.00 

BuyNow

(CCA-500.vce & CCA-500.pdf : 60 Questions & Answers)

VCE files + PDF Files : $20.00

New Question Updates for : 15 days FREE

ExamCollection

  1. Latest questions, authenticity guaranteed;
  2. Refund policy applies;
  3. Great savings – Corporate Clients welcome;
  4. Immediate download after purchase;
  5. 95% of VCEplus clients pass their exams;
  6. Guaranteed by VCEplus & ExamCollection;
  7. Email Support : [email protected];
  8. Facebook fan page to support

Free Exam Questions in PDF Format

CCA-500 Cloudera Certified Administrator for Apache Hadoop (CCAH) 2015-07-14.pdf

 

Summary

A Cloudera Certified Administrator for Apache Hadoop (CCAH) certification proves that you have demonstrated your technical knowledge, skills, and ability to configure, deploy, maintain, and secure an Apache Hadoop cluster.

Cloudera Certified Administrator for Apache Hadoop (CCA-500)
Number of Questions: 60 questions
Time Limit: 90 minutes
Passing Score: 70%
Language: English, Japanese

Notes: Hadoop ecosystem items are no longer treated separately as their own section and are integrated throughout the exam. Both CCA–500 and CCA–505 share the same proportion of items per section.

1. HDFS (17%)

  • Describe the function of HDFS daemons
  • Describe the normal operation of an Apache Hadoop cluster, both in data storage and in data processing
  • Identify current features of computing systems that motivate a system like Apache Hadoop
  • Classify major goals of HDFS Design
  • Given a scenario, identify appropriate use case for HDFS Federation
  • Identify components and daemon of an HDFS HA-Quorum cluster
  • Analyze the role of HDFS security (Kerberos)
  • Determine the best data serialization choice for a given scenario
  • Describe file read and write paths
  • Identify the commands to manipulate files in the Hadoop File System Shell

2. YARN and MapReduce version 2 (MRv2) (17%)

  • Understand how upgrading a cluster from Hadoop 1 to Hadoop 2 affects cluster settings
  • Understand how to deploy MapReduce v2 (MRv2 / YARN), including all YARN daemons
  • Understand basic design strategy for MapReduce v2 (MRv2)
  • Determine how YARN handles resource allocations
  • Identify the workflow of MapReduce job running on YARN
  • Determine which files you must change and how in order to migrate a cluster from MapReduce version 1 (MRv1) to MapReduce version 2 (MRv2) running on YARN

3. Hadoop Cluster Planning (16%)

  • Principal points to consider in choosing the hardware and operating systems to host an Apache Hadoop cluster
  • Analyze the choices in selecting an OS
  • Understand kernel tuning and disk swapping
  • Given a scenario and workload pattern, identify a hardware configuration appropriate to the scenario
  • Given a scenario, determine the ecosystem components your cluster needs to run in order to fulfill the SLA
  • Cluster sizing: given a scenario and frequency of execution, identify the specifics for the workload, including CPU, memory, storage, disk I/O
  • Disk Sizing and Configuration, including JBOD versus RAID, SANs, virtualization, and disk sizing requirements in a cluster
  • Network Topologies: understand network usage in Hadoop (for both HDFS and MapReduce) and propose or identify key network design components for a given scenario

4. Hadoop Cluster Installation and Administration (25%)

  • Given a scenario, identify how the cluster will handle disk and machine failures
  • Analyze a logging configuration and logging configuration file format
  • Understand the basics of Hadoop metrics and cluster health monitoring
  • Identify the function and purpose of available tools for cluster monitoring
  • Be able to install all the ecoystme components in CDH 5, including (but not limited to): Impala, Flume, Oozie, Hue, Cloudera Manager, Sqoop, Hive, and Pig
  • Identify the function and purpose of available tools for managing the Apache Hadoop file system

5. Resource Management (10%)

  • Understand the overall design goals of each of Hadoop schedulers
  • Given a scenario, determine how the FIFO Scheduler allocates cluster resources
  • Given a scenario, determine how the Fair Scheduler allocates cluster resources under YARN
  • Given a scenario, determine how the Capacity Scheduler allocates cluster resources

6. Monitoring and Logging (15%)

  • Understand the functions and features of Hadoop’s metric collection abilities
  • Analyze the NameNode and JobTracker Web UIs
  • Understand how to monitor cluster daemons
  • Identify and monitor CPU usage on master nodes
  • Describe how to monitor swap and memory allocation on all nodes
  • Identify how to view and manage Hadoop’s log files
  • Interpret a log file

You may also like...