`

Talend Big Data v6 Certified Developer Exam

Talend certification exams measure candidates' skills to ensure that they have the knowledge to successfully implement quality projects. Preparation is critical to passing.

Certification Exam Details

Exam content is updated periodically. The number and difficulty of questions may change. The passing score is adjusted to maintain a consistent standardfor example, a new exam version with more difficult questions may have a lower passing score.

Approximate number of questions: 60
Time limit: 1 hour (60 minutes)
Types of questions:

  • Multiple choice
  • Multiple response
  • True or false

Recommended Experience

General knowledge of Hadoop: HDFS, Map Reduce v1 and v2, Hive, Pig, HBase, Hue, Zookeeper, and Sqoop.

Experience with Talend Big Data 6.x solutions and Talend Studio, including metadata creation, configuration, and troubleshooting.

Preparation

To prepare for the certification exam, Talend recommends:

  • Take the Big Data Basics training course
  • Study the training material
  • Read the product documentation
  • Acquire experience by using the product for at least 6 months

Certification Exam Topics

Big Data—general concepts

  • YARN daemons
  • How HDFS works
  • The Hadoop ecosystem: Pig, Hive, Hue, Hbase, and Sqoop
  • The process to create cluster metadata in Talend Studio
  • Ways to test connectivity to a Hadoop cluster

HDFS

  • What is HDFS?
  • Talend components dedicated to HDFS: names, how they work, how to configure them
  • Mandatory configuration to connect to HDFS
  • Troubleshooting common issues

Hive

  • What is Hive?
  • Talend components dedicated to Hive: names, how they work, how to configure them
  • How to create, profile, and preview Hive tables
  • Troubleshooting common issues

Pig

  • What is Pig?
  • Talend components dedicated to Pig: names, how they work, how to configure them
  • Troubleshooting common issue

HBase

  • What is HBase?
  • Talend components dedicated to HBase: names, how they work, how to configure them
  • Troubleshooting common issues

Sqoop

  • What is Sqoop?
  • Talend components dedicated to Sqoop: names, how they work, how to configure them
  • Troubleshooting common issues

Big Data batch Jobs

  • Configuration of Job execution: cluster information, memory limits
  • How to run a Big Data batch Job

Sample Questions

1. You designed a Big Data batch using the MapReduce framework. You plan to execute it on a cluster using Map Reduce v1. What mandatory configuration must be specified on the Hadoop Configuration tab of the Run view?

Choose all that apply.

a. Name Node

b. Data Node

c. Resource Manager

d. Job Tracker

2. What is HDFS?

a. A data warehouse infrastructure tool for processing structured data in Hadoop

b. A tool for importing/exporting tables from/to the Hadoop file system

c. A column-oriented key/value data store built to run atop the Hadoop file system

d. The primary storage system used by Hadoop applications

3. In which perspective of Studio can you run an analysis on Hive table content?

a. Profiling

b. Integration

c. Big Data

d. Mediation

4. HDFS components can only be used in Big Data batch or Big Data streaming Jobs.

a. True

b. False

5. ZooKeeper service is mandatory to coordinate transactions between Talend Studio and HBase.

a. True

b. False

6. Select the outgoing links allowed for a tSqoopImport component:

a. Main

b. Iterate

c. OnSubjobOk

d. SqoopCombine

7. What is the name of the Pig scripting language?

a. Pig Latin

b. Pig Scripting Language

c. Pig QL

d. Pig Code

Answers:

  1. a and d
  2. d
  3. a
  4. b
  5. a
  6. b and c
  7. a