Cloudera Developer Training for Spark and Hadoop

 

This four-day hands-on training course delivers the key concepts and expertise developers need to develop high-performance parallel applications with Apache Spark 2. Participants will learn how to use Spark SQL to query structured data and Spark Streaming to perform real-time processing on streaming data from a variety of sources. Developers will also practice writing applications that use core Spark to perform ETL processing and iterative algorithms. The course covers how to work with large datasets stored in a distributed file system, and execute Spark applications on a Hadoop cluster. After taking this course, participants will be prepared to face real-world challenges and build applications to execute faster decisions, better decisions, and interactive analysis, applied to a wide variety of use cases, architectures, and industries.

With this course update, we streamlined the agenda to help you quickly become productive with the most important technologies, including Spark 2.

Overview

Learn how to import data into your Apache Hadoop cluster and process it with Spark, Hive, Flume, Sqoop, Impala, and other Hadoop ecosystem tools This four-day hands-on training course delivers the key concepts and expertise participants need to ingest and process data on a Hadoop cluster using the most up-to-date tools and techniques. Employing Hadoop ecosystem projects such as Spark, Hive, Flume, Sqoop, and Impala, this training course is the best preparation for the real-world challenges faced by Hadoop developers. Participants learn to identify which tool is the right one to use in a given situation, and will gain hands-on experience in developing using those tools.

Get Hands-On Experience

Hands-on exercises take place on a live cluster, running in the cloud. A private cluster will be built for each student to use during the class.

Through instructor-led discussion and interactive, hands-on exercises, participants will navigate the Hadoop ecosystem, learning how to:

  • Distribute, store, and process data in a Hadoop cluster
  • Write, configure, and deploy Spark applications on a cluster
  • Use the Spark shell for interactive data analysis
  • Process and query structured data using Spark SQL
  • Use Spark Streaming to process a live data stream

What to Expect

This course is designed for developers and engineers who have programming experience, but prior knowledge of Hadoop and/or Spark is not required.

  • Apache Spark examples and hands-on exercises are presented in Scala and Python. The ability to program in one of those languages is required.
  • Basic familiarity with the Linux command line is assumed
  • Basic knowledge of SQL is helpful

Get Certified

Upon completion of the course, attendees are encouraged to continue their study and register for the CCA Spark and Hadoop Developer exam. Certification is a great dierentiator. It helps establish you as a leader in the field, providing employers and customers with tangible evidence of your skills and expertise.

Course Fee : 2,180 €
Duration : 4 days

Quick Info

  • Laptop and more tools available during the courses
  • Affordable course fee
  • Dedicated Course Trainer