Data Engineering using Kafka and Spark Structured Streaming

Udemy Data Engineering using Kafka and Spark Structured Streaming

Register & Get access to index
lMSd8Vf.jpg


A comprehensive Data Engineering course on building streaming pipelines using Kafka and Spark Structured Streaming

What you'll learn​

  • Setting up self support lab with Hadoop (HDFS and YARN), Hive, Spark, and Kafka
  • Overview of Kafka to build streaming pipelines
  • Data Ingestion to Kafka topics using Kafka Connect using File Source
  • Data Ingestion to HDFS using Kafka Connect using HDFS 3 Connector Plugin
  • Overview of Spark Structured Streaming to process data as part of Streaming Pipelines
  • Incremental Data Processing using Spark Structured Streaming using File Source and File Target
  • Integration of Kafka and Spark Structured Streaming - Reading Data from Kafka Topics

Requirements​

  • Laptop with decent configuration
  • Decent internet speed to watch the lessons
  • Self Support lab (instructions will be provided as part of the course) or ITVersity labs
  • Knowledge about Functional Programming (preferably Python or Scala)
  • Knowledge or experience using Spark

Description​

As part of this course, you will be learning to build streaming pipelines by integrating Kafka and Spark Structured Streaming. Let us go through the details about what is covered in the course.
  • First of all, we need to have the proper environment to build streaming pipelines using Kafka and Spark Structured Streaming on top of Hadoop or any other distributed file system. As part of the course, you will start with setting up a self-support lab with all the key components such as Hadoop, Hive, Spark, and Kafka on a single node Linux-based system.
  • Once the environment is set up you will go through the details related to getting started with Kafka. As part of that process, you will create a Kafka topic, produce messages into the topic as well as consume messages from the topic.
  • You will also learn how to use Kafka Connect to ingest data from web server logs into Kafka topic as well as ingest data from Kafka topic into HDFS as a sink.
  • Once you understand Kafka from the perspective of Data Ingestion, you will get an overview of some of the key concepts of related Spark Structured Streaming.
  • After learning Kafka and Spark Structured streaming separately, you will build a streaming pipeline to consume data from Kafka topic using Spark Structured Streaming, then process and write to different targets.
  • You will also learn how to take care of incremental data processing using Spark Structured Streaming.
Course Outline
Here is a brief outline of the course. You can choose either Cloud9 or GCP to provision a server to set up the environment.
  • Setting up Environment using AWS Cloud9 or GCP
  • Setup Single Node Hadoop Cluster
  • Setup Hive and Spark on top of Single Node Hadoop Cluster
  • Setup Single Node Kafka Cluster on top of Single Node Hadoop Cluster
  • Getting Started with Kafka
  • Data Ingestion using Kafka Connect - Web server log files as a source to Kafka Topic
  • Data Ingestion using Kafka Connect - Kafka Topic to HDFS a sink
  • Overview of Spark Structured Streaming
  • Kafka and Spark Structured Streaming Integration
  • Incremental Loads using Spark Structured Streaming
Udemy based support
In case you run into technical challenges while taking the course, feel free to raise your concerns using Udemy Messenger. We will make sure that issue is resolved in 48 hours.

Who this course is for:​

  • Experienced ETL Developers who want to learn Kafka and Spark to build streaming pipelines
  • Experienced PL/SQL Developers who want to learn Kafka and Spark to build streaming pipelines
  • Beginner or Experienced Data Engineers who want to learn Kafka and Spark to build streaming pipelines
Author
TUTProfessor
Downloads
98
Views
1,240
First release
Last update
Rating
0.00 star(s) 0 ratings

More resources from TUTProfessor