Building Kafka Solutions with Confluent Training Course
This instructor-led, live training (online or onsite) is aimed at engineers who wish to use Confluent (a distribution of Kafka) to build and manage a real-time data processing platform for their applications.
By the end of this training, participants will be able to:
- Install and configure Confluent Platform.
- Use Confluent's management tools and services to run Kafka more easily.
- Store and process incoming stream data.
- Optimize and manage Kafka clusters.
- Secure data streams.
Format of the Course
- Interactive lecture and discussion.
- Lots of exercises and practice.
- Hands-on implementation in a live-lab environment.
Course Customization Options
- This course is based on the open source version of Confluent: Confluent Open Source.
- To request a customized training for this course, please contact us to arrange.
Course Outline
Introduction
- Running Apache Kafka at scale with Confluent.
Setting up Confluent
Overview of Confluent Features and Architecture
Building a Streaming Platform
The Publish and Subscribe Process
How Kafka Stores Data
Processing Data On-the-Fly
Case Study: Twitter Analytics
Implementing Kafka's APIs
- Producer, Consumer, Streams, and Connect
Building Applications on Top of Kafka
Monitoring Kafka
Administration Tools
Case Study: Netflix Movie Recommendations
Adding New Systems
Detecting Problems with Message Delivery
Enterprise Security
Disaster Recovery
Developer Features
Troubleshooting
Summary and Conclusion
Requirements
- A general understanding of Apache Kafka
- Java programming experience
Audience
- Developers
- Architects
- System Administrators
Open Training Courses require 5+ participants.
Building Kafka Solutions with Confluent Training Course - Booking
Building Kafka Solutions with Confluent Training Course - Enquiry
Building Kafka Solutions with Confluent - Consultancy Enquiry
Consultancy Enquiry
Testimonials (5)
The labs and the slides combine well with Jorge's knowledge and love for Kafka.
Willem - BMW SA
Course - Apache Kafka for Developers
Sufficient hands on, trainer is knowledgable
Chris Tan
Course - A Practical Introduction to Stream Processing
During the exercises, James explained me every step whereever I was getting stuck in more detail. I was completely new to NIFI. He explained the actual purpose of NIFI, even the basics such as open source. He covered every concept of Nifi starting from Beginner Level to Developer Level.
Firdous Hashim Ali - MOD A BLOCK
Course - Apache NiFi for Administrators
Grate skills, examples, very good exercises
Marek Konieczny - G2A.COM Limited
Course - Kafka for Administrators
That I had it in the first place.
Peter Scales - CACI Ltd
Course - Apache NiFi for Developers
Upcoming Courses
Related Courses
Apache Kafka Connect
7 HoursThis instructor-led, live training in Norway (online or onsite) is aimed at developers who wish to integrate Apache Kafka with existing databases and applications for processing, analysis, etc.
By the end of this training, participants will be able to:
- Use Kafka Connect to ingest large amounts of data from a database into Kafka topics.
- Ingest log data generated by an application servers into Kafka topics.
- Make any collected data available for stream processing.
- Export data from Kafka topics into secondary systems for storage and analysis.
Unified Batch and Stream Processing with Apache Beam
14 HoursApache Beam is an open source, unified programming model for defining and executing parallel data processing pipelines. It's power lies in its ability to run both batch and streaming pipelines, with execution being carried out by one of Beam's supported distributed processing back-ends: Apache Apex, Apache Flink, Apache Spark, and Google Cloud Dataflow. Apache Beam is useful for ETL (Extract, Transform, and Load) tasks such as moving data between different storage media and data sources, transforming data into a more desirable format, and loading data onto a new system.
In this instructor-led, live training (onsite or remote), participants will learn how to implement the Apache Beam SDKs in a Java or Python application that defines a data processing pipeline for decomposing a big data set into smaller chunks for independent, parallel processing.
By the end of this training, participants will be able to:
- Install and configure Apache Beam.
- Use a single programming model to carry out both batch and stream processing from withing their Java or Python application.
- Execute pipelines across multiple environments.
Format of the Course
- Part lecture, part discussion, exercises and heavy hands-on practice
Note
- This course will be available Scala in the future. Please contact us to arrange.
Building Data Pipelines with Apache Kafka
7 HoursApache Kafka is a distributed streaming platform. It is de facto a standard for building data pipelines and it solves a lot of different use-cases around data processing: it can be used as a message queue, distributed log, stream processor, etc.
We'll start with some theory behind data pipelines in general, then continue with fundamental concepts behind Kafka. We'll also discover important components like Kafka Streams and Kafka Connect.
Apache Flink Fundamentals
28 HoursThis instructor-led, live training in Norway (online or onsite) introduces the principles and approaches behind distributed stream and batch data processing, and walks participants through the creation of a real-time, data streaming application in Apache Flink.
By the end of this training, participants will be able to:
- Set up an environment for developing data analysis applications.
- Understand how Apache Flink's graph-processing library (Gelly) works.
- Package, execute, and monitor Flink-based, fault-tolerant, data streaming applications.
- Manage diverse workloads.
- Perform advanced analytics.
- Set up a multi-node Flink cluster.
- Measure and optimize performance.
- Integrate Flink with different Big Data systems.
- Compare Flink capabilities with those of other big data processing frameworks.
A Practical Introduction to Stream Processing
21 HoursIn this instructor-led, live training in Norway (onsite or remote), participants will learn how to set up and integrate different Stream Processing frameworks with existing big data storage systems and related software applications and microservices.
By the end of this training, participants will be able to:
- Install and configure different Stream Processing frameworks, such as Spark Streaming and Kafka Streaming.
- Understand and select the most appropriate framework for the job.
- Process of data continuously, concurrently, and in a record-by-record fashion.
- Integrate Stream Processing solutions with existing databases, data warehouses, data lakes, etc.
- Integrate the most appropriate stream processing library with enterprise applications and microservices.
Distributed Messaging with Apache Kafka
14 HoursThis course is for enterprise architects, developers, system administrators and anyone who wants to understand and use a high-throughput distributed messaging system. If you have more specific requirements (e.g. only system administration side), this course can be tailored to better suit your needs.
Kafka for Administrators
21 HoursThis instructor-led, live training in Norway (online or onsite) is aimed at sysadmins who wish to set up, deploy, manage and optimize an enterprise-grade Kafka cluster.
By the end of this training, participants will be able to:
- Set up and administer a Kafka Cluster.
- Evaluate the benefits and disadvantages of deploying Kafka on-premise vs in the cloud.
- Deploy and monitor Kafka in using various on-premise and cloud environment tools.
Apache Kafka for Developers
21 HoursThis instructor-led, live training in Norway (online or onsite) is aimed at intermediate-level developers who wish to develop big data applications with Apache Kafka.
By the end of this training, participants will be able to:
- Develop Kafka producers and consumers to send and read data from Kafka.
- Integrate Kafka with external systems using Kafka Connect.
- Write streaming applications with Kafka Streams & ksqlDB.
- Integrate a Kafka client application with Confluent Cloud for cloud-based Kafka deployments.
- Gain practical experience through hands-on exercises and real-world use cases.
Apache Kafka for Python Programmers
7 HoursThis instructor-led, live training in Norway (online or onsite) is aimed at data engineers, data scientists, and programmers who wish to use Apache Kafka features in data streaming with Python.
By the end of this training, participants will be able to use Apache Kafka to monitor and manage conditions in continuous data streams using Python programming.
Stream Processing with Kafka Streams
7 HoursKafka Streams is a client-side library for building applications and microservices whose data is passed to and from a Kafka messaging system. Traditionally, Apache Kafka has relied on Apache Spark or Apache Storm to process data between message producers and consumers. By calling the Kafka Streams API from within an application, data can be processed directly within Kafka, bypassing the need for sending the data to a separate cluster for processing.
In this instructor-led, live training, participants will learn how to integrate Kafka Streams into a set of sample Java applications that pass data to and from Apache Kafka for stream processing.
By the end of this training, participants will be able to:
- Understand Kafka Streams features and advantages over other stream processing frameworks
- Process stream data directly within a Kafka cluster
- Write a Java or Scala application or microservice that integrates with Kafka and Kafka Streams
- Write concise code that transforms input Kafka topics into output Kafka topics
- Build, package and deploy the application
Audience
- Developers
Format of the course
- Part lecture, part discussion, exercises and heavy hands-on practice
Notes
- To request a customized training for this course, please contact us to arrange
Confluent KSQL
7 HoursThis instructor-led, live training in Norway (online or onsite) is aimed at developers who wish to implement Apache Kafka stream processing without writing code.
By the end of this training, participants will be able to:
- Install and configure Confluent KSQL.
- Set up a stream processing pipeline using only SQL commands (no Java or Python coding).
- Carry out data filtering, transformations, aggregations, joins, windowing, and sessionization entirely in SQL.
- Design and deploy interactive, continuous queries for streaming ETL and real-time analytics.
Apache NiFi for Administrators
21 HoursIn this instructor-led, live training in Norway (onsite or remote), participants will learn how to deploy and manage Apache NiFi in a live lab environment.
By the end of this training, participants will be able to:
- Install and configure Apachi NiFi.
- Source, transform and manage data from disparate, distributed data sources, including databases and big data lakes.
- Automate dataflows.
- Enable streaming analytics.
- Apply various approaches for data ingestion.
- Transform Big Data and into business insights.
Apache NiFi for Developers
7 HoursIn this instructor-led, live training in Norway, participants will learn the fundamentals of flow-based programming as they develop a number of demo extensions, components and processors using Apache NiFi.
By the end of this training, participants will be able to:
- Understand NiFi's architecture and dataflow concepts.
- Develop extensions using NiFi and third-party APIs.
- Custom develop their own Apache Nifi processor.
- Ingest and process real-time data from disparate and uncommon file formats and data sources.
Spark Streaming with Python and Kafka
7 HoursThis instructor-led, live training in Norway (online or onsite) is aimed at data engineers, data scientists, and programmers who wish to use Spark Streaming features in processing and analyzing real-time data.
By the end of this training, participants will be able to use Spark Streaming to process live data streams for use in databases, filesystems, and live dashboards.