Project Metamorphosis: Unveiling the next-gen event streaming platform.Learn More

Fundamentals for Apache Kafka®

Register Now

What is Apache Kafka® and how does it work?

Apache Kafka was built with the vision to become the central nervous system that makes real-time data available to all the applications that need to use it, with numerous use cases like stock trading and fraud detection, to transportation, data integration, and real-time analytics.

This four-part online talk series provides an overview of what Kafka is, what it's used for, and the core concepts that enable it to power a highly scalable, available and resilient real-time event streaming platform. The series begins with an introduction to the shift toward real-time data streaming, and continues all the way through to best practices for developing applications with Apache Kafka and how to integrate Kafka into your environment.

Whether you’re just getting started or have already built stream processing applications, you will find actionable insights in this series that will enable you to further derive business value from your data systems.

Register now to learn Apache Kafka from Confluent, the company founded by Kafka’s original developers.

Benefits of Stream Processing and Apache Kafka Use Cases

Tuesday, May 26
3 sessions to accommodate your preferred time zone:

  • 10:00am PDT / 1:00pm ET
  • 11:00am SGT / 1:00pm AEST
  • 10:00am London

This talk explains how companies are using event-driven architecture to transform their business and how Apache Kafka serves as the foundation for streaming data applications.

Learn how major players in the market are using Kafka in a wide range of use cases such as microservices, IoT and edge computing, core banking and fraud detection, cyber data collection and dissemination, ESB replacement, data pipelining, ecommerce, mainframe offloading and more.

Also discussed in this talk are the differences between Apache Kafka and Confluent Platform.

Apache Kafka Architecture & Fundamentals Explained

Tuesday, June 2
3 sessions to accommodate your preferred time zone:

  • 10:00am PDT / 1:00pm ET
  • 11:00am SGT / 1:00pm AEST
  • 10:00am London

This session explains Apache Kafka’s internal design and architecture. Companies like LinkedIn are now sending more than 1 trillion messages per day to Apache Kafka. Learn about the underlying design in Kafka that leads to such high throughput.

This talk provides a comprehensive overview of Kafka architecture and internal functions, including:

  • Topics, partitions and segments
  • The commit log and streams
  • Brokers and broker replication
  • Producer basics
  • Consumers, consumer groups and offsets

How Apache Kafka Works

Tuesday, June 9
3 sessions to accommodate your preferred time zone:

  • 10:00am PDT / 1:00pm ET
  • 11:00am SGT / 1:00pm AEST
  • 10:00am London

Pick up best practices for developing applications that use Apache Kafka, beginning with a high level code overview for a basic producer and consumer. From there we’ll cover strategies for building powerful stream processing applications, including high availability through replication, data retention policies, producer design and producer guarantees.

We’ll delve into the details of delivery guarantees, including exactly-once semantics, partition strategies and consumer group rebalances. The talk will finish with a discussion of compacted topics, troubleshooting strategies and a security overview.

Integrating Apache Kafka Into Your Environment

Tuesday, June 16
3 sessions to accommodate your preferred time zone:

  • 10:00am PDT / 1:00pm ET
  • 11:00am SGT / 1:00pm AEST
  • 10:00am London

Integrating Apache Kafka with other systems in a reliable and scalable way is a key part of an event streaming platform. This session will show you how to get streams of data into and out of Kafka with Kafka Connect and REST Proxy, maintain data formats and ensure compatibility with Schema Registry and Avro, and build real-time stream processing applications with Confluent KSQL and Kafka Streams.

Sign Up Now

Recevez jusqu'à 50 $ US de réduction sur votre facture chaque mois calendaire pour le premier trimestre.

Nouvelles inscriptions uniquement.

By clicking “sign up” above you understand we will process your personal information in accordance with our Politique de confidentialité.

En cliquant sur « Inscription » ci-dessus, vous acceptez les termes du/de la Conditions d'utilisation et de recevoir occasionnellement des e-mails publicitaires de la part de Confluent. Vous comprenez également que nous traiterons vos informations personnelles conformément à notre Politique de confidentialité.

Gratuit à vie sur un seul broker Kafka
i

Le logiciel permettra une utilisation illimitée dans le temps de fonctionnalités commerciales sur un seul broker Kafka. Après l'ajout d'un second broker, un compteur de 30 jours démarrera automatiquement sur les fonctionnalités commerciales. Celui-ci ne pourra pas être réinitialisé en revenant à un seul broker.

Sélectionnez un type de déploiement
Manual Deployment
  • tar
  • zip
  • deb
  • rpm
  • docker
ou
Déploiement automatique
  • kubernetes
  • ansible

By clicking "download free" above you understand we will process your personal information in accordance with our Politique de confidentialité.

En cliquant sur « Téléchargement gratuit » ci-dessus, vous acceptez la Contrat de licence Confluent et de recevoir occasionnellement des e-mails publicitaires de la part de Confluent. Vous acceptez également que vos renseignements personnels soient traitées conformément à notre Politique de confidentialité.

Ce site Web utilise des cookies afin d'améliorer l'expérience utilisateur et analyser les performances et le trafic sur notre site Web. Nous partageons également des informations concernant votre utilisation de notre site avec nos partenaires publicitaires, analytiques et de réseaux sociaux.