Modern businesses have data at their core, and this data is changing continuously. How can we harness this torrent of continuously changing data in real-time? The answer is stream processing, and Apache Kafka® is a core hub for streaming data.
This talk will provide a brief introduction to Apache Kafka and describe its usage as a platform for streaming data. It will explain how Kafka serves as a foundation for both streaming data pipelines and applications that consume and process real-time data streams. It will introduce some of the newer components of Kafka that help make this possible, including Kafka Connect, a framework for capturing continuous data streams, and Kafka Streams, a lightweight stream processing library.
Mitch Henderson, Sr. Technical Account Manager, Confluent
Mitch Henderson is a Sr. Technical Account Manager for Confluent. Mitch’s background is rooted in DevOps roles mostly focused on inserting new technologies into the enterprise. Before joining Confluent Mitch was a field engineer at DataStax and played **integral roles** at Thompson Reuters and CenturyLink where he built large-scale deployments, planned auditing services, and was a key player doing financial data delivery with a focus on quality, maintainability, stability and scalability.
Ce site Web utilise des cookies afin d'améliorer l'expérience utilisateur et analyser les performances et le trafic sur notre site Web. Nous partageons également des informations concernant votre utilisation de notre site avec nos partenaires publicitaires, analytiques et de réseaux sociaux.