Deep Dive into Apache Kafka - 2 out of 6
Thursday, October 6, 2016
10:00am PT | 1:00pm ET | 7:00pm CET
Recording Time: 59:50
In the last few years, Apache Kafka has been used extensively in enterprises for real-time data collecting, delivering, and processing. This talk will provide a deep dive on some of the key internals that help make Kafka popular.
- Companies like LinkedIn are now sending more than 1 trillion messages per day to Kafka. Learn about the underlying design in Kafka that leads to such high throughput.
- Many companies (e.g., financial institutions) are now storing mission critical data in Kafka. Learn how Kafka supports high availability and durability through its built-in replication mechanism.
- One common use case of Kafka is for propagating updatable database records. Learn how a unique feature called compaction in Apache Kafka is designed to solve this kind of problem more naturally.
This is talk 2 out of 6 from the Kafka Talk Series. This was recording from October 6, 2016.
プレゼンター
Jun Rao
Co-founder of Confluent and Co-creator of Kafka
Jun Rao は、Apache Kafka をコアとしたストリームデータプラットフォームを提供する Confluent の共同創設者です。Confluentの前は、Jun Rao は LinkedIn のシニアスタッフエンジニアを務め、Kafka の開発を主導していました。LinkedIn の前は、IBM のアルマデン研究データセンターの研究者であり、データベースと分散システムの研究を行っていました。Jun Rao は Apache Kafka Committee (PMC) の議長を務めており、Apache Cassandra のコミッターです。