Project Metamorphosis: Unveiling the next-gen event streaming platform.Learn More
Log Compaction

Log Compaction | Highlights in the Kafka and Stream Processing Community | January 2016

Happy 2016! Wishing you a wonderful, highly scalable, and very reliable year. Log Compaction is a monthly digest of highlights in the Apache Kafka and stream processing community. Got a newsworthy item? Let us know.

Many things have happened since we last shared the state of Apache Kafka and the streams ecosystem. Lets take a look!

  • Kafka 0.9 was released and so was Confluent Platform 2.0
  • The call for proposals for Kafka Summit is closing soon – submit your abstract by Monday, January 11! Make it your new year’s resolution to participate more in the Kafka community and you can start by registering for the conference. Catch the Early Bird price (save $100) before it expires on January 15. 
  • Congratulations to Ewen Cheslack-Postava who joined Apache Kafka as a committer! We wish him much success and many patch reviews.
  • Kafka 0.9 added protocol support for managing groups so that clients no longer need to interact with ZooKeeper directly. KIP-40 gives a design for adding protocol support for clients to list available groups and to show the group members and their lag. The protocol is implemented in the new ConsumerGroupCommand (kafka-consumer-groups.sh), and is also available for use by 3rd party clients.
  • Kafka currently only supports a “logical” notion of time – the message offset, which indicates a relative order of messages. Many users want to be able to know the physical time that a message was produced. KIP-32, which is in active discussions and voting, will add a timestamp field to each Kafka message, indicating the time the client created the message. This would eventually allow adding special indexes and also support consuming messages based on their timestamp.
  • Kafka Connect is a new feature introduced in 0.9 that makes it really easy to directly integrate Kafka with external data systems like RDBMS’s or Hadoop. This tutorial shows how to use Kafka Connect to get events from a relational database to Kafka, and from there to HDFS and Hive, including automated partitioning of the data and updates to the Hive schema.
  • Looking to use Kafka with Spring for stream processing? There’s a 5-part blog post on how to do just that.
  • Yahoo published a benchmark comparing popular stream processing frameworks – Storm, SparkStreaming, and Flink.
  • At the first Seattle Kafka Meetup this past November, Microsoft shared how they use Kafka in Bing. One trillion messages per day!
  • Kafka, The Definitive Guide is now available for pre-order from O’Reilly.

Did you like this blog post? Share it now

Subscribe to the Confluent blog

More Articles Like This

Kafka Streams Interactive Queries Go Prime Time

What is stopping you from using Kafka Streams as your data layer for building applications? After all, it comes with fast, embedded RocksDB storage, takes care of redundancy for you, […]

From Eager to Smarter in Apache Kafka Consumer Rebalances

Everyone wants their infrastructure to be highly available, and ksqlDB is no different. But crucial properties like high availability don’t come without a thoughtful, rigorous design. We thought hard about […]

Walmart’s Real-Time Inventory System Powered by Apache Kafka

Consumer shopping patterns have changed drastically in the last few years. Shopping in a physical store is no longer the only way. Retail shopping experiences have evolved to include multiple […]

Jetzt registrieren

Erhalten Sie in den ersten drei Monaten einen Rabatt von bis zu 50 USD pro Kalendermonat auf Ihre Rechnung.

Nur neue Registrierungen.

By clicking “sign up” above you understand we will process your personal information in accordance with our und bin damit einverstanden.

Indem Sie oben auf „Registrieren“ klicken, akzeptieren Sie die Nutzungsbedingungen und den gelegentlichen Erhalt von Marketing-E-Mails von Confluent. Zudem ist Ihnen bekannt, dass wir Ihre personenbezogenen Daten gemäß unserer und bin damit einverstanden.

Auf einem einzigen Kafka Broker unbegrenzt kostenlos verfügbar
i

Die Software ermöglicht die unbegrenzte Nutzung der kommerziellen Funktionen auf einem einzelnen Kafka Broker. Nach dem Hinzufügen eines zweiten Brokers startet automatisch ein 30-tägiger Timer für die kommerziellen Funktionen, der auch durch ein erneutes Herunterstufen auf einen einzigen Broker nicht zurückgesetzt werden kann.

Wählen Sie den Implementierungstyp aus
Manual Deployment
  • tar
  • zip
  • deb
  • rpm
  • docker
oder
Automatische Implementierung
  • kubernetes
  • ansible

By clicking "download free" above you understand we will process your personal information in accordance with our Datenschutzerklärung zu.

Indem Sie oben auf „kostenlos herunterladen“ klicken, akzeptieren Sie die Confluent-Lizenzvertrag und den gelegentlichen Erhalt von Marketing-E-Mails von Confluent. Zudem erklären Sie sich damit einverstanden, dass wir Ihre personenbezogenen Daten gemäß unserer Datenschutzerklärung zu.

Diese Website verwendet Cookies zwecks Verbesserung der Benutzererfahrung sowie zur Analyse der Leistung und des Datenverkehrs auf unserer Website. Des Weiteren teilen wir Informationen über Ihre Nutzung unserer Website mit unseren Social-Media-, Werbe- und Analytics-Partnern.