Log Compaction

Log Compaction | Highlights in the Kafka and Stream Processing Community | November 2015

Gwen Shapira
Last Updated: 

kafka-logo-wide (1)

The Apache Kafka community just concluded its busiest month ever. As we are preparing for the upcoming release of Kafka 0.9.0.0, the community worked together to close a record number of bugs and added quite a few last-minute improvements to new features. Here’s a summary of some of the community highlights:

  • SASL/Kerberos-based authentication was added to Kafka and will be included in the 0.9.0.0 release
  • The new Kafka Consumer has a client-side partition-assignment support, allowing consumer groups to implement their own partition assignment strategy, in addition to those provided in Apache Kafka. You can read the very detailed design document or take a look at the patch.
  • Huge improvements on the Copycat export/import feature – Copycat now works in distributed mode and there is a nice REST API too.
  • The community also worked hard on making Kafka more testable. In the last few month the old spagetti-like system tests slowly got replaced by the more modular Ducktape-based integration tests. Last week we removed the old tests completely and moving forward we will only maintain the new test suite.
  • This is the month of the Kafka Meetup! New Kafka meetups are popping up in cities all over the world including Seattle, Austin and London. I encourage you to attend if you happen to be in one of those cities.
  • If you are a Kafka internals geek, you’ll enjoy reading the deep details behind Kafka’s delayed response mechanism.
  • For those interested in the latest in stream processing, Data Artisans shared interesting highlights from the first Apache Flink conference.
  • Remember that the Kafka Summit “Call for Proposals” is open so make sure you get your entry in before the January 11, 2016 deadline. If you’re planning to attend and have ideas of sessions you’d like to see on the program tweet us @ConfluentInc #kafkasummit to let us know.

If you are interested in contributing to Apache Kafka, check out our contributor guide to help you get started.


Apache Kafka

Download the Confluent Platform 

Subscribe to the Confluent Blog

Abonnieren

More Articles Like This

Streams and Tables in Apache Kafka: Elasticity, Fault Tolerance, and Other Advanced Concepts
Michael Noll

Streams and Tables in Apache Kafka: Elasticity, Fault Tolerance, and Other Advanced Concepts

Michael Noll

Now that we’ve learned about the processing layer of Apache Kafka® by looking at streams and tables, as well as the architecture of distributed processing with the Kafka Streams API […]

Streams and Tables in Apache Kafka: Processing Fundamentals with Kafka Streams and ksqlDB
Michael Noll

Streams and Tables in Apache Kafka: Processing Fundamentals with Kafka Streams and ksqlDB

Michael Noll

Part 2 of this series discussed in detail the storage layer of Apache Kafka: topics, partitions, and brokers, along with storage formats and event partitioning. Now that we have this […]

Streams and Tables in Apache Kafka: A Primer
Michael Noll

Streams and Tables in Apache Kafka: A Primer

Michael Noll

This four-part series explores the core fundamentals of Kafka’s storage and processing layers and how they interrelate. In this first part, we begin with an overview of events, streams, tables, […]

Fully managed Apache Kafka as a Service!

Try Free