Rise of the Kafka Heroes ! Participez à la révolution du streaming de données | Lisez la bande dessinée
This blog post is the third in a four-part series that discusses a few new Confluent Control Center features that are introduced with Confluent Platform 6.2.0. It focuses on inspecting Apache Kafka® topics better with a newly added last-produced timestamp column in the Control Center UI. The series highlights the following new features that make managing clusters via Control Center an even smoother experience:
If you are not too familiar with Control Center, you can always refer to the Control Center overview first. Having a running Control Center instance at hand helps you explore the features discussed in this blog series better.
Now that you are ready, let’s delve into the third feature here in part 3: improved topic inspection via last-produced timestamp.
The “Topics” overview page gives a summary—health, throughput—of all the topics for a cluster. Confluent Platform 6.2.0 introduces a new column, “Last produced,” which reports the timestamp of the latest message produced to each topic. Using this report, you can easily compare and identify the topics that have not been produced in a long time.
There are a few important caveats to note:
--
means that no timestamp was obtained within a certain timeoutThe following section provides the technical rationale behind these specific limitations.
Currently, KafkaConsumer
does not provide a direct way of accessing the last message, or its timestamp, produced to a topic. Therefore, Control Center uses an internal KafkaConsumer
to actually consume messages and find the last produced timestamp for each topic.
To obtain a last-produced timestamp, the internal Control Center consumer performs the following:
KafkaConsumer::endOffsets()
.KafkaConsumer::seek()
.KafkaConsumer::poll()
. It’s important to note that poll(Duration timeout)
either returns immediately if there are any messages available from the given partitions, or it waits until timeout.The internal Control Center consumer performs the above procedure to obtain the last-produced timestamp per topic, giving each topic—active or dormant—an equal amount of time to poll and return a valid timestamp (dormant means that no messages or rarely any messages are produced to a topic). More importantly, the internal Control Center consumer performs the above procedure asynchronously for each topic, such that if one dormant topic’s poll
is awaiting timeout, it would not hinder other topics’ poll to return valid timestamps.
There are specific reasons behind the three caveats described earlier:
Because KafkaConsumer::poll()
could timeout in the case of dormant topics, obtaining the last-produced timestamp is a relatively expensive API call. Therefore, we only obtain new timestamps when necessary, for example, a page refresh.
The Control Center UI does not currently support sorting on a column, where its values are obtained separately per topic. Because we obtain the last-produced timestamp asynchronously per topic, we do not currently support sorting the topics by this column.
--
means that no timestamp was obtained within a certain timeout.
Currently, the timeout of KafkaConsumer::poll()
is set to five seconds in the internal Control Center consumer. This means that if no messages are returned by poll
within five seconds, we assume that fetching the last-produced timestamp failed for this topic. This could happen to any dormant topic. We do not try to obtain the last-produced timestamp for this topic until the next time that the page is refreshed.
In summary, improved topic inspection via last-produced timestamp is a beta feature that allows you to better distinguish dormant versus active topics by seeing when a message is last produced. Enhancements are on the way for future releases to eliminate all the current limitations.
To learn about other new features of Control Center 6.2.0, check out the remaining blog posts in this series:
Today, we’re introducing Confluent Cloud’s fully managed service for Apache Flink®, improvements to Kora Engine, how AI and streaming work together, and much more.
Learn the best practices for integrating Confluent with AWS Lambda to build event-driven architectures.