r/apachekafka

Hi. I'm curious how the adoption of these open table formats is going - who's using it in prod, who in dev, who's just curious. If anyone is willing to share where their Kafka cluster stands on the matter, I think it'll make for an interesting discussion!

We have certainly seen a plethora of vendors support OTF integration. One thing that's unclear to me is how useful those are, and which tradeoffs matter (e.g low ingestion lag, low management overhead, etc)

reddit.com
u/2minutestreaming — 14 days ago
▲ 0 r/apachekafka+2 crossposts

Hey folks,

I tried a different way to explain Kafka by creating a short video where a Kafka-like object explains the system itself.

Synopsis (relevance to Kafka):

The video walks through the core Kafka concepts:

  • Producers sending data into the system
  • Topics and partitions for scalability
  • Replication across brokers for fault tolerance
  • Consumers and consumer groups for parallel processing
  • Offsets for tracking consumption progress

The goal was to simplify how these pieces fit together in a real-time streaming pipeline, especially for people new to Kafka.

Would really appreciate feedback from people using Kafka:

  • Is this abstraction accurate enough?
  • What important concepts am I missing (retention, lag, rebalancing, etc.)?
  • Would this help when onboarding new engineers?
u/mock_coder — 10 days ago

Just shipped Jikkou 1.0: the Resource-as-Code framework for Apache Kafka 🎉.

Substantial new capabilities anchor the release:

🧊 Apache Iceberg provider : manage namespaces, tables, and views declaratively. Schema evolution runs in two passes so renames + type promotions land as a single safe change instead of drop-and-add. Works with REST, Hive, JDBC, AWS Glue, and Nessie catalogs.

🌍 Multi-cluster orchestration : group your providers and roll changes across a whole fleet in one command: jikkou apply --provider-group production --continue-on-error. Fleet-wide diffs in one shot too.

🔐 Confluent Cloud RBAC : RoleBinding resources to manage Confluent Cloud role bindings as code, next to your topics and ACLs.

📦 Plus: resource dependency ordering, JSON Schema export, provider-grouped get commands, hardened deps

Release note : https://www.jikkou.io/docs/releases/release-v1.0.0/

Happy to answer questions, hear what you'd want from the 1.x line, or just discuss how you're managing Kafka resources today 🙂.

medium.com
u/fhussonnois — 9 days ago

TL;DR: During blue-green broker deployments with 50K+ partition movements, Cruise Control moves all replicas over X hours, then executes all 10K+ leadership changes in a concentrated burst at the end, causing client latency spikes. Looking for ways to spread leadership movements throughout the rebalance.

Background scenario:

We run a 9-broker Kafka cluster and do blue-green deployments where we add 9 new brokers and rebalance the entire cluster. Our typical rebalance involves ~55,000 partition movements.

The execution is sequential:

  1. Move ALL replicas (X hours)
  2. Then move ALL leadership (concentrated burst)

This causes a "leadership storm" at the end where thousands of leadership changes happen rapidly, leading to client connection disruptions and request timeouts.

Questions:

  1. Is this sequential execution (replicas → leadership) fundamental to CC's architecture, or are we missing a config option?
  2. Has anyone else dealt with this during large rebalances or blue-green deployments?
reddit.com
u/TacticalObserver — 8 days ago

How do you handle SOC 2 / PCI-DSS evidence collection for Kafka?

Genuinely curious how teams here approach this.

For context, I've been spending a lot of time on the audit side of Kafka — SOC 2, PCI-DSS, ISO 27001 — and the recurring pain seems to be:

  1. Inventory: nobody's quite sure how many topics, clusters, or principals exist

  2. ACL audit: someone granted User:* during an incident a year ago and nobody undid it

  3. Inter-broker TLS: enabled on the dev cluster, mysteriously not on prod

  4. Audit logs: enabled, but no retention policy, so the auditor's "who consumed from this topic last quarter" question can't be answered

Some questions I'd love to hear answers to from this community:

- Do you run a pre-audit checklist? If yes, manual or automated?

- How do you prove inter-broker is encrypted, in writing, to an auditor?

- What's your strategy for ACL drift? (Periodic review? Diff against IaC?)

- Has anyone tied control evidence to CI/CD — i.e., the build won't merge if compliance breaks?

I work on an open-source project in this space (KafkaGuard) but I'm asking because the *questions* keep coming up identically across teams and I'd like to know what's working in the wild — tools, scripts, processes, anything.

Will share aggregated patterns from the replies if there's enough discussion.

reddit.com
u/jacksparrowon — 6 days ago

RFC: What do we want to do about AI-generated content on r/apachekafka?

There's been a sharp rise in the quantity of AI-generated content being shared on this sub. This includes blogs, videos, and tools. My concern is that we are frogs in a pan, and the water temperature is approaching boiling without us realising. Low-quality posts dilute visibility of useful content; people stop bothering to even downvote; the sub slowly dies. (Related: AI Slop is Killing Online Communities.)

For some context, visits to this sub are down over the last four months straight, and nearly 50% since October's peak. Perhaps this is unrelated. Perhaps not.

People sharing content built with AI are often not ill-intentioned. They are really excited about the thing they just created. But often it's not actually that novel or useful for the Apache Kafka community, and more of a "hey look what happened when I prompted my AI tool!". Which is cool, but doesn't belong on r/apachekafka.

I'm a member of other subs who have similar challenges, and see various approaches — all with their advantages and disadvantages.

What are the options?

  1. Ban anything AI-created.
  2. Mandate self-disclosure and labelling — poster must include Built with AI flair.
  3. No drive-by link dumping. Contributors must be already active in the sub, and engage with comments on their posts. A first-time post in r/apachekafka may not be sharing one's own content.
  4. Do nothing. Let downvotes do their thing.
  5. Other?

My opinion

Doing nothing is not an option. Downvotes are but a paper cocktail umbrella against a deluge; ineffective at scale. The community will disengage and over time disintegrate.

An outright ban is not the positive engagement environment that we seek to foster on r/apachekafka. StackOverflow tried the absolutist route, and I along with many others simply walked away because it sucks.

My proposal is that we adopt rules 2 and 3 above.

Your opinion?

This post is literally a Request for Comments 😄

Reply with your thoughts and votes for the above options (or other suggestions). I'll summarise next week and review with the rest of the mod team before any further action is taken.

u/rmoff — 3 days ago

Apache Kafka® Deserves Topic Types

Reposting my colleague Juha's blog all about the evolving language we use to describe Kafka, it's an interesting read :) He digs into how to way we describe Kafka has changed and the issues introduced by new innovations.

aiven.io
u/HughEvansDev — 2 days ago

Apache Kafka Community Events at Current London

Calling all Apache Kafka users (devs, architects, operators, etc) who are attending Current London

Besides all the amazing talks lined up, we wanted to share 2 Apache focused sessions that will provide you an opportunity to engage with AK committers, PMC members, adn the community at large.

  • The Apache Kafka AMA (Tuesday | 12:30 PM | Expo Hall - Meetup Hub)
    • Come ask tough questions to all the PMC members
  • Office Hours : The Apache Kafka Guildhall (Tuesday | 3:00 PM | Expo Hall - Sponsor Theater)
    • Come and share your Kafka stories with other practitioners and PMC members and learn along the way. This is an open session, not a presentation, located at the Sponsor Theater.
u/LoudCat5895 — 1 day ago
▲ 10 r/apachekafka+2 crossposts

If you're consuming from Kafka and writing into ClickHouse, sync inserts at high message rates will hurt you. Async insert mode helps a lot but the buffering and dedupe behavior isn't always obvious.

Wrote this up from our my experience building a stream processing pipeline.

Curious how others are handling the Kafka → ClickHouse write path.

u/Marksfik — 17 hours ago