Apache Kafka is marketed as a messaging system, but most teams run it as an ETL backbone and central data hub. That is why Kafka clusters store tens of terabytes, sit between every database and engine, and show up in platform cost and reliability discussions instead of messaging design reviews. This article explains how Kafka actually gets used in modern stacks, what problems that creates for integration, execution and governance, and concrete steps leaders can take to simplify architectures, control Kafka related costs and prepare for a federated execution layer on top. Kafka solved integration, not messaging On paper Kafka sits next to message brokers. In practice it earned adoption because it fixed an integration problem. As organisations added more systems that needed to exchange data, point to point ETL pipelines multiplied beyond control. With ten source systems and ten destinations you quickly end up with something close to one hundred individual jobs. Each pipeline ca...
Fractional Chief Architect for Big Data Systems & Distributed Data Processing