Viktor Gamov is a Principal Developer Advocate at Confluent, founded by the original creators of Apache Kafka®. With a rich background in implementing and advocating for distributed systems and cloud-native architectures, Viktor excels in open-source technologies. He is passionate about assisting architects, developers, and operators in crafting systems that are not only low in latency and scalable but also highly available.
As a Java Champion and an esteemed speaker, Viktor is known for his insightful presentations at top industry events like JavaOne, Devoxx, Kafka Summit, and QCon. His expertise spans distributed systems, real-time data streaming, JVM, and DevOps. Viktor has co-authored “Enterprise Web Development” and “Apache Kafka® in Action”. Follow Viktor on X - @gamussa to stay updated with Viktor’s latest thoughts on technology, his gym and food adventures, and insights into open-source and developer advocacy.
Streaming data with Apache Kafka has become the backbone of modern applications. While streams are ideal for continuous data flow, they lack built-in querying capabilities. Unlike databases with indexed lookups, Kafka’s append-only logs are designed for high-throughput processing—not for on-demand queries. This necessitates additional infrastructure to query streaming data effectively.
Traditional approaches replicate stream data into external stores: relational databases like PostgreSQL for operational queries, object storage like S3 accessed via Flink, Spark, or Trino for analytics, and Elasticsearch for full-text search and log analytics. Each serves a purpose—but they also introduce silos, schema mismatches, freshness issues, and complex ETL pipelines that increase system fragility.
In this session, we’ll explore solutions that aim to unify operational, analytical, and search workloads across real-time data. We’ll demonstrate the following:
stream processing with Kafka Streams, Apache Flink, and SQL engines real-time analytics with Apache Pinot and ClickHouse search capabilities with Elasticsearch modern lakehouse approaches using Apache Iceberg with Tableflow to represent Kafka topics as queryable tables While there’s no one-size-fits-all solution, understanding the tools and trade-offs will help you design more robust and flexible architectures.