Anchor: Navigation

Retail Issue 4.0 - February 2025 Follow the Stream Your Dedicated Community Content Scroll 

The Year of DSP: Data Streaming in 2025 & Shift Left

Welcome to the fourth issue of Follow the Stream! To kick off 2025, this edition is dedicated to the concept of "Shift Left”—a data processing and governance paradigm that moves data processing closer to the source and ingests fresh trustworthy data into any downstream consumers. Covering data streaming predictions and trends for the upcoming year, we explore how Shift Left architecture is revolutionizing data quality and providing reusable data products for both operational and analytical use cases in financial services.
See Industry Updates

Data Streaming Fundamentals

Data Streaming Crash Course

Let's explore the basics of Shift Left & Data Streaming Platforms.

Heading 1

Subtitle 1

Heading 2

Subtitle 2

Heading 3

Subtitle 3
Shift Left is a rethink of how to circulate, share, and manage data. Along with a data streaming platform, it enables organizations to build data once, build it right, and reuse it anywhere within moments of its creation.
Data governance plays an important role ensuring data quality, security, and compliance across the data lifecycle, while data products, built around real-time streams, facilitate reusable, high-quality data assets for efficiency and collaboration.
Subtitle 1
Subtitle 2
Subtitle 3
Anchor: General-Updates

Industry in Motion: General Updates

Available in German, French, and Spanish!

3 Data Engineering Trends Riding Apache Kafka®, Apache Flink®, and Apache Iceberg™

The Apache Kafka, Flink, and Iceberg communities are continuously evolving, offering innovative ways for engineers to manage and process data at scale. From re-envisioning microservices with Flink streaming applications to enabling real-time AI model applications, these tools are reshaping data integration. With strong community contributions, especially to Iceberg, data governance and real-time analytics are set to accelerate, revolutionizing how businesses manage their data infrastructure.
Read Article

The Data Streaming Landscape 2025

“Data streaming is a new software category. It has grown from niche adoption to becoming a fundamental part of modern data architecture, leveraging open source technologies like Apache Kafka® and Apache Flink®. With real-time data processing transforming industries, the ecosystem of tools, platforms, and cloud services has evolved significantly.” Kai Waehner, Confluent’s Field CTO, “explores the data streaming landscape of 2025, analyzing key players, trends, and market dynamics shaping this space.”
Read Blog

Ushering in a New Era of Data Streaming Confluent Recognized as a Challenger in 2024 Gartner® Magic Quadrant™ for Data Integration Tools

Recognized as a Challenger in Gartner’s 2024 Magic Quadrant for Data Integration Tools, Confluent predicts that the future of data integration will focus on universal data products and event-driven flows, with real-time streaming bridging gaps between operational and analytical data. This shift will drive innovation in areas like fraud detection, personalized experiences, and supply chain optimization, making data streaming the standard for faster, more efficient decision-making.
Read Blog

New in Confluent Cloud: Extending Private Networking Across the Data Streaming Platform

Confluent’s Q4 2024 launch wrapped up the year with a host of powerful features paving the way for even more innovation in 2025. Focusing on delivering private networking and enhanced security across the data streaming platform, key features include mTLS authentication, AWS PrivateLink for Schema Registry, Apache Flink® UDFs for custom stream processing, and WarpStream Orbit for seamless Apache Kafka® migrations.
Read Blog
Anchor: Executives-Brief

Executive's Brief: Data Strategy & Leadership

Predictive Analytics: How GenAI and Data Streaming Work Together to Forecast the Future

Predictive analytics, enhanced by generative AI and data streaming, is transforming decision-making in retail. By utilizing historical data and machine learning models, predictive analytics allows retailers to forecast customer behavior, demand trends, and sales outcomes with greater precision. Generative AI takes this a step further by generating diverse scenarios, filling data gaps, and continuously adapting to real-time data for more accurate, dynamic forecasting.
Read Blog

Shift Left: Unifying Operations and Analytics with Data Products

The need for high-quality business data is greater than ever, so preventing and mitigating bad data—across the entire business—has become a critical capability. Shifting data processing and governance “left” eliminates duplicate pipelines, reduces the risk and impact of bad data at the source, and leverages high-quality data products for both operational and analytical use cases.
Read Ebook

Good Teams Manage Kafka – Efficient Teams Use Confluent

Managing open source Apache Kafka® in-house or through basic cloud-hosted services can lead to significant costs, operational challenges, and downtime risks, especially under tight engineering budgets. Using a fully managed, cloud-native platform can help reduce these challenges, enabling teams to focus on strategic initiatives while improving cost efficiency and accelerating time to market for real-time data projects.
Download White Paper
Available in German, French, Spanish, and Portuguese!

Hungry for Data – Huel Decouples Infrastructure with Data Streaming 

Huel, a rapidly expanding food subscription service, transformed its infrastructure by adopting an event-driven microservices powered by Apache Kafka®. This shift allowed Huel to decouple its systems and leverage real-time data, enhancing its subscription platform and inventory management. As a result, Huel optimized operational efficiency, enabling smarter decision-making and personalized customer experiences. 
Read Article
Anchor: Customer-Experience

Data Streaming for Customer Experience

How Data Streaming Elevates the Omnichannel Customer Experience

To realize its vision of a frictionless omnichannel experience, Thrivent invested in an agile data architecture that supports real-time data ingestion and processing. Alongside digital services consulting firm Improving, Thrivent discussed the challenges of building a next-gen data streaming architecture and best practices for developing reusable APIs to ensure a seamless customer experience across channels.
Watch Webinar

Data Streaming’s Role in Shaping and Scaling the Customer Experience for Fnac Darty

Fnac Darty uses data streaming to deliver real-time updates on pricing, inventory, and customer navigation events, ensuring a seamless and responsive shopping experience. By integrating siloed data and enabling instant access to actionable insights, the company can better understand customer intentions and optimize operations to meet their needs. This approach enhances agility, improving customer satisfaction and loyalty.
Read Article

Vimeo Revolutionizes Real-Time Experiences for Over 260 Million Users with Confluent

Vimeo uses data streaming to deliver real-time insights that improve user experiences, optimize video playback, and enable agile decision-making. By utilizing adaptive bitrate streaming, Vimeo dynamically adjusts video quality based on real-time network conditions, ensuring smooth playback without buffering. This real-time approach empowers the video platform to iterate faster, monitor campaigns, and make data-driven product decisions.
Read Case Study
Anchor: Architects-Blueprint

Architect's Blueprint: Data Systems Design

The Streaming Data Quality Guidebook

Maintaining high data quality is critical, especially with the rise of AI/ML, where poor data can lead to system failures, inaccurate decisions, and costly outages. By implementing best practices for stream governance—identifying schema issues, establishing data contracts, and adopting decentralized data ownership—organizations ensure the integrity of their data streams and significantly reduce risks.
Download Guidebook
Available in German, French, Spanish, and Japanese!

From Batch and Lakehouse to Real-Time Data Products with Data Streaming

“The Shift Left Architecture enables a data mesh with real-time data products to unify transactional and analytical workloads with Apache Kafka, Flink, and Iceberg. Consistent information is handled with streaming processing or ingested into Snowflake, Databricks, Google BigQuery, or any other analytics/AI platform to increase flexibility, reduce cost, and enable a data-driven company culture with faster time-to-market building innovative software applications.”
Read Blog

Real-Time Data Streaming for Smart Warehouses

Data streaming empowers retailers to revolutionize warehouse operations by enabling real-time inventory tracking, predictive maintenance, and optimized resource allocation. Real-time data enhances automation, reduces errors, and ensures timely fulfillment to meet customer demands, such as same-day delivery. By breaking down data silos and enabling actionable insights, data streaming supports agile and customer-centric logistics in the retail sector.
Read Blog
Anchor: must-read

Data Streaming Journey at Migros | Data Products, Observability, Event-Driven Microservices, IoT

Migros, one of the largest retailers and employers in Switzerland, started its data streaming journey in 2016 for tracking customer events across touchpoints. In an effort to continue optimizing its global supply chain, the supermarket chain leveraged data streaming for “customer master data management, product and store information updates, and synchronization with Migros Online.” By 2022, Migros kicked off the Enterprise Integration Platform, enabling data products across the organization.
Read the full story and many more in the new book by Kai Waehner: The Ultimate Data Streaming Guide. Download is available for free.
Download Ebook for Free
Available in German, French, and Spanish!

How the Retailer Intersport uses Apache Kafka® as Database with Compacted Topic

Intersport uses Apache Kafka's compacted topics to efficiently manage and query real-time data across global operations. By leveraging this feature, the retailer ensures accurate, up-to-date stock information throughout its supply chain, enabling seamless integration between systems like ERP, POS, and inventory management. This approach helps streamline operations, improve data consistency, and scale flexibly during high-demand periods, such as holidays.
Read Blog
Anchor: Developers-Desk

Developer's Desk: Building Applications

How Apache Iceberg and Flink Can Ease Developer Pain

Although developing stateful applications can be chaotic due to complex upstream and downstream interactions, tools like Apache Iceberg and Apache Flink help simplify this process. Iceberg optimizes data querying by defining efficient table structures, while Flink enables real-time data processing, improving speed and reliability. Together, they provide a robust framework that enhances both the efficiency and reliability of stateful application development.
Watch Podcast

Shift Processing and Governance Left: A DSP Product Demo

“Shift left" is gaining traction as a trend by addressing the increasing need for high-quality business data. This approach solves common challenges like redundant datasets, poor data quality, and high costs associated with data warehouses and data lakes by cleaning and aggregating data closer to the point of data generation.
Watch Demo

Handling the Producer Request: Kafka Producer and Consumer Internals | Part 2

The Apache Kafka® producer request process involves a series of stages, including handling by network threads, queuing, data validation by I/O threads, and eventually writing to disk. Throughout this journey, configurations such as buffer sizes, thread counts, and flush intervals play key roles in optimizing performance. Monitoring metrics help track the efficiency of each step, ensuring that the data is properly processed and replicated for durability.
Read Blog

Key Concepts of a Schema Registry

Schema Registry is a tool that manages and enforces data schemas in the Apache Kafka ecosystem, ensuring data compatibility and quality across applications. Understanding key concepts like schemas as contracts, the role of Schema Registry, schema workflow, and integration with client applications can help developers enhance system reliability and overall data streaming capabilities.
Watch Free Course

How to Add a Connector to Confluent Cloud

Kafka Connectors provide a way to get data flowing between sources and sinks, and Confluent Cloud. Choosing between available options like fully managed connectors for easy setup, or custom connectors that suit specific requirements, ultimately depends on the required level of control and specific data pipeline goals.
Watch on YouTube

Learn About Data Streaming With Apache Kafka® and Apache Flink®

High-throughput low latency distributed event streaming platform. Available locally or fully-managed via Apache Kafka on Confluent Cloud.

High-performance stream processing at any scale. Available via Confluent Cloud for Apache Flink.

Explore Developer Hub

Request Flink Workshop or Tech Talk

Anchor: Innovation-Research

Data Streaming for Innovation & Research

A New Era in Dynamic Pricing: Real-Time Data Streaming with Kafka and Flink

Dynamic pricing is “a flexible model that adjusts prices based on real-time market changes to enable real-time responsiveness to demand, competitor prices, and customer behaviors.” Leveraging data streaming allows companies like AO to “to seize pricing opportunities and align closely with customer expectations,” increasing customer conversion rates by 30%.
Read Blog

Real-Time Data Streaming is Key to a Successful Christmas Campaign

Data streaming is the “invisible but essential technology” for retailers, especially during critical times like the Christmas season. By analyzing customer behavior in real time, businesses can launch timely flash sales, suggest complementary products, and blend omnichannel discounts, creating a seamless shopping experience that drives conversions and boosts customer loyalty throughout the holiday rush.
Read Article

Join the community

Sign up for updates below and check out previous issues!

Share content suggestions and new uses cases in the Comments Section