Systems | Development | Analytics | API | Testing

March 2024

How to Unlock the Power of Event-Driven Architecture | Designing Event-Driven Microservices

An Event-Driven Architecture is more than just a set of microservices. Event Streams should represent the central nervous system, providing the bulk of communication between all components in the platform. Unfortunately, many projects stall long before they reach this point.

Set your Data in Motion with Confluent on Google Cloud

Confluent is pioneering a fundamentally new category of data infrastructure focused on data in motion. Confluent’s cloud-native offering is the foundational platform for data in motion – designed to be the intelligent connective tissue enabling real-time data, from multiple sources, to constantly stream across the organization. With Confluent, organizations can meet the new business imperative of delivering rich, digital front-end customer experiences and transitioning to sophisticated, real-time, software-driven backend operations.

Confluent announces general availability of Confluent Cloud for Apache Flink®, simplifying stream processing to power next-gen apps

Confluent Cloud for Apache Flink®, a leading cloud-native, serverless Flink service is now available on AWS, Google Cloud, and Microsoft Azure. Confluent's fully managed, cloud-native service for Flink helps customers build high-quality data streams for data pipelines, real-time applications, and analytics.

Confluent Cloud for Apache Flink Is Now Generally Available

Last year, we announced our plan to build a cloud-native Apache Flink® service to meet the growing demand for scalable and efficient stream processing solutions in the cloud. Today, we're thrilled to announce the general availability of Confluent Cloud for Apache Flink across all three major clouds. This means that you can now experience Apache Kafka® and Flink as a unified, enterprise-grade platform to connect and process your data in real time, wherever you need it.

Introducing Tableflow

We’re excited to talk about our vision for Tableflow, which makes it push-button simple to take Apache Kafka® data and feed it directly into your data lake, warehouse, or analytics engine as Apache Iceberg® tables. Making operational data accessible to the analytical world is traditionally a complex, expensive, and brittle process and we believe we can do better to unify the operational and analytical estates.

Streams Forever: Kafka Summit London 2024 Keynote | Jay Kreps, Co-founder & CEO, Confluent

Join the Confluent leadership team as they share their vision of streaming data products enabled by a data streaming platform built around Apache Kafka. Jay Kreps, Co-creator of Apache Kafka and CEO of Confluent, will present his vision of unifying the operational and analytical worlds with data streams and showcase exciting new product capabilities. During this keynote, the winner and finalists of the $1M Data Streaming Startup Challenge will showcase how their use of data streaming is disrupting their categories.

Build, Connect, and Consume Intelligent Data Pipelines Seamlessly and Securely

We’re excited to share the latest and greatest features on Confluent Cloud, in our first launch of 2024. This Cloud Launch comes to you from Kafka Summit London, where we talked about the latest updates highlighted in our launch, including serverless Apache Flink®, some exciting pricing changes, updates to connectors, and more! We also shared our vision for a future offering, Tableflow.

Exploring Apache Flink 1.19: Features, Improvements, and More

The Apache Flink® community unveiled Apache Flink version 1.19 this week! This release is packed with numerous new features and enhancements. In this blog post, we'll spotlight some of the standout additions. For a comprehensive rundown of all updates, don't forget to review the release notes.

Confluent Cloud for Apache Flink | Simple, Serverless Stream Processing

Stream processing plays a critical role in the infrastructure stack for data streaming. Developers can use it to filter, join, aggregate, and transform their data streams on the fly to power real-time applications and streaming data pipelines. Among stream processing frameworks, Apache Flink has emerged as the de facto standard because of its performance and rich feature set. However, self-managing Flink (like self-managing other open source tools like Kafka) can be challenging due to its operational complexity, steep learning curve, and high costs for in-house support.

The Confluent Q1 '24 Launch

The Confluent Q1 ’24 Launch is packed with new features that enable customers to build, connect, and consume intelligent data pipelines seamlessly and securely Our quarterly launches provide a single resource to learn about the accelerating number of new features we’re bringing to Confluent Cloud, our cloud-native data streaming platform.

4 Key Types of Event-Driven Architecture

Adam Bellemare compares four main types of Event-Driven Architecture (EDA): Application Internal, Ephemeral Messaging, Queues, and Publish/Subscribe. Event-Driven Architectures have a long and storied history, and for good reason. They offer a powerful way to build scalable and decoupled architectures. But thanks to its long history, people often have different ideas of what EDA means depending on when they first encountered this architecture.

How to Evolve your Microservice Schemas | Designing Event-Driven Microservices

Schema evolution is the act of modifying the structure of the data in our application, without impacting clients. This can be a challenging problem. However, it gets easier if we start with a flexible data format and take steps to avoid unnecessary data coupling. When we find ourselves having to make breaking changes, we can always fall back to creating new versions of our APIs and events to accommodate those changes.

What is a Kafka Consumer and How does it work?

Now that your data is inside your Kafka cluster, how do you get it out? In this video, Dan Weston covers the basics of Kafka Consumers: what consumers are, how they get your data flowing, and best practices for configuring consumers in a real-time data streaming system. You will also learn about offsets, consumer groups, and partition assignment.