Systems | Development | Analytics | API | Testing

Confluent

Exploring Apache Flink 1.20: Features, Improvements, and More

The Apache Flink community released Apache Flink 1.20 this week. In this blog post, we'll highlight some of the most interesting additions and improvements. You’ll find a comprehensive rundown of all of the updates in the official release announcement. Recent Flink releases have emphasized improving Flink’s usability, not only for developers, but also for those operating Flink clusters, and this theme continues in this latest release.

How BT Group Built a Smart Event Mesh with Confluent

BT Group's Smart Event Mesh: Centralized Event Streaming With Decentralized Customer Experience, Automation, and a Foundation Built on Confluent. BT Group is a British multinational telecommunications holding company headquartered in London, England. It has operations in around 180 countries and is one of the largest telecommunications companies in the world, providing a range of products and services including fixed-line, broadband, mobile, and TV.

How to Set Idle Timeouts | Apache Flink in Action

This video covers setting an idle timeout on a watermark generator when joining data in Apache Flink. This can be used when you have two streams, one that has frequent updates, and one that has infrequent updates, and you need to join data without waiting for a fresh watermark from the infrequent one.

New with Confluent Platform: Enhanced security with OAuth Support, Confluent Platform for Apache Flink (LA), a new HTTP Source Connector, and More

At Confluent we’re committed to building the world's leading data streaming platform that's cloud-native, complete, and available everywhere your data and applications reside. We offer this data streaming platform as a fully managed service in the cloud—Confluent Cloud; as a self-managed software that runs in your own environments—Confluent Platform; or as a hybrid of each of these.

Flink AI: Real-Time ML and GenAI Enrichment of Streaming Data with Flink SQL on Confluent Cloud

Modern data platforms enable enterprises to extract valuable business insights from data, sourced from various origins. Data engineers, data scientists, and other data practitioners utilize both data streaming and batch processing frameworks as a means to provide these insights. While batch processes work on historical data, stream processing extracts insights in real time, enabling businesses to react faster with respect to changing events.

Introducing Apache Kafka 3.8

We are proud to announce the release of Apache Kafka 3.8.0. This release contains many new features and improvements. This blog post will highlight some of the more prominent features. For a full list of changes, be sure to check the release notes. See the Upgrading to 3.8.0 from any version 0.8.x through 3.7.x section in the documentation for the list of notable changes and detailed upgrade steps. In a previous release, 3.6, Tiered Storage was released as early access feature.

Accelerate your data streaming journey with the latest in Confluent Cloud

The Q2 2024 Confluent Cloud launch introduces a suite of enhancements across the four key pillars of a Data Streaming Platform - Stream, Connect, Process, and Govern – alongside some significant work we have been doing with our partner ecosystem to help customers unlock new possibilities. Confluent has helped over 4,900+ global enterprises start their data streaming journey and was recently named a Leader by Forrester Research in The Forrester Wave: Streaming Data Platforms, Q4 2023.

Confluent Cloud for Apache Flink | Interactive Tables for Flink SQL Workspaces

When developing or debugging a stream processing pipeline with Flink SQL, it’s common to inspect each processing step's output to ensure data is being transformed properly. However, comprehending the resulting data stream's structure, distribution, and characteristics entails executing multiple ad-hoc SQL queries, which can be time-consuming and tedious. Additionally, isolating specific subsets of the stream for analysis or debugging often involves even more queries, adding to the complexity and time required.