Systems | Development | Analytics | API | Testing

Confluent Supercharges Apache Flink Offering with New Developer Tools and Enterprise-Ready Security

Confluent adds Table API support for Apache Flink® making it even easier for developers to use Java or Python to build streaming applications. New private networking and encryption features safeguard data streams for enterprises in highly regulated industries.

The Confluent Q3 '24 Launch

The Confluent Q3 ’24 Launch is packed with the new features and tools needed to make stream processing a developer’s best friend. We’re also introducing Client-Side Field Level Encryption, WarpStream, and more! Our quarterly launches provide a single resource to learn about the accelerating number of new features we’re bringing to Confluent Cloud, our cloud-native data streaming platform.

New research from Confluent sees IT leaders share their biggest AI implementation challenges

12th September 2024 - Skills shortages are the #1 challenge facing IT leaders looking to implement artificial intelligence in 2024. That's according to research from Confluent, released ahead of this year's BigData LDN event. The research, which surveyed over 500 UK IT leaders, explores the top challenges facing IT departments when it comes to adopting and implementing AI.

Shift left to write data once, read as tables or streams

Shift Left is a rethink of how we circulate, share and manage data in our organizations using DataStreams, Change Data Capture, FlinkSQL and Tableflow. It addresses the challenges with multi-hop and medallion architectures using batch pipelines by shifting the data preparation, cleaning and schemas to the point where data is created and as a result, you can build fresh trustworthy datasets as streams for operational use cases or Apache Iceberg tables for analytical use cases.

How to source data from AWS DynamoDB to Confluent using DynamoDB Streams and AWS Lambda

This is a one-minute video showing an animated architectural diagram of the integration between Amazon DynamoDB and Confluent Cloud using DynamoDB Streams and AWS Lambda. Details of the integration are provided via narration.

How Producers Work: Kafka Producer and Consumer Internals, Part 1

I shouldn’t have to convince anyone that Apache Kafka is an incredibly useful and powerful technology. As a distributed event streaming platform, it’s adept at storing your event data and serving it up for downstream consuming applications to make sense of that information––in real time or as close to real time as your use case permits. The real beauty of Kafka as a technology is that it can do it with very little effort on your part. In effect, it’s a black box.

How to source data from AWS DynamoDB to Confluent using the DynamoDB CDC Source Connector

This is a one-minute video showing an animated architectural diagram of the integration between Amazon DynamoDB and Confluent Cloud using the all new, fully managed DynamoDB CDC Source connector. This real-time data pipeline doesn’t require you to write or maintain code.

Watermark Alignment Explained in 2 Minutes | Apache Flink in Action

Watermark alignment is a relatively new feature in Apache Flink. It lets you cope with the problem of needing to temporally join streams with mismatched event frequencies, e.g., one stream’s updates are much more frequent than those of the stream(s) with which you need to join it. In this video we’ll break the feature down, and relate how it can help you better manage your Apache Flink integration.