Systems | Development | Analytics | API | Testing

Resume tokens and last-event IDs for LLM streaming: How they work & what they cost to build

When an AI response reaches token 150 and the connection drops, most implementations have one answer: start over. The user re-prompts, you pay for the same tokens twice, and the experience breaks. Resume tokens and last-event IDs are the mechanism that prevents this. They make streams addressable – every message gets an identifier, clients track their position, and reconnections pick up from exactly where they left off. The concept is straightforward.

Why ELT Can't Keep Up in the Era of High-Scale Data Engineering

While winning in artificial intelligence (AI) is critical to the future of business, old-school analytics—visualizations, dashboards, and infrequent reports—are still core to an organization's data needs. Behind the scenes, this analytics ecosystem remains heavily hydrated by batch-based ELT data integration. For a long time, this made perfect sense, as data sources were fewer, data volumes were manageable, and analytics consumers were limited.

Why Databox MCP Wins for AI Analytics Over Individual Connector MCPs

The Model Context Protocol (MCP) has given AI assistants something they’ve never had before: a standardized way to pull live data from external systems. Instead of just generating text, an AI agent can now query your CRM, check ad performance, or pull revenue numbers in real time. The industry’s response has been predictable. Every major platform is racing to build their own MCP server.

Analytics Beyond Reporting: How Embedded BI Drives Executive Action

Most executives are drowning in dashboards but starving for insights. We’ve been conditioned to view “analytics” as a rear-view mirror, a report on what happened, rather than a steering wheel for what should happen next. Traditional BI creates a “reporting tax,” where scaling insights requires a proportional increase in data analyst headcount to interpret the noise.

Leveraging the MCP Registry in Kong Konnect for Dynamic Tool Discovery

As enterprises start deploying AI agents into real systems, a new architectural challenge is emerging. Agents need a reliable way to discover tools, services, and capabilities dynamically, instead of relying on hardcoded integrations. This is where the Model Context Protocol (MCP) ecosystem is rapidly evolving. MCP servers expose tools and capabilities that AI agents can use. However, once organizations begin deploying multiple MCP servers across environments, the question becomes clear.

Configuring Kong Dedicated Cloud Gateways with Managed Redis in a Multi-Cloud Environment

A persistent challenge arises as businesses adopt multicloud architectures and agentic AI: the need for state synchronization. API and AI gateways require a robust persistence layer to synchronize data, whether it's for governing AI token usage, facilitating agent-to-agent communication, or boosting performance through caching.

Kong Simplifies Multicloud Cloud Gateways with Managed Redis Cache

As enterprises race to deploy multicloud architectures and Agentic AI, they face a common bottleneck: "state." To govern AI token usage, manage agent-to-agent communication, or optimize performance via caching, API and AI gateways require a persistence layer to synchronize data. We’re excited to share the GA of Managed Redis cache for Kong Dedicated Cloud Gateways (DCGW).

Best AI test automation tools for fast, high-quality releases

The promise of test automation was simple: automate repetitive testing tasks, catch bugs faster, and ship quality software at scale. Yet for most development teams, that promise remains unfulfilled. Traditional test automation frameworks demand specialized coding skills, require constant maintenance when applications change, and create bottlenecks that slow down release cycles rather than accelerate them.