Systems | Development | Analytics | API | Testing

ROI of Digital Twin Testing: Cut Testing Costs by 50%

When engineering leaders review their cloud bills, they often focus on production costs—the infrastructure serving real users, processing real transactions, generating real revenue. But there’s a shadow cost lurking in every cloud environment that often goes unnoticed until it becomes painful: non-production infrastructure.
Sponsored Post

Peeking Under the Hood with Claude Code

Claude is one of the go-to AI-native code editors for developers. Because it's a simple chatbot interface housed inside a familiar CLI, it provides a pretty smooth path between traditional IDEs and agentic AI. But what's actually happening behind the scenes when you ask it to write code, generate a test, or debug an issue? Who and what is it talking to behind the scenes? Can I prevent data leakage or do I need to add another layer to my tin foil hat? To answer these questions, I used proxymock to inspect the network traffic flowing from the Claude IDE.

Moving Our Observability Data Collector from Sidecars to eBPF

For years, the Kubernetes sidecar pattern has been a practical way to capture observability data. Running a collector alongside each application pod gave us deep visibility into traffic, including full request and response payloads across supported protocols. However, as cloud-native environments have grown more complex, the limitations of sidecars—such as resource overhead, operational complexity, and scaling challenges—have become more apparent.

How VPS Architecture Solves The Problem

Virtual private servers approach this differently. They tackle contention head-on through dedicated resource allocation. Instead of letting resources slosh around freely between all sites like shared hosting does, VPS technology carves a physical server into isolated virtual machines. Each one operates independently. Each gets its own slice of CPU cores, RAM, and storage that nobody else can touch.

Mock vs Stub: Essential Differences

When discussing the process of testing an API, one of the most common sets of terms you might encounter are “mocks” and “stubs.” These terms are quite ubiquitous, but understanding exactly how they differ from one another - and when each is the correct method for software testing - is critical to building an appropriate test and validation framework. In this blog, we’re going to talk about the differences and similarities between mocks and stubs.

The CES Hangover: 3 Expensive Hardware Fails That Were Actually Software Problems

The dust has settled on Las Vegas. We saw transparent TVs, cars that drive sideways, and enough “AI-powered” toothbrushes to confuse a dentist. CES is incredible at selling the dream of hardware. The demos are slick, the lighting is perfect, and everything works on the showroom floor. But as engineers, we know the dirty secret of CES: The hardware is the easy part.

Supercharge your LLM Using Production Data Context

Are your LLM coding agents (like Cursor or Claude Code) hallucinating fixes because they don't know what's actually happening in production? In this video, Matt from Speedscale shows you how to bridge the gap between your local IDE and live production traffic using the Model Context Protocol (MCP). Most observability tools just give you telemetry. Speedscale’s MCP server gives your agent the "inner workings" of actual API calls and payloads, so it can check its assumptions against reality. No more "vibe-coding" and hoping it works; let your agent find the 500 errors and rate limits for you.