Atlanta, GA, USA
2020
  |  By JT, Matthew LeRay, and Hugh Brien
Part of the motivation for this post is how cloud API economics are shifting: Anthropic is moving large enterprise customers toward per-token, usage-based billing (unbundled from flat seat fees), which makes “always call the API” a moving cost line for teams at scale. A hybrid or local layer is one way to keep spend bounded while you still use premium models where they matter.
  |  By Josh Thornton
The terminal is a developer’s most trusted tool. It sees your source code, your secrets, the commands you run in production. When a terminal adds AI and cloud features, it’s worth asking what it’s doing on the network.
  |  By Matthew LeRay
The alert fires at 2 AM. Your observability platform’s synthetic test just failed. Login is broken. So you open your laptop, pull up the dashboard, and stare at a single red dot: the browser test. You know the problem is somewhere in the stack, but not where. Is it the auth service? The token validator? The user profile API? The API gateway timing out? You’re now about to spend the next 45 minutes correlating traces, tailing logs, and manually hitting endpoints until you find it.
  |  By Josh Thornton
The biggest risk to your product isn’t AI-generated code that doesn’t work. It’s generated code that seems fine. AI doesn’t optimize for correctness. It creates something passable. Something that passes the smell test. And when everybody in the industry is pushed to move faster and do more with less, you end up shipping software that looks correct. It passed your quick visual check. It passed all the tests. But no one ever fully understood it.
  |  By Ken Ahrens
Every week there is another model drop, another agent framework, and another workflow tweak you are supposed to evaluate. Meanwhile, the largest companies, the ones operating at the highest scale and leaning hardest on AI, are also the ones making headlines for reliability strain: capacity limits, outages, and services that buckle under load.
  |  By Alan Mon
Last week, a fleet of autonomous robotaxis in China suddenly stopped working—at scale. Over a hundred vehicles stalled across a city, stranding passengers in traffic and raising immediate concerns about safety, reliability, and trust in autonomous systems. This wasn’t just a bad day for self-driving cars. It was a distributed systems failure, one that happened in the physical world, not just in dashboards.
  |  By Kush Mansingh
OpenTelemetry is great at answering one question: “what just broke?” The problem is that most teams need a different answer first: “what is about to break in this release?” That is where trace-based testing comes in, especially for teams running a vendor-neutral OTel stack (Collector + Tempo/Jaeger + Prometheus) and needing deterministic release gates.
  |  By Matthew LeRay
The era of the “copilot” is ending. We are moving rapidly toward the era of the autonomous software factory, where autonomous agents don’t just autocomplete our code—they investigate, plan, test, and merge entire features while we sleep. But this shift has exposed a critical flaw in how we consume AI. For the past decade, the default motion for enterprise software has been SaaS. It’s easy, frictionless, and managed by someone else.
  |  By Ken Ahrens
I worked in observability for years, and the same pattern showed up across teams. An alert fired, the on-call rotation scrambled, and everyone did what they had to do to stabilize production. Then came the retrospective. Once the immediate pressure was gone, the conversation shifted to one question: how do we make sure this never happens again? My friend Jade Rubick coined a name for that principle: DRI, “don’t repeat the incident”.
  |  By Ken Ahrens
Most founders have a build problem. Very few have solved the distribution problem. That gap is where startups die. When someone comes to me with a startup idea (and this happens more than you’d think), I always ask the same question first.
  |  By Speedscale
The AI-native stack is currently facing a "Distrust Paradox": while 84% of developers use AI coding tools, trust in the results has plummeted to 29%. Ken demonstrates how to stabilize rapid software delivery by replacing unreliable "AI vibes" with a deterministic testing foundation.
  |  By Speedscale
If it passes the tests, it’s not my problem. If you’re still manually checking every line of code in 2026, you’re just wasting company time. Let the AI cook and go touch some grass. Check out: speedscale.com.
  |  By Speedscale
The "Buy vs. Build" rule is dead. Generic CRMs are too slow for lean startups, so we built our own. In this video, Ken breaks down "Radar," the custom AI dashboard we use at Speedscale to automate prospecting and outreach. Stop fighting bloated SaaS and start building the exact tools you need to solve your distribution problem. Learn more: speedscale.com.
  |  By Speedscale
Forecast latency, throughput and headroom before every deploy.

Continuous Resiliency from Speedscale gives you the power of a virtual SRE-bot working inside your automated software release pipeline. Forecast the real-world conditions of every build, and know you’ll hit your SLO’s before you go to production.

Feed Speedscale traffic (or let us listen) and we’ll turn it into traffic snapshots and corresponding mock containers. Insert your own service container in between for a robust sanity check every time you commit. Understand latency, throughput, headroom, and errors -- before you release! The best part? You didn’t have to write any scripts or talk to anyone!

Automated Traffic Replay for Every Stakeholder:

  • DevOps / SRE Pros: Understand if your app will break or burn up your error budget before you release.
  • Engineering Leads: Let Speedscale use traffic to autogenerate tests and mocks. Introduce Chaos testing and fuzzing.
  • Application Executives: Understand regression/performance, increase uptime and velocity with automation.

Before you go to production, run the projection.