Systems | Development | Analytics | API | Testing

Koyeb

Serverless Postgres Public Preview

We're excited to announce the Koyeb Serverless Postgres public preview - a fully managed, fault-tolerant, and scalable serverless Postgres Database Service. What do all modern applications have in common? They all have APIs, workers, and databases. Deploying APIs and workers with Koyeb has long been possible. Starting today, you can spin up databases too! Using Koyeb Serverless Postgres, you can easily start a resilient Database Service alongside your apps in a few seconds.

We raised $7M to Simplify App Deployment with our Global Serverless Platform

We are thrilled to share that we’ve raised $7M in seed funding! At Koyeb, we simplify app deployment with our global serverless platform. We provide an easy way to deploy full-stack applications and databases in production, everywhere, in minutes. We’re focused on allowing developers and businesses to seamlessly build, run, and scale any application globally, with no code rewrite or infrastructure management.

Sustaining free compute in a hostile environment

One year ago, Heroku sunsetted its free tier. Today, we want to reaffirm our commitment to maintaining our free tier, dive into why offering a free tier for compute is complicated (we are looking at you crypto miners), take the time to explain how we intend to sustain it, and explain why we are so committed to providing a free tier. Long story short: we aim to keep a free tier thanks to how we control our costs.

Building a global deployment platform is hard, here is why

If you ever tried to go global, you have probably faced a reality check. A whole new set of issues starts to appear when you start to operate a workload over multiple locations across the globe: So it looks like a great idea in theory, but in practice, all of this complexity multiplies the number of failure scenarios to consider!

The Global Deployment Engine: How We Deploy Across Continents

We previously explored how we built our own Serverless Engine and a multi-region networking layer based on Nomad, Firecracker, and Kuma. But what about the architecture of the engine that orchestrates these components across the world? This is an interesting topic to work on and we thought it could be useful to share some internals out there. Put on your scuba equipment, this is a deep dive into our architecture and the story of how we built our own global deployment engine.

Koyeb Metrics: Built-in Observability to Monitor Your Apps Performances

At Koyeb, we're working to build the most seamless way to deploy apps to production without worrying about infrastructure. But there's still plenty to keep you busy at the application layer with performance tuning and troubleshooting. That's why we're introducing Metrics — an easy way to monitor and troubleshoot application performance. Deploying on Koyeb makes thinking about infrastructure or orchestration unnecessary.

Accelerate Docker builds with cache

Speed and efficiency are paramount during the build process. If you use a Dockerfile to build your container images from source code, you want to know about build cache. In this blog post, we’ll talk about what happens when you create a Docker image using a Dockerfile, how caching works with Docker, and how to optimize your Dockerfiles to maximize the benefits of build cache with Docker and on Koyeb.

Dockerfile Deployment on High-Performance MicroVMs is GA

Today, we are excited to announce the support of Dockerfile based deployments in general availability. You can now deploy any GitHub repository that contains a Dockerfile across all our locations worldwide. It can be used to deploy APIs, full-stack applications as well as workers with no extra cost. Building and deploying using Dockerfiles offers more flexibility: you can deploy any kind of application, framework, and runtime, including with custom system dependencies.

Deploy and scale high-performance background jobs with Koyeb Workers

Today, we are thrilled to announce workers are generally available on Koyeb! You can now easily deploy high performance workers to process background jobs in all of our locations. It's now simple to deploy workers from a GitHub repository and rely on our built-in CI/CD engine: simply connect your repository and we build, deploy, and scale your workers on high-performance servers all around the world.