Systems | Development | Analytics | API | Testing

Koyeb

eBPF: The future of the service mesh and network innovation

The conversations around eBPF and how this technology will shape the future of the service mesh caused a huge buzz in the last year — yes, bee pun intended. eBPF lets you run sandboxed programs in an operating system kernel. Imagining how eBPF could improve the service mesh brings exciting possibilities, but it also raises security and operational concerns given the current state and limitations of eBPF.

Introducing the Koyeb Terraform Provider

Today, we are glad to release the Koyeb Terraform Provider and announce that we are a Hashicorp Technology partner. Terraform has become an essential tool for many developers who manage infrastructure and applications running in the cloud. We are excited to offer our users a way to manage their Koyeb resources using Terraform.

US-East region is live: deploy your apps in Washington, DC

Our US-East region is now available to deploy all your full-stack apps and APIs! 🇺🇸 Connect a Git repository, select your region, we will build your app and run it inside of MicroVMs on top of high-performance BareMetal servers in the most interconnected location worldwide: Washington, US-East. Sign up and give it a try!

New Frankfurt location: deploy high-performance apps in Europe

Today, we are super excited to announce the grand opening of our Frankfurt core location in Germany! 🇩🇪🇪🇺 All users can now deploy their applications in the heart of Europe with one click, or API call, and enjoy all of the platform's features with the same pricing. Give it a try, sign up and deploy your full-stack apps now! As usual, we will seamlessly deploy your services across a fleet of high-performance MicroVMs running on top of BareMetal servers.

What is continuous deployment?

After working on a new feature or an improvement, it's normal that you want to deliver these changes to your users as fast as possible. Depending on your deployment pipeline, confidently deploying changes to production might include coordinating with whoever is leading on-call, getting their approval, manually deploying your changes, spending hours to monitor how it goes, and performing a rollback in case something goes wrong.

What is a service mesh?

A service mesh is a dedicated infrastructure layer that manages traffic, also known as communication, between services in applications composed of containerized microservices. It is a critical component in a microservices architecture, responsible for the secure, fast, and reliable communication between services. This article answers a lot of the questions you may have about service meshes: What are they and how do they work? Who is using them and why?

Distributed tracing with Envoy, Kuma, Grafana Agent, and Jaeger

As a cloud service provider, observability is a critical subject as it's strongly related to the availability of the services running on the platform. We need to understand everything that is happening on our platform to troubleshoot errors as fast as possible and improve performance issues. A year ago, while the platform was still in private beta, we faced a tough reliability issue: users were facing random 500 errors when accessing their applications.

Heroku's free tier legacy: The shoulders we stand on 15 years later

On August 15, 2022, we learned the shocking news that free Heroku Dynos, Heroku Postgres and Heroku Data for Redis will no longer be available and existing services will either need to upgrade to a paid service or find another hosting solution by November 28, 2022. Since its conception, Heroku has changed the way developers, hobbyists, students, and indie hackers deploy applications.

What is an API Gateway?

An API gateway is a server that sits between your backend services and your users. It provides an abstraction layer that helps you manage the communication between clients and your services. API gateways are a single point of entry into a microservices application and they work like a reverse proxy: They receive API calls from clients, route them to the right microservices in the backend, and return aggregated responses to clients.