Systems | Development | Analytics | API | Testing

ClearML

How to Optimize Huggingface Models for Production

Deploying models is becoming easier every day, especially thanks to excellent tutorials like Transformers-Deploy. It talks about how to convert and optimize a Hugging face model and deploy it on the Nvidia Triton inference server. Nvidia Triton is an exceptionally fast and solid tool and should be very high on the list when searching for ways to deploy a model. If you haven’t read the blogpost yet, do it now first, I will be referencing it quite a bit in this blogpost.

How You Can Contribute to ClearML's MLOps Platform

ClearML is an open source MLOps platform, and we love the community that’s been growing around us over the last few years. In this post, we’ll give you an overview of the structure of the ClearML codebase so you know what to do when you want to contribute to our community. Prefer to watch the video? Click below: First things first. Let’s take a look at our GitHub page and corresponding repositories. Later on, we’ll cover the most important ones in detail.

How ClearML Helps Daupler Optimize Their MLOps

We recently had a chance to catch up with Heather Grebe, Senior Data Scientist at Daupler, which offers Daupler RMS, a 311 response management system, used by more than 200 cities and service organizations across North America and internationally. This platform helps utilities, public works, and other service organizations coordinate and document response efforts while reducing workload and collecting insights into response operations.