Systems | Development | Analytics | API | Testing

ClearML

How to Optimize Huggingface Models for Production

Deploying models is becoming easier every day, especially thanks to excellent tutorials like Transformers-Deploy. It talks about how to convert and optimize a Hugging face model and deploy it on the Nvidia Triton inference server. Nvidia Triton is an exceptionally fast and solid tool and should be very high on the list when searching for ways to deploy a model. If you haven’t read the blogpost yet, do it now first, I will be referencing it quite a bit in this blogpost.

How You Can Contribute to ClearML's MLOps Platform

ClearML is an open source MLOps platform, and we love the community that’s been growing around us over the last few years. In this post, we’ll give you an overview of the structure of the ClearML codebase so you know what to do when you want to contribute to our community. Prefer to watch the video? Click below: First things first. Let’s take a look at our GitHub page and corresponding repositories. Later on, we’ll cover the most important ones in detail.