Systems | Development | Analytics | API | Testing

Scaling Kafka Brokers in Cloudera Data Hub

This blog post will provide guidance to administrators currently using or interested in using Kafka nodes to maintain cluster changes as they scale up or down to balance performance and cloud costs in production deployments. Kafka brokers contained within host groups enable the administrators to more easily add and remove nodes. This creates flexibility to handle real-time data feed volumes as they fluctuate.

A Guide to Principal Component Analysis (PCA) for Machine Learning

Principal Component Analysis (PCA) is one of the most commonly used unsupervised machine learning algorithms across a variety of applications: exploratory data analysis, dimensionality reduction, information compression, data de-noising, and plenty more. In this blog, we will go step-by-step and cover: Before we delve into its inner workings, let’s first get a better understanding of PCA. Imagine we have a 2-dimensional dataset.

7 Best Change Data Capture (CDC) Tools of 2022

As your data volumes grow, your operations slow down. Data ingestion - extraction of all underlying datasets, transformation, and loading in a storage destination (such as a PostgreSQL or MySQL database) - becomes sluggish, impacting processes down the line. Affecting your data analytics and time to insights. Change Data Capture (CDC) makes data available faster, more efficiently, and without sacrificing data accuracy. In this blog we are going to overview the 7 best change data capture tools of 2022.

Software Quality Management Best Practices | 5 Do's & Don'ts

Achieving optimal software reliability and quality management processes sit at the core of a memorable digital experience. Quality management in software can be summarized in two points: Stakeholders and management always want their digital products to successfully launch. Software testing is normally seen as rejecting builds and stretching out the delivery date. Why is that?

How to Do Data Labeling, Versioning, and Management for ML

It has been months ago when Toloka and ClearML met together to create this joint project. Our goal was to showcase to other ML practitioners how to first gather data and then version and manage data before it is fed to an ML model. We believe that following those best practices will help others build better and more robust AI solutions. If you are curious, have a look at the project we have created together.

How to Distribute Machine Learning Workloads with Dask

Tell us if this sounds familiar. You’ve found an awesome data set that you think will allow you to train a machine learning (ML) model that will accomplish the project goals; the only problem is the data is too big to fit in the compute environment that you’re using. In the day and age of “big data,” most might think this issue is trivial, but like anything in the world of data science things are hardly ever as straightforward as they seem.