Systems | Development | Analytics | API | Testing

How to Run Workloads on Spark Operator with Dynamic Allocation Using MLRun

With the Apache Spark 3.1 release in early 2021, the Spark on Kubernetes project has been production-ready for a few years. Spark on Kubernetes has become the new standard for deploying Spark. In the Iguazio MLOps platform, we built the Spark Operator into the platform to make the deployment of Spark Operator much simpler.

What is Self Service Analytics? The Role of Accessible BI Explained

Self service analytics (also called self-service business intelligence, or self-service BI) is a term commonly used among analytics vendors and organizations adopting BI, often in the context of being the next big thing in driving more people to use data to find insights. But what is self service analytics? How does self service analytics work? And why does self service analytics matter?

AI at Scale isn't Magic, it's Data - Hybrid Data

A recent VentureBeat article , “4 AI trends: It’s all about scale in 2022 (so far),” highlighted the importance of scalability. I recommend you read the entire piece, but to me the key takeaway – AI at scale isn’t magic, it’s data – is reminiscent of the 1992 presidential election, when political consultant James Carville succinctly summarized the key to winning – “it’s the economy”.

Pros & Cons of Using a Customer Data Platform as Your Data Warehouse

Does your Ecommerce business team understand the customer journey? By tracking the history of individual customer behavior and customer interactions across different channels, your organization can better understand what motivates your audience — and cater to them with the right marketing campaigns.

How Twitter maximizes performance with BigQuery

How does a tweet go from one person to hundreds of millions of people? How does the data process so quickly? In this episode of Architecting with Google Cloud, Priyanka chats with Gary and Saurabh from Twitter about how data from over 200 million users goes through the Twitter data center and Google Cloud. Watch along and learn how data stored across tens of thousands of BigQuery tables in Google Cloud runs millions of queries each month.

How to Accelerate HuggingFace Throughput by 193%

Deploying models is becoming easier every day, especially thanks to excellent tutorials like Transformers-Deploy. It talks about how to convert and optimize a Huggingface model and deploy it on the Nvidia Triton inference engine. Nvidia Triton is an exceptionally fast and solid tool and should be very high on the list when searching for ways to deploy a model. Our developers know this, of course, so ClearML Serving uses Nvidia Triton on the backend if a model needs GPU acceleration.