Systems | Development | Analytics | API | Testing

Accelerating and Scaling AI Deployments Across Hybrid Environments - MLOps Live #40 with Safaricom

Safaricom, one of the most AI-mature mobile operators, delivers predictive modeling and hyper-personalized financial services to millions of users. But operational challenges were slowing down deployments—limiting their ability to scale and act in real time. In this session, Safaricom’s AI team shares how they: Watch now to learn how they overcame bottlenecks, scaled faster, and unlocked real-time impact at massive scale with the Iguazio technology.

Best Practices to Develop, Deploy, and Manage Gen AI Copilots

Generative AI copilots are moving from experimental tools to core enterprise solutions. But too often, organizations rush into development, only to discover adoption stalls because the copilot doesn’t solve a specific user problem, lacks trust safeguards, or can’t scale reliably. This guide lays out best practices across the entire lifecycle, from planning and building, to deployment, monitoring, and long-term maintenance.

From Bias to Breach: Navigating the Challenges in Machine Learning | Deepika Hanumanthu

As machine learning models continue to shape critical decisions in areas like healthcare, finance, and security, understanding their vulnerabilities has become paramount. “Breaking the Machine” delves into adversarial attacks—carefully crafted actions designed to exploit model weaknesses, leading to incorrect predictions. This talk explores the two main categories of adversarial attacks, White Box and Black Box, and their subcategories of targeted and untargeted attacks.

Build Custom AI Workflows in Minutes with ClearML's Native Application Ecosystem

By Erez Schnaider, Technical Product Marketing Manager, ClearML The number of AI applications are rapidly increasing, and it can be difficult to keep up. Every month brings a new protocol, LLM, or tool. In this environment, the true strength of a platform is measured not only by its core features but also by its extensibility and adaptability to change. Many platforms address this challenge by hosting OSS tools or exposing API connections.

Powering the Next Generation of AI Agents with ClearML's GenAI App Engine

The era of simple, scripted AI is swiftly fading. We’re now witnessing the dawn of AI Agents: sophisticated, self-governing digital entities that possess the capacity to comprehend their surroundings, navigate intricate problems, and execute purposeful actions. Multi-agent systems take this even further, multiplying these capabilities by enabling teams of AI agents to collaborate, delegate tasks, and solve challenges collectively in ways a single agent cannot achieve alone.

Seamless AI Portability: Lift-and-Shift AI Workloads Without the Headaches

Every week brings a new breakthrough in AI, and a new strain on infrastructure. One day, you’re fine-tuning a small model on a local machine. The next, you’re trying to schedule workloads that consume dozens of GPUs across multiple locations. And that doesn’t include the pace of new hardware, which increases what you can do.

Orchestrating Multi-Agent Workflows with MCP & A2A

Multi-agent workflows are the latest technological gen AI advancements. In this blog, we explore how to develop such systems, overcome operational challenges, improve system observability, and enable seamless collaboration between agents in complex AI pipelines. We’ll cover architecture, A2A and MCP protocols and introduce Google Cloud’s agentic marketplace.

ClearML Enterprise 3.26 Is Here: Static Routes, NIM Deployment, SGLang Support, and More

ClearML Enterprise v3.26 brings powerful upgrades across model deployment, NIMs container deployment, and dataset management – all part of our end-to-end platform for managing and scaling AI in the enterprise.

LLM Evaluation and Testing for Reliable AI Apps

As LLMs become central to AI-driven products like copilots and customer support chatbots, data science teams need to ensure the LLM performs well for the use case. The process of LLM evaluation ensures reliability, safety and performance in production AI systems. In this guide, we explore how to approach evaluations across development and production lifecycles, what frameworks to use, and how the integration between open-source MLRun and Evidently AI enables more scalable, structured testing.