Systems | Development | Analytics | API | Testing

Using Agentic Frameworks to Build New AI Services

The original promise of AI was that it would write most of the code for us. In reality, we’re not there yet. So where can AI meaningfully improve developer productivity today? In this post, we look at how AI powers development productivity across the SDLC, practical tools to use and frameworks for overcoming AI operationalization bottlenecks.

7 RAG Evaluation Tools You Must Know

RAG evaluation measures how effectively a system retrieves relevant context and uses it to generate grounded answers. These evaluations detect hallucinations, measure retrieval precision and reveal where pipelines degrade after model updates or knowledge-base changes. Engineers rely on these tools to maintain output quality, prevent regressions, validate prompt and architecture choices and ensure that production answers stay aligned with trusted sources.

Introducing MLRun v1.10: New tools for building agents and monitoring gen AI

MLRun 1.10, the latest version of our open source AI orchestration framework, is available today to all users. Iguazio started out as a platform to operationalize enterprise machine learning projects. Though we’ve been through quite a few waves of AI in just a short time, the underlying challenges are the same: getting from experimentation to production remains a major blocker.

Banking on Gen AI: Driving Profitable and Scalable Client Engagement with Gen AI Copilots

Wealth management has always been about personal touch. Relationship managers provide a white-glove service to elite clientele - guiding investments, financial plans, and more. However, they’re under growing pressure to serve more clients and drive bank revenue, without diluting that personal connection and service quality. This dual mandate is placing relationship managers in a catch-22 situation. If they serve more clients their ability to provide personalized services diminishes, and vice versa.

LLM Observability Tools in 2025

1. Organizations have moved beyond pilots and are embedding LLMs into production workflows across customer support, finance, security, and software delivery. 2. LLM observability mitigates risks like hallucinations, bias, compliance breaches, and runaway costs. 3. LLM observability requires prompt/response tracking, hallucination detection, drift monitoring, RAG pipeline visibility, and long-term context tracing. 4.

Managing AI Risks When Implementing Gen AI

As enterprises embed gen AI into their workflows, many are discovering a minefield of risks. Data privacy breaches, misinformation, adversarial attacks and hidden bias are just a few of the challenges that can derail gen AI initiatives. These aren't just technical concerns, they're business-critical issues that can erode trust, trigger legal consequences, and tarnish reputations.

Best Practices to Develop, Deploy, and Manage Gen AI Copilots

Generative AI copilots are moving from experimental tools to core enterprise solutions. But too often, organizations rush into development, only to discover adoption stalls because the copilot doesn’t solve a specific user problem, lacks trust safeguards, or can’t scale reliably. This guide lays out best practices across the entire lifecycle, from planning and building, to deployment, monitoring, and long-term maintenance.

Orchestrating Multi-Agent Workflows with MCP & A2A

Multi-agent workflows are the latest technological gen AI advancements. In this blog, we explore how to develop such systems, overcome operational challenges, improve system observability, and enable seamless collaboration between agents in complex AI pipelines. We’ll cover architecture, A2A and MCP protocols and introduce Google Cloud’s agentic marketplace.

LLM Evaluation and Testing for Reliable AI Apps

As LLMs become central to AI-driven products like copilots and customer support chatbots, data science teams need to ensure the LLM performs well for the use case. The process of LLM evaluation ensures reliability, safety and performance in production AI systems. In this guide, we explore how to approach evaluations across development and production lifecycles, what frameworks to use, and how the integration between open-source MLRun and Evidently AI enables more scalable, structured testing.

13 Best Free Datasets for Call Centers and Telcos

Customer service chatbots and co-pilots and smart call center analysis applications are prime use cases for AI and generative AI. These AI systems and agents can provide real-time recommendations, support customer service at scale, generate insights that can be used in downstream applications to reduce churn and increase revenue, and more. How can customer service organizations grow and optimize their use of data and AI?