Systems | Development | Analytics | API | Testing

Latest News

RAG vs Fine-Tuning: Navigating the Path to Enhanced LLMs

RAG and Fine-Tuning are two prominent LLM customization approaches. While RAG involves providing external and dynamic resources to trained models, fine-tuning involves further training on specialized datasets, altering the model. Each approach can be used for different use cases. In this blog post, we explain each approach, compare the two and recommend when to use them and which pitfalls to avoid.

Commercial vs. Self-Hosted LLMs: A Cost Analysis & How to Choose the Right Ones for You

As can be inferred from their name, foundation models are the foundation upon which developers build AI applications for tasks like language translation, text summarization, sentiment analysis and more. Models such as OpenAI's GPT, Google's Gemini, Meta’s Llama and Anthropic’s Claude, are pre-trained on vast amounts of text data and have the capability to understand and generate human-like language.

Manage Resource Utilization and Allocation with ClearML

Written by Noam Wasersprung, Head of Product at ClearML Last month we released the Resource Allocation & Policy Management Center to help teams visualize their compute infrastructure and understand which users have access to what resources. This new feature makes it easy for administrators to visualize their resource policies for enabling workload prioritization across available resources.

Unparalleled Productivity: The Power of Cloudera Copilot for Cloudera Machine Learning

In the fast-evolving landscape of data science and machine learning, efficiency is not just desirable—it’s essential. Imagine a world where every data practitioner, from seasoned data scientists to budding developers, has an intelligent assistant at their fingertips. This assistant doesn’t just automate mundane tasks but understands the intricacies of your workflows, anticipates your needs, and dramatically enhances your productivity at every turn.

Transforming Enterprise Operations with Gen AI

Enterprises are beginning to implement gen AI across use cases, realizing its enormous potential to deliver value. Since we are all charting new technological waters, being mindful of recommended strategies, pitfalls to avoid and lessons learned can assist with the process and help drive business impact and productivity. In this blog post, we provide a number of frameworks that can help enterprises effectively implement and scale gen AI while avoiding risk.

Snowflake ML Now Supports Expanded MLOps Capabilities for Streamlined Management of Features and Models

Bringing machine learning (ML) models into production is often hindered by fragmented MLOps processes that are difficult to scale with the underlying data. Many enterprises stitch together a complex mix of various MLOps tools to build an end-to-end ML pipeline. The friction of having to set up and manage separate environments for features and models creates operational complexity that can be costly to maintain and difficult to use.

Future-Proofing Your App: Strategies for Building Long-Lasting Apps

The generative AI industry is changing fast. New models and technologies (Hello GPT-4o) are emerging regularly, each more advanced than the last. This rapid development cycle means that what was cutting-edge a year ago might now be considered outdated. The rate of change demands a culture of continuous learning and technological adaptation.

LLM Validation and Evaluation

LLM evaluation is the process of assessing the performance and capabilities of LLMs. This helps determine how well the model understands and generates language, ensuring that it meets the specific needs of applications. There are multiple ways to perform LLM evaluation, each with different advantages. In this blog post, we explain the role of LLM evaluation in AI lifecycles and the different types of LLM evaluation methods. In the end, we show a demo of a chatbot that was developed with crowdsourcing.

How ClearML Helps Teams Get More out of Slurm

It is a fairly recent trend for companies to amass GPU firepower to build their own AI computing infrastructure and support the growing number of compute requests. Many recent AI tools now enable data scientists to work on data, run experiments, and train models seamlessly with the ability to submit their jobs and monitor their progress. However, for many organizations with mature supercomputing capabilities, Slurm has been the scheduling tool of choice for managing computing clusters.