Systems | Development | Analytics | API | Testing

Get Your AI to Production Faster: Accelerators For ML Projects

One of the worst-kept secrets among data scientists and AI engineers is that no one starts a new project from scratch. In the age of information there are thousands of examples available when starting a new project. As a result, data scientists will often begin a project by developing an understanding of the data and the problem space and will then go out and find an example that is closest to what they are trying to accomplish.

Artificial Intelligence vs. Intelligent Automation: What's the Difference?

AI injects “intelligence” into automation, enabling systems to execute tasks, comprehend complex data, make informed decisions, and learn from outcomes. Unlike technologies such as robotic process automation (RPA), which follow predetermined rules, AI leverages data to evaluate situations and determine the best course of action. Now that we've explored how AI augments traditional automation tools, let's delve deeper into the realm of intelligent automation.

A Software Engineer's Tips and Tricks #3: CPU Utilization Is Not Always What It Seems

Hey there! We're back for our third edition of Tips and Tricks. As we said in our first posts on Drizzle ORM and Template Databases in PostgreSQL, our new Tips and Tricks mini blog series is going to share some helpful insights and cool tech that we've stumbled upon while working on technical stuff. Today's topic is short and sweet. It'll be on CPU utilization and what that metric indicates. If you enjoy it and want to learn more, I encourage you to check out the "further reading" links.

Confluent Unveils New Capabilities to Apache Flink Offering to Simplify AI and Bring Stream Processing to Workloads Everywhere

Confluent's new AI Model Inference seamlessly integrates AI and ML capabilities into data pipelines. Confluent's new Freight clusters offer cost-savings for high-throughput use cases with relaxed latency requirements.

Behind The Scenes Of Snowflake Open Source LLM Arctic

Snowflake CEO Sridhar Ramaswamy and Snowflake Head of AI Baris Gultekin join Adrien Treuille, Director of Product Management, to discuss the launch of Snowflake Arctic, the latest enterprise-grade, truly open large language model (LLM). Snowflake Arctic stands out in the competitive landscape with its exceptional efficiency and cost-effectiveness, emphasizing Snowflake's commitment to open-source development and the future of enterprise AI. Join us to explore how Arctic is set to revolutionize industries by making advanced AI more accessible and trustworthy.

Snowflake's Arctic-TILT: A State-of-the-Art Document Intelligence LLM in a Single A10 GPU

The volume of unstructured data — such as PDFs, images, video and audio files — is surging across enterprises today. Yet documents, which represent a substantial portion of this data and hold significant value, continue to be processed through inefficient and manual methods.

Using Moesif, Kong, and Stripe to Monetize Your AI APIs - Part 1: Integrating The Platforms

As the wave of AI sweeps through the technology landscape, many have hopped on board. Interestingly enough, and often overlooked, is that many AI capabilities are served through APIs. Fancy user interfaces integrate with the actual mechanisms where the magic happens: the APIs. So, when generating revenue through AI platforms, the APIs drive the revenue.