Systems | Development | Analytics | API | Testing

Snowflake: Automate tuning for data cloud speed and scale

40% of companies surveyed will increase their AI investment because of advances in GenAI (McKinsey). And 80% plan to maintain or increase their investment in data quality/observability (dbt). With this in mind, Unravel is hosting a live event to help you leverage data observability to achieve speed and scale with Snowflake. Join Unravel Data for this event about automating tuning with AI-powered data performance management for Snowflake with Eric Chu, Unravel Data VP of Product, and Clinton Ford, Unravel Data VP of Product Marketing.

Addressing the Elephant in the Room - Welcome to Today's Cloudera

Hadoop. The first time that I really became familiar with this term was at Hadoop World in New York City some ten or so years ago. There were thousands of attendees at the event – lining up for book signings and meetings with recruiters to fill the endless job openings for developers experienced with MapReduce and managing Big Data. This was the gold rush of the 21st century, except the gold was data.

Which Language Should Testers Use?

Should you design tests in the same language as the application you’re testing, or should you use the language you’re best at? @Hanson Ho recommends using the language that’s most popular in the application’s platform. This way, you’ll have more help available from the community. If you want more insights like this one, check out Test Case Scenario.

Software Test Estimation & 6 Techniques

Software testing evolved from a simple debugging activity in the 1950s to becoming integral to software development with advanced testing tools and test estimation techniques. As a C-level executive or business developer, ensuring your teams provide accurate QA effort estimates is crucial. This precision influences the project outcome and bolsters your credibility with clients. Underestimating QA efforts can lead to potential underperformance and unclear requirements.

Ensuring the performance of your Kafka-dependent applications

In today’s data-driven world, Apache Kafka has emerged as an essential component in building real-time data pipelines and streaming applications. Its fault tolerance, scalability, and ability to handle high throughput makes it a great choice for businesses handling high volumes of data.

Ingest Data Faster, Easier and Cost-Effectively with New Connectors and Product Updates

The journey toward achieving a robust data platform that secures all your data in one place can seem like a daunting one. But at Snowflake, we’re committed to making the first step the easiest — with seamless, cost-effective data ingestion to help bring your workloads into the AI Data Cloud with ease. Snowflake is launching native integrations with some of the most popular databases, including PostgreSQL and MySQL.