Systems | Development | Analytics | API | Testing

February 2023

Logilica becomes Synopsys Alliance Partner

SYDNEY, NSW — 28 February 2023. Logilica, a leader in value stream analytics, today announced becoming a Synopsys Technology Alliance Partner. Renowned for their leadership in the application security testing market Synopsys features flagship products such as Coverity, Black Duck and CodeDx to enable secure development across the supply chain.

Winter '23: Accelerate time to value and data modernization efforts

We recognize that extracting the most return from your data isn’t always easy. In fact, for many organizations, it rarely is. According to a recent survey, 97% of respondents said they face challenges using data effectively. And NewVantage Partners found that only 24% of executives say their organizations are data-driven. There are simply too many data sources, too much data volume, and too few resources for these organizations to effectively manage the growing demand for trusted data.

Data Engineer vs Analytics Engineer: How to choose the career that's right for you

A little over a year ago, I found myself feeling stuck in my role as a data engineer. I had majored in business in college and was looking to connect more with that side of things. I enjoyed my tasks as a data engineer but I wanted more flexibility and creativity. I wanted to be involved in business decisions rather than my tasks already being decided for me.

Three Ways to Use Data to Drive a More Strategic Approach to People In 2023

Business priorities, and therefore people priorities, are shifting at a pace like never before thanks to increasing interest rates, energy prices, political uncertainty, global supply constraints, and issues finding and funding expensive talent. People and HR leaders are going to be under considerable pressure to adapt and evolve quickly throughout 2023 in order to support their CEO’s much bigger focus on the workforce.

The Role Data Will Play In Driving Future Growth In The Telecom Industry

What will it take for companies in the telecom industry to return to top-line growth? According to Karl Bjurstrom, EVP and Global Head of Tech & Telecom Industries at Capgemini Invent, it’s all about the data. Capgemini is a leading strategic partner to companies worldwide looking to effect a digital transformation of their businesses. In this episode of “Data Cloud Now,” host Ryan Green asks Bjurstrom for his insights on the challenges and opportunities for growth in the telecom industry.

Master Data Management: Why CXOs Should Embrace MDM

Data has always been an important part of any business. A company’s data is its most valuable asset as it contains information about the procurement of resources to the customer relationship. With the rise of Big Data, more companies are trying to collect and store all information about their customers, products, employees, and other aspects of their operations to better understand them.

Power BI Data Preparation in 5 steps

Microsoft Power BI is a fantastic tool for data visualization and business intelligence. But as any data analyst can attest, before you can build metrics and dashboards, you need to spend more than 1/3 of your time preparing and cleaning the dataset for your Power BI instance. Luckily, there is a better way to cut down on this time-consuming task. In this article you will learn.

Snowflake Workloads Explained: Data Lakes

Snowflake’s cross-cloud platform breaks down silos by supporting a variety of data types and storage patterns. Data engineers, data scientists, analysts, and developers across organizations can access governed structured, semi-structured, and unstructured data for a variety of workloads, without resource contention or concurrency issues.

Isn't the Data Warehouse the Same Thing as the Data Lakehouse?

A data lakehouse is a data storage repository designed to store both structured data and data from unstructured sources. It allows users to access data stored in different forms, such as text files, CSV or JSON files. Data stored in a data lakehouse can be used for analysis and reporting purposes.

The 7 Best Airflow Alternatives in 2023

Who doesn’t love Apache Airflow? The Python-based open-source tool allows us to schedule and automate workflows with DAGs (Directed Acyclic Graphs). Data teams use Airflow for a myriad of use cases: from building ETL data pipelines to launching machine learning apps. The open-source tool makes workflow management easy: it is extensible, easy to monitor from the intuitive user interface in real time, and it allows you to build dependencies between jobs.

DataFinOps: More on the menu than data cost governance

IT and data executives find themselves in a quandary about deciding how to wrangle an exponentially increasing volume of data to support their business requirements – without breaking an increasingly finite IT budget. Like an overeager diner at a buffet who’s already loaded their plate with the cheap carbs of potatoes and noodles before they reach the protein-packed entrees, they need to survey all of the data options on the menu before formulating their plans for this trip.

Business Metric Strategies: How To Choose the Right Framework To Measure Success

Business metrics provide a quantifiable way to measure the success of a business. They help organizations to track their progress internally while also serving as a way to communicate the performance of a business to stakeholders and external parties. There are hundreds of metrics that could be factored into these calculations, but they need to be specific to an organization to be effective.

Transaction Support Using Apache Phoenix

This video provides a short demo on Apache Phoenix transaction support in Cloudera Operational Database (COD). COD supports Apache OMID (Optimistically Transaction Management In Datastores) transactional framework. The transaction support in COD enables you to perform complex distributed transactions and run atomic cross-row and cross-table database operations. The atomic database operations ensure that your database operations must either be completed or terminated.

Computer Vision 101: What It Is and Why It Matters

10 years ago, it would be ridiculous for people to believe that someday they would be able to use their faces to unlock their phones. That’s because it had been extremely difficult to create cartoon characters without profound drawing skills – but now we can easily turn photos into cartoon characters. Struggling with parallel parking? No worries, because self-parking systems are becoming standard equipment in vehicles.

Leveraging Data Analytics in the Fight Against Prescription Opioid Abuse

Every day in the US thousands of legitimate prescriptions for the opioid class of pharmaceuticals are written to mitigate acute pain during post-operation recovery, chronic back and neck pain, and a host of other cases where patients experience moderate-to-severe discomfort.

Seven Ways to Gain Data Clarity in An Uncertain Climate

It’s been a rollercoaster ride for everyone over the last few years, with particular pressure on Chief Financial Officers (CFOs) to support CEOs steering their organizations through things none of us expected to experience in our lifetime. Unfortunately, with the financial markets going into turmoil over the last few months and consumers of all shapes and sizes starting to cut back on spending, the uncertainty isn’t going to stop anytime soon.

Implementing and Using UDFs in Cloudera SQL Stream Builder

Cloudera’s SQL Stream Builder (SSB) is a versatile platform for data analytics using SQL. As apart of Cloudera Streaming Analytics it enables users to easily write, run, and manage real-time SQL queries on streams with a smooth user experience, while it attempts to expose the full power of Apache Flink. SQL has been around for a long time, and it is a very well understood language for querying data.

Snowflake's Phil Kippen Weighs In on Launch of the Telecom Data Cloud

Today Snowflake is officially launching the Telecom Data Cloud. Snowflake’s newest Data Cloud helps telecommunications service providers break down data silos within the business and across the ecosystem, allowing organizations to easily and securely access data in near real time, enrich it with machine learning models, and then share and analyze it to drive better decision-making.

Reverse ETL - A Must-Have for Modern Businesses?

Extract, Transform, Load (ETL), and Extract, Load, Transform (ELT) pipelines are standard data management techniques among data engineers. Indeed, organizations have long been using these processes to create effective data models. However, there has recently been a remarkable rise in the use of Software-as-a-Service (SaaS) based customer relationship management (CRM) apps, such as Salesforce, Zendesk, Hubspot, Zoho, etc., to store and analyze customer data.

The Evolution from DevOps to DataOps

By Jason Bloomberg, President, Intellyx Part 2 of the Demystifying Data Observability Series for Unravel Data In part one of this series, fellow Intellyx analyst Jason English explained the differences between DevOps and DataOps, drilling down into the importance of DataOps observability. The question he left open for this article: how did we get here? How did DevOps evolve to what it is today, and what parallels or differences can we find in the growth of DataOps?

The 7 best Python ETL tools in 2023

In a fast-paced world that produces more data than it can ingest, the right Python ETL tool makes all the difference. But not all Python tools are made the same. Some Python ETL tools are great for writing parallel load jobs for data warehousing, others are specialized for unstructured data extraction. In this article, we’ll explore the 7 best tools for ETL tasks and what business requirements they help you fulfill: Let’s dive right into the best tools and see how they compare.

The Snowflake Telecom Data Cloud

As Snowflake rolls out its new Telecom Data Cloud, “Data Cloud Now” host Ryan Green sits down with Phil Kippen, Global Head of Industry, Telecom, at Snowflake, to discuss what it all means for telecom service providers. During the interview, Kippen notes that the arrival of 5G creates new market opportunities but also new operational complexities for telecom service providers as they take on the task of rolling out new services and managing new infrastructure. He explains that the rollout of the Telecom Data Cloud will help companies achieve operational efficiencies by providing a single, unified platform across all domains and business functions within the telecom service provider environment and across all clouds. In addition, Snowflake will help service providers create a new marketplace that will enhance their ability to find new ways to monetize their data and applications and will help them work with partners across the telecom ecosystem to develop new opportunities for collaboration and data sharing.

SaaS In 60 - The Business Glossary

The Business Glossary helps eliminate data confusion by providing a comprehensive library of terms and descriptions that clearly identify how an organization defines its metrics measures and dimensions. It can streamline data-based decisions by eliminating misunderstandings due to competing terminologies or inconsistencies between technology definitions and business language. And a glossary can simplify regulatory compliance and serve as an important tool for data governance.

Traditional BI vs Self-Service Analytics: What's the Difference?

Data has historically been in the hands of a select few in most businesses - until recently. Business intelligence (BI) solutions have evolved dramatically in the last few years to not only be more sophisticated, but simpler and more accessible for regular professionals to use analytics tools and get the insights they need to make decisions.

13 Skills Needed for any Data Engineer According to ChatGPT

Overview With the increasing use and discussion surrounding ChatGPT and its applications, I decided to test out what it says about important skillsets for data engineers. I conducted a search about both soft and hard skills and here is what it came up with. I have added a lot of commentary to each of the 13 skills identified.

How to Create a Dashboard in Kibana

Wondering how to create a dashboard in Kibana to visualize and analyze your log data? In this blog post, we’ll provide a step-by-step explanation of how to create a dashboard in Kibana. You’ll learn how to use Kibana to query indexed application and event log data, filter query results to highlight the most critical and actionable information, build Kibana visualizations using your log data, and incorporate those visualizations into a Kibana dashboard.

Countly Receives SOC 2 Type 2 Certification

As a product analytics solution, establishing a strong sense of trust is vital to what we do. It provides the foundation for the relationships we build with our clients, which in turn allows our clients to confidently provide the services they need to their customers. Data privacy and security are both key components of building a level of trust that is essential for both our customers and ourselves as a company.

Snowflake Snowpark Explained In Under 2 Minutes

Learn about Snowflake Snowpark. What if there was a way to enable your entire team to collaborate securely on the same data in a single platform that just works, regardless of language? Snowpark is here to help. Supercharge your data team to securely build scalable, optimized pipelines in your language. Quickly and efficiently execute machine learning workflows, with your choice of language, from Python to Scala and more. For more details visit: snowflake.com/snowpark

The Top 5 Risks of In-House Development for Data Integration

Data integration is essential for businesses to achieve efficient and effective processes. For successful data integration, businesses can choose between in-house software development or outsourcing their data integration needs. Due to the potential risks of developing in-house, it may make sense for your company to outsource your data integration needs to a trusted third party to ensure all your needs are met on time and on budget.

How Seagate Runs Advanced Manufacturing at Scale With Iguazio

Seagate is the world’s leading data storage solution. Together with Iguazio, Seagate is able to manage data engineering at scale while harnessing petabytes of data, efficiently utilize resources, bridge the gap between data engineering and data science and create one production-ready environment with enterprise capabilities. In this new webinar, Vamsi Paladugu, Sr.

What Is Deep Learning? A Guide to Deep Learning Use Cases, Applications, and Benefits

Deep learning has become a buzzword in the field of artificial intelligence (AI) in recent years. It has achieved impressive results in a variety of tasks, including image and speech recognition, natural language processing, and even playing games.

On-premise vs. Cloud: Finding the Best Solution for your Product

When thinking about a product analytics solution, an essential component of the process lies in deciding between two types of data storage: cloud or on-premise. Each has its advantages and disadvantages depending on a number of factors, such as the stage or size of a business, budget, etc. Before we can tackle these variables, we should understand what each option entails.

How to List Your Product on Snowflake Marketplace | Demo for Providers

Accelerate the growth of your business by monetizing your data and apps in the Data Cloud while serving customers more effectively across clouds and without costly integrations or data pipelines. In this product demo, Snowflake's Lena Pennington shows you how to list your product on Snowflake Marketplace to reach thousands of organizations in the Data Cloud.

What's new in ThoughtSpot Analytics Cloud 9.0.0.cl

Check out ThoughtSpot Analytics Cloud 9.0, our biggest cloud release ever! Sync your data to HubSpot and ServiceNow, connect to PostgreSQL via live query, share your data connections, and embed a new Liveboard experience with tabbed navigation, flexible tile sizing, and more! Also, preview early access features - simply ask your admin to switch them on.

To Data Fabric or not to Data Fabric, is it really a question?

Data fabric is a term used to describe a set of technologies and practices that enable organizations to manage and access data across multiple platforms and environments. This includes supporting an organization’s need to break down data silos, gain more insight into metadata, optimize data sharing across apps and data platforms. Organizations are starting to explore more flexible ways of managing their data ecosystems and ensuring they can leverage data more effectively.

Classifying DNA Sequences into Gene Families on SageMaker

The cost of DNA sequencing continues to decline exponentially. With the average cost of sequencing mammalian DNA hovering around $1,000 in the beginning of 2023, startups like Ultima Genomics and Illumina are working to decrease the cost to between $100-$200. That’s about the same as a new pair of Brooks running shoes! As the sequencing cost drops, the quantity of genetic data to study and analyze explodes, making it even more important to leverage machine learning techniques.

Why do we need DataOps Observability?

DevOps was started more than a decade ago as a movement, not a product or solution category. DevOps offered us a way of collaborating between development and operations teams, using automation and optimization practices to continually accelerate the release of code, measure everything, lower costs, and improve the quality of application delivery to meet customer needs.

ThoughtSpot and Databricks make governed, self-service analytics a reality with new Unity Catalog integration

Two years ago, we announced our Databricks partnership—including the launch of ThoughtSpot for Databricks, which gives joint customers the ability to run ThoughtSpot search queries directly on the Databricks Lakehouse without the need to move any data. Since then, we’ve empowered teams at companies like Johnson & Johnson, NASDAQ, and Flyr to safely self-serve business-critical insights on governed and reliable data.

Talend in 2023: Customer-focused, customer-forward

Prepping for a sales kick-off puts the focus on an organization’s customers unlike any other event. There is no selling — or success — without understanding and fulfilling the needs of our prospects and customers. As our gathering in Nashville last month proved, Talend is laser-focused on meeting customers wherever they are in their data journey, with everything they need to solve their most pressing data management challenges.

Data Talks Helps Sports Brands Score More Sales with Snowflake's Data Cloud

Data Talks fills a gap in the sports market by giving brands the data they need to truly understand their customers. But when your platform is responsible for continued ticket, merchandise, and sponsorship revenue, it needs to deliver insights fast. Here’s how Data Talks delivers data quickly with a rapid insights platform powered by Snowflake. From football fans filling out stadiums to baseball followers streaming every game, few customers are as avid and loyal as sports fans.

An Overview of Streaming Analytics in AWS for Logging Applications

Streaming analytics in AWS gives enterprises the ability to process and analyze log data in real time, enabling use cases that range from delivering personalized customer experiences to anomaly and fraud detection, application troubleshooting, and user behavior analysis. In the past, real-time log analytics solutions could process just a few thousand records per second and it would still take minutes or hours to process the data and get answers.

Maximize Business Results with FinOps

As organizations run more data applications and pipelines in the cloud, they look for ways to avoid the hidden costs of cloud adoption and migration. Teams seek to maximize business results through cost visibility, forecast accuracy, and financial predictability. Watch the breakout session video from Data Teams Summit and see how organizations apply agile and lean principles using the FinOps framework to boost efficiency, productivity, and innovation. Transcript available below.

Enabling Strong Engineering Practices at Maersk

As DataOps moves along the maturity curve, many organizations are deciphering how to best balance the success of running critical jobs with optimized time and cost governance. Watch the fireside chat from Data Teams Summit where Mark Sear, Head of Data Platform Optimization for Maersk, shares how his team is driving towards enabling strong engineering practices, design tenets, and culture at one of the largest shipping and logistics companies in the world.

Getting Up to Speed on Snowpark for Python with Educational Services

In today's livestream, Evan Troyka and Melanie Klein will introduce the 1-day Snowpark DataFrame Programming course on Snowflake. This 1-day course covers concepts, features, and programming constructs intended for practitioners building DataFrame data solutions in Snowflake.

Spark Technical Debt Deep Dive

Once in a while I stumble upon Spark code that looks like it has been written by a Java developer and it never fails to make me wince because it is a missed opportunity to write elegant and efficient code: it is verbose, difficult to read, and full of distributed processing anti-patterns. One such occurrence happened a few weeks ago when one of my colleagues was trying to make some churn analysis code downloaded from GitHub work.

Revamping Data Management Strategies with Data Pipelines

1. Data pipelines can improve data management strategies by enabling quick and easy data flow, transformation, and analysis. 2. Considerations when building a data pipeline include real-time data ingestion, scalability, performance optimization, data security and governance, and support for multiple sources. 3. Data mesh is a decentralized data architecture that organizes data sources by their specific business domains and must comply with the principles of the architecture. 5.

How Retailers Modernize Operations and Reporting with the Snowflake Retail Data Cloud

A modern data infrastructure is essential for retailers looking to stay competitive today. Companies are abandoning more traditional, on-premises IT infrastructures and moving to more centralized “as a service” (XaaS) models of delivery enabled by cloud technologies, according to McKinsey. Aging on-premises infrastructures are unable to meet demands for agility and innovation, eating up too much time and too many resources for teams trying to maintain them.

Learn How Snowflake's Own IT Department Built a Solution to Optimize Software Licenses

In this episode of “Data Cloud Now,” host Ryan Green sits down with Snowflake CIO and CDO Sunny Bedi to discuss how, in today’s uncertain economy, organizations use data and machine learning to maximize operational efficiency and control costs. Sunny and Ryan discuss how to use data and applications to optimize provisioning employees with the right technical resources and software. Snowflake itself has recently developed SnowPatrol — an internal application to analyze, predict, and optimize software spend.

How Banks are Using Technologies to Help Underserved Communities

Financial inclusion, defined as the availability and accessibility of financial services to underserved communities, is a critical issue facing the banking industry today. According to the World Bank, 1.7 billion adults around the world do not have access to formal financial services, meaning that they cannot open a bank account or access credit, insurance, or other financial products.

How Iodine Uses ClearML to Enable Better Healthcare Delivery

Improving patient care is right up there with the importance of optimizing the allocation and efficiency of precious resources when it comes to today’s healthcare. But that’s been difficult for people alone to accomplish, even with automation. The good news is that machine learning is now addressing these challenges and a company called Iodine Software is leading the innovation.

What is data mining and what are the best techniques to follow?

The most successful organizations today know they need to use business analytics to make decisions and drive outcomes. Often, however, these decisions must be driven by insights that can remain hidden in data. That’s where data mining comes into play. Data mining is a powerful tool to help extract meaningful insights from even the largest, most complex data sets.

BUILD Tel Aviv Panel: Bright Data and Wand.ai Discuss the Journey of Building Products on Snowflake

At Snowflake's biggest Data Cloud Developer Summit in Tel Aviv, lead Snowflake evangelist Eva Murray interviewed Or Lenchner, CEO of Bright Data, and Rotem Alaluf, CEO of Wand.ai, to get their thoughts about their experience as start-up companies building products on Snowflake. #Snowflake #DataCloud

7 Important Capabilities for Data Observability

Organizations need to manage data across ecosystems, develop data pipelines, APIs, insight into their metadata, and try to make sure that silos and data quality issues are managed effectively. Enter data observability platforms. This blog post looks at what drives many organizations to adopt data observability to ensure the health of your data across systems and providers.

Building Custom ITSM Dashboards for BMC Remedy

The importance of timely and accurate IT insights is increasing rapidly in the modern era. Organizations often desire a customizable solution to meet the specific needs of their business best and increasingly want to mix data from multiple sources and match data to sales, HR, asset management and other sources in order to better understand root causes and drive better decision-making. To do so, they need a solution that is able to be flexibly tailored for all use cases.

Effective Business Intelligence Application Testing Techniques

Software testing techniques help us ensure that our software meets all the requirements. The goal of testing techniques is to find out missing requirements, gaps, and errors in comparison to the actual requirements. Finding the right testing technique can be a little challenging. These techniques help us to identify test conditions. Note that there are multiple types of testing. One example is the black box testing technique. In this process, developers test their applications through different inputs.

Demystifying BigQuery BI Engine

BigQuery BI Engine is a fast, in-memory analysis system for BigQuery currently processing over 2 billion queries per month and growing. BigQuery has its roots in Google's Dremel system and is a data warehouse built with scalability as a goal. On the other hand BI Engine was envisioned with data analysts in mind and focuses on providing value on Gigabyte to sub-Terabyte datasets, with minimal tuning, for real time analytics and BI purposes.

Taming Cloud Costs for Data Analytics with FinOps

Uncontrolled cloud costs pose an enormous risk for any organization. The longer these costs go ungoverned, the greater your risk. Volatile, unforeseen expenses eat into profits. Budgets become unstable. Waste and inefficiency go unchecked. Making strategic decisions becomes difficult, if not impossible. Uncertainty reigns.

My 2023 Predictions for Chief Data Officers

The role of the CDO will not suffer the slow rise to prominence other emerging C-level roles have seen in previous years. Modern technologies designed to easily centralize, access, analyze, share, and monetize data have arrived. As a result, data has become the most powerful resource to drive an organization’s transformation. That puts CDOs at the wheel, their foot on the pedal, and their fellow execs holding on tight for the ride of their lives.

40 Characteristics of a Successful Product Manager

If you currently work as a product manager, are thinking about a job in product management, or work with product managers, one thing is certain: you will hear a different definition of what a ‘product manager’ is (or does) from each person you ask. Is the product manager the “CEO of the product” (which is a broad and misunderstood definition), the person who manages the product (a much narrower view of product management), or both?

What Are the Top 10 SQL Server Data Tools for 2023?

Five things you need to know about SQL Server data tools: Microsoft SQL Server is an RDBMS (relational database management system) for managing and storing information, transaction processing, and business intelligence. Managing a database on SQL Server or moving data to and from the platform can be challenging at the best of times. However, various SQL Server data tools can streamline these processes and improve scalability, functionality, and features.

Extending BMC Smart Reporting to Yellowfin BI: Webinar Recap

Welcome to the BMC Smart Reporting to Yellowfin BI webinar recap! Following our recent webinar for BMC customers held on January 18, 2023, this blog provides a helpful summary for those who could not attend or who would like a quick text recap on the most important points. Speakers at the event included: Missed the event? You can watch the webinar on-demand for free at your convenience. If you prefer to read the highlights, we've got you covered in our handy recap - read on.

Cybersyn Brings Exciting New Data Listings to the Snowflake Marketplace

In this episode of “Data Cloud Now,” Cybersyn Founder Alex Izydorczyk chats with DCN host Ryan Green about his passion for using data to drive business decisions and how the formation of the Snowflake Marketplace has dramatically enhanced data usage and data sharing opportunities. It’s leading to what he calls the “Moneyball-ization” of every industry. Tune in to hear him elaborate on that concept and to learn how Cybersyn itself is using publicly available sources of economic and demographic data to create new data sets in the Marketplace that hold valuable potential for almost any business.