While COVID-19 continues to cause devastating disruption to the global economy more than two years into the pandemic, it is also continuing to force remarkable innovation across different industries. Companies have found new ways to sell, service and operate during the crisis. For me, there is one common theme for these innovative companies, including Qlik, and it is “Jobs to be Done.”
In an era where the passage of time seems to have changed somehow, it definitely feels strange to already be reflecting on another year gone by. It’s a cliche for a reason–the world definitely feels like it’s moving faster than ever, and in some completely unexpected directions. Sometimes it feels like we’re living in a time lapse when I consider the pace of technological progress I’ve witnessed in just a year.
Since I’m now migrating NodeGraph’s processes to Qlik, I thought it may be a good time to talk about migrating data during a merger or acquisition. There are many aspects to consider. Here are some of my thoughts on why companies merge or migrate data landscapes, common M&A migration pitfalls and how to avoid them, the time and cost involved migrating data during a merger or acquisition, and other topics.
Data is everywhere. As the sheer volume and number of data sources continue to explode, so do new opportunities for modern businesses to create and act on insights. That is if they are equipped with the right analytics technology. Historically, many businesses have settled for “good enough” analytics tools, putting up with lackluster bundles from full-stack vendors in an attempt to minimize cost or risk.
Since the release of Cloudera Data Engineering (CDE) more than a year ago, our number one goal was operationalizing Spark pipelines at scale with first class tooling designed to streamline automation and observability. In working with thousands of customers deploying Spark applications, we saw significant challenges with managing Spark as well as automating, delivering, and optimizing secure data pipelines.
“Water, water, everywhere, nor any drop to drink.” The famous line from Samuel Taylor Coleridge’s epic poem “The Rime of the Ancient Mariner” has a fitting application to today’s data problem. Enterprises are deluged with data, but they often have no way to leverage it. According to most experts, only a small percentage of data is usable and made useful, and most of it is in the dark — thus the term, “dark data.”
Out with the old; in with the new! If you haven’t already checked out the new Snowflake® interface (aka Snowsight®), make it your New Year’s resolution. Set yourself up for success in 2022 by spending a few minutes getting to know the new features and experiences that are in public preview—available when you click the Snowsight button at the top of your console’s menu bar.
With interest in big data and cloud increasing around the same time, it wasn’t long until big data began being deployed in the cloud. Big data comes with some challenges when deployed in traditional, on-premises settings. There’s significant operational complexity, and, worst of all, scaling deployments to meet the continued exponential growth of data is difficult, time-consuming, and costly.
In the BigQuery Spotlight series, we talked about Monitoring. This post focuses on using Audit Logs for deep dive monitoring. BigQuery Audit Logs are a collection of logs provided by Google Cloud that provide insight into operations related to your use of BigQuery. A wealth of information is available to you in the Audit Logs. Cloud Logging captures events which can show “who” performed “what” activity and “how” the system behaved.
The right set of tools helps businesses utilize data to drive insights and value. But balancing a strong layer of security and governance with easy access to data for all users is no easy task. Retrofitting existing solutions to ever-changing policy and security demands is one option. Another option — a more rewarding one — is to include centralized data management, security, and governance into data projects from the start.
For decades, the technologies and systems that deliver analytics have undergone massive change. What hasn’t changed, however, is the goal: using data-driven insights to drive actions. Insight to action has been a consistent vision for the industry. Everyone from data practitioners to technology developers have sought this elusive goal, but as Chief Data Strategy Officer Cindi Howson points out, it has remained unfulfilled — until now.
Enterprises using Microsoft Azure can now deliver secure, automated data pipelines while boosting compliance with privacy and industry mandates using Fivetran Business Critical.
Today we’re excited to announce the public beta launch of Continual, the first operational AI platform built specifically for modern data teams and the modern data stack. We’re also announcing our $4M Series Seed, led by Amplify Partners, and joined by Illuminate Ventures, Wayfinder, DCF, and Essence, as well as new partnerships with Snowflake and dbt Labs.
Data teams across companies have continuous challenges of consolidating data, processing it and making it useful. They deal with challenges such as a mixture of multiple ETL jobs, long ETL windows capacity-bound on-premise data warehouses and ever-increasing demands from users. They also need to make sure that the downstream requirements of ML, reporting and analytics are met with the data processing.
After a year packed with one machine learning and data science event after another, it’s clear that there are a few different definitions of the term ‘MLOps’ floating around. One convention uses MLOps to mean the cycle of training an AI model: preparing the data, evaluating, and training the model. This iterative or interactive model often includes AutoML capabilities, and what happens outside the scope of the trained model is not included in this definition.
The Cloudera Enterprise Data Maturity Report is a global survey of 3,150 business and IT decision makers assessing organizations’ maturity when it comes to their current capabilities and handling of data and analytics.
Stitch is pleased to announce the availability of Microsoft SQL Server as a destination. MS SQL Server joins nine other data destinations (including Microsoft Azure Synapse) that Stitch supports to help execute all your data modeling and analysis projects. Stitch customers can immediately benefit from the new destination, which supports both Azure SQL Server and standard SQL Server editions reaching as far back as SQL Server 2012.
About the report The Cloudera Enterprise Data Maturity Report is a global survey of 3,150 business and IT decision makers assessing organizations’ maturity when it comes to their current capabilities and handling of data and analytics.
The following is Part III of a three-part series. Welcome to the final installment of a three-part series discussing the areas to take seriously when you want to drive business with analytics. In Part I of this series, I discussed how to prioritize data accessibility and how to address the challenges that come with it. Those challenges include: Part II discussed where the disconnect is and addressed how organizations can bridge the gap.
Over the course of my next two blog posts, I would like to share my thoughts around a debate raging in data architecture circles. The bone of contention? That the 21st century needs a new data management paradigm for modern analytics. First up, I’ll frame the argument and explain the two prominent approaches of data hub and data fabric. Then, I’ll cover data mesh and compare all three architectures. As always, I’d love to get your input, feedback, queries and comments!
Today we’re pleased to announce Continual Integration for dbt. We believe this is a radical simplification of the machine learning (ML) process for users of dbt and presents a well-defined path that bridges the gap between data analytics and data science. Read on to learn more about this integration and how you can get started.
Artificial Intelligence (AI) has revolutionized how various industries operate in recent years. But with growing demands, there’s a more nuanced need for enterprise-scale machine learning solutions and better data management systems. The 2021 Data Impact Awards aim to honor organizations who have shown exemplary work in this area.
One of the most challenging and complex elements of operating a financial services institution is compliance. Managing risk, security and privacy to earn customers’ trust has long been at the core of financial services, but this foundation has been shaken over recent years.
One of the key cornerstones of the emerging field of ethical, explainable AI is recognizing and avoiding bias. As AI takes on a greater role in organizations with sometimes opaque calculations, there is an increased urgency in many businesses to get ahead of these challenges, and companies, such as IBM, Salesforce and Microsoft, have already added roles specifically with the aim of ensuring that ethics are a key consideration of AI.
In today’s tech-forward companies, CFOs are tasked with managing and overseeing an increasingly expansive domain of systems and technologies to thrive. The rise of regulatory considerations, novel market drivers and a globally connected business environment is creating an entirely new set of pressures on both the structure of the department and on leadership.
It’s easy to take continuous integration (CI) and continuous delivery/deployment (CD) for granted these days, but these have been transformational concepts that have drastically changed the face of software development over the past thirty years.
While CI/CD is synonymous with modern software development best practices, today’s machine learning (ML) practitioners still lack similar tools and workflows for operating the ML development lifecycle on a level on par with software engineers. For background, follow a brief history of transformational CI/CD concepts and how they’re missing from today’s ML development lifecycle.
Coming off of our Snowday event, we’ve unveiled a number of new product capabilities that expand what is possible in the Data Cloud. From helping businesses operate globally with improved replication efficiency, empowering developers with new functionality in Snowpark, and improving the security and governance of data through native object tagging, there is no shortage of exciting advancements coming to Snowflake.
Marketing teams routinely struggle with monitoring the performance and cost of their ad campaigns. Now, they have a solution that can be as easy as just a few clicks. We recently joined our partners at Rivery for a webinar demonstrating the new Anodot Markering Analytics Monitoring Kit. The kit allows users to track marketing campaigns in real-time and take the action needed to make the most of ad spend.
Hello, I’m Imogen Heap – musician and tech founder. Now, I know what you might be thinking. Why is a musician and composer talking to me about data? Well, for any of you that follow my work, you will know that I am a bit of a data nerd, and that’s why I loved speaking to Joe DosSantos on the latest Data Brilliant podcast. A lot of my music is powered by data. In fact, Joe and I discussed my love for how technology and data inspires me to be more creative. Want to know how?
Business analytics (BA) is the process of evaluating data in order to gauge business performance and to extract insights that may facilitate strategic planning. It aims to identify the factors that directly impact business performance, such as ie. revenue, user engagement, and technical availability. BA takes data from all business levels, from product and marketing, to operations and finance.
As organizations look to improve business operations and outcomes, global industries are pushing for data-driven transformation. The 2021 Cloudera Data Impact Awards recognize those organizations that have pulled ahead of the pack with efforts to leverage the power of data to improve operations and better serve their customers. The finalists in the “Industry Transformation” category are MTN, National Payments Corporation of India (NPCI), Sberbank, and Bank Negara Indonesia (BNI).
The Cloudera Enterprise Data Maturity Report is a global survey of 3,150 business and IT decision makers assessing organizations’ maturity when it comes to their current capabilities and handling of data and analytics.
CDP Operational Database (COD) is a real-time auto-scaling operational database powered by Apache HBase and Apache Phoenix. It is one of the main Data Services that runs on Cloudera Data Platform (CDP) Public Cloud. You can access COD right from your CDP console. With COD, application developers can now leverage the power of HBase and Phoenix without the overheads related to deployment and management.
One of the best features in ThoughtSpot Everywhere is the Developer Playground. The Playground lets frontend Developers visually configure elements and generate JavaScript code to add into your web app. It is an amazing tool for testing and iterating configuration options before adding final elements such as Search, Liveboards, and visualizations into your web app. But what about a backend Developer who might be building solutions that utilize the Platform’s APIs?
Snowpark has generated significant excitement and interest since it was announced. Snowpark is a developer framework that enables data engineers, data scientists, and data developers to code in their language of choice, and execute pipelines, machine learning (ML) workflows, and data applications faster and more securely. While many parts of Snowpark are in preview stages, External Functions entered General Availability earlier this year.
When looking at API analytics and monitoring platforms, many seem to be so similar that it’s hard to figure out the differences between them. We often hear this confusion from users and prospects. In a world with so many tools available, how do we figure out which ones are necessary and which are redundant? One of the most common questions we are asked revolves around how Moesif compares to Datadog and how they could work together.
It takes vision, purpose, and skill to unlock the power of data. It also takes the right strategy. For ExxonMobil, Ares Trading (Merck), and the University of California San Diego (UCSD), the right strategy is taking full advantage of the cloud. All three organizations have partnered with Cloudera, leveraging a hybrid or cloud-based architecture to improve the lives of the people who depend on their organizations’ data.
Organizations in the financial services sector face a unique set of challenges as they consider how to wrangle and process the vast amount of data they collect. During our Financial Services Summit, I was lucky enough to speak to Brian Anthony, chief data officer for the Municipal Securities Rulemaking Board (MSRB), to learn how the MSRB is integrating technologies such as artificial intelligence (AI) and machine learning to modernize its data.
Machine learning (ML) models have become key drivers in helping organizations reveal patterns and make predictions that drive value across the business. While extremely valuable, building and deploying these models remains in the hands of only a small subset of expert data scientists and engineers with deep programming and ML framework expertise.
At Snowflake, putting the customer first is an essential company value. But “customer-centric” is more than just a buzzword: We use a data-driven, outside-in lens on everything we do, at all levels of the company. In particular, here’s how Snowflake Support is listening to you—our customers—and continuously improving the Snowflake customer experience at every touchpoint.
The business world is rapidly pivoting all the time. Strategic shifts, reprioritization and being first all require being smart while moving fast. The value of agility has never stood out more due to the need to react to new realties in everything from public health, remote and in-office business policies and workflows, to broader economic concerns like supply chain as we move into recovery and revitalization.
In part 1 of this blog post, we discussed the need to be mindful of data bias and the resulting consequences when certain parameters are skewed. Surely there are ways to comb through the data to minimise the risks from spiralling out of control. We need to get to the root of the problem. In 2019, the Gradient institute published a white paper outlining the practical challenges for Ethical AI.
For Cloudera ensuring data security is critical because we have large customers in highly regulated industries like financial services and healthcare, where security is paramount. Also, for other industries like retail, telecom or public sector that deal with large amounts of customer data and operate multi-tenant environments, sometimes with end users who are outside of their company, securing all the data may be a very time intensive process.
Financial products are no longer characterized by the steps of filling out a form, waiting for a credit decision and, if successful, watching the monthly payments leaving your account.
There is an explosion of data from a myriad of sources and an insatiable demand to consume it. Traditional manual ETL methods are too brittle to keep up. Leaving many a BI team struggling to provide meaningful business insights quickly.
As customers grow their analytical workloads and footprint on BigQuery, their monitoring and management requirements evolve - they want to be able to manage their environments at scale, take action in context. They also desire capacity management capabilities to optimize their BigQuery environments. With our BigQuery Administrator Hub capabilities, customers can now better manage BigQuery at scale.
Without data, your business cannot survive. You need it to understand your customers, to inform your product development, and to plan the future of your company. So why would you tolerate technology that artificially throttles access to your own data and stunts your company’s growth? It just doesn't make sense. That’s why Talend is bringing you Stitch Unlimited and Stitch Unlimited Plus.