Improve Data Quality With Sandbox Analytics

Originally published March 16, 2020. Updated March 29, 2024
Søren Block Olsen
Søren Block Olsen
3 min read
TopBlogImage

Experimenting with new data in smaller groups is the key to a successful bimodal business intelligence (BI) strategy and high data quality. To err is human. That applies to love, history, politics…and data management.

When integrating data in an ERP system or other back-end data source, there are myriad ways data can be purposefully or accidentally manipulated, outdated, or cultivated in incongruous formats. As a result, any report or analysis delivered by a company’s BI solution may not reflect the realities of the company, its customers, or the industry in which it’s operating.

 

What's Your Mission?

Data quality is mission-critical for a company’s BI strategy. An analysis is only as good as the data behind it. When implementing a BI strategy, data must be accurate, authentic, and trustworthy. Standards of data quality may vary across verticals, but the integrity of the information must be able to stand solidly.

In recent years, BI has become increasingly operational, playing an important role in daily tactical and strategic decision-making. Decisionmakers can’t act unless they know they’re using reliable information to make those decisions.

 

Do More With Bimodal BI

Today, businesses are outgrowing the silos that were previously the status quo within the company. In years past, it was easy enough for employees in particular departments to wrap their heads around the data that was relevant to their role. Now, that information is shifting externally; it’s touching connections that have previously been undiscovered in daily business life.

As new and strengthening trends like big data take over the BI environment, companies must learn how to handle the unstructured, semi-structured, incomplete, and massive amounts of data that are suddenly at their fingertips. A deliberate, strong, and strategic method for dealing with data security and quality is more important than ever.

A bimodal BI solution tackles both the necessity of agile, user-friendly analytics and the reliability and security of data. A bimodal BI strategy should not only facilitate traditional business operations — the classic data warehouse and continuous decision loops — but also discovery and innovation. Bimodal BI is both centralized (company-wide initiatives) and decentralized (change, innovation, and exploration). A proper BI strategy is designed with flexibility and mass distribution in mind.

 

Enter: Sandbox Analytics

But not all data — no matter how potentially useful — should immediately be shared company- or even department-wide until the right experimentation, finessing, and cleansing for quality has been performed.

That’s why I suggest making “sandbox analytics” a part of your BI practice. In other words, break up small, isolated groups to produce, experiment with, and share data before considering wider adoption. These groups should be intra-department to help better shape how this data would affect the greater company.

For example, the marketing department will need to share information with the Sales department. Start with a smaller group from both departments when incorporating new and experimental data. With the right BI tool, there’s no need to compromise who gets what information when. Either/or is a thing of the past; it’s about embracing the “and.” Marketing and Sales. Sales and Finance.

The ability to play with big data sets and analyze them on top of what’s already in the data warehouse encourages employees to think strategically without the need to pull in IT, a must when employing a bimodal BI strategy. And a proper data governance model enables powerful sandbox analytics.

 

The Data Cleansing Strategy

But what about the quality of the data being experimented with? As new data is pulled into the BI system to be analyzed, it likely won’t mesh seamlessly with current ERP system data without first going through a data cleansing.

However, not all data needs to immediately meet industry or company standards, especially during the period in which users are determining whether it’s worthwhile to fully integrate it into the BI strategy. Cleansing data can be a massive undertaking and should be reserved for data that’s guaranteed to make a positive impact on the BI strategy.

Because of this, some BI consumers are not comfortable with data that’s not entirely standardized and should be left out of these sandbox security models. When the “sand castle” looks good, that is the time to bring the data quality up to 100 percent and display the results to the rest of the teams.

The trick is to gradually increase the quality of data along the way as it proves to be useful to the organization. This strategy ensures short-term ROI compared to what would typically be a massive data-cleansing project of new data.

It is, however, a necessary project. Without high-quality, standardized data, it’s impossible to create reports and analyses that can be trusted or that show the whole truth of a business scenario. Customer data, for example, doesn’t work if there are different versions of that customer within an ERP system.

I commonly see BI users who are frustrated with their inability to use data effectively. To actually use the data, they need to understand the process of connecting, cleaning, and correlating it. Once they fix or establish this process, they can focus on the KPIs, knowing they can trust the data.

Get a little dirty playing in the sand, then clean up what matters most to see the full potential of your BI strategy. Next step: Learn how to incorporate the most important KPIs into your company's data strategy.

Ready to Become a Data-Driven Organization?

DOWNLOAD GUIDE
Originally published November 24, 2015. Updated March 29, 2024

Søren Block Olsen
Written by

Søren Block Olsen

Director of Marketing & Sales Operations