Lessons learned from a master data management implementation


It’s clear what the benefits are: a single, correct source of the truth—created and maintained through collaboration, stored centrally and shared everywhere. If having the right information at the right time is key to your business success, there is a lot of sense in your whole company having access to (and contributing to the currency of) your most important data. However, the realization of this vision can be difficult to achieve and require more than just implementing a software solution.

Master Data Management (MDM) software companies tend to be quite clear about what an organisation needs to implement or have in place to successfully leverage an MDM solution. Vendors state the need for data governance, defined business rules, trusted sources and universal provisioning. But what perhaps isn’t always so clear is the effort needed to implement and maintain these practices. Considerable thought must be given not only to how to implement an MDM solution, but also to how it’s going to live and breathe well into the future.

3gamma recently helped a large client implement an MDM solution. These are our key learning points from the project:

  1. Iteration is the best way to move forwards
  2. Don’t overestimate the quality of the existing data and recognize the value of humans in cleaning and maintaining it
  3. Beware of consumers driving your integration design
  4. You probably don’t need as much master data as you think you do.

1. Iteration is the best way to move forwards

Attempting to get a committee of stakeholders to agree on what data they want to store (and what data they can live without) takes time. It’s easy to get stuck debating all the possibilities but we need to move forward quickly and build interest and momentum. By taking an iterative approach we can leave the starting line faster and monitor further need in the real world. The solution then creates a compelling case for itself through continuous delivery and feedback. Planning for iterations and ensuring your solution and methodology can handle successive changes in a flexible way is more important than getting it all captured first time.

2. Don’t overestimate the quality of existing data and recognize the value of humans in cleaning and maintaining it

The project involved moving data from an existing set of disparate, non-data management systems (CRM tools, finance systems, data warehouses, spreadsheets etc.) as well as importing public data from commercial data providers into a single enterprise MDM system.

As data is often not correct in the source, you can’t just map from one field to another, and the MDM can’t easily clean the data by spotting errors and correcting it, because it can’t logically make the distinction between e.g. an address and someone’s middle name without a lot of code. There may also be discrepancies between source systems.

So how do we handle the conflict but crucially not lose any information?

Cleaning up data using a data mapping process and a robot to calculate which data is correct will be imperfect and complex. Using the MDM system to achieve this will probably not work well either. There are recent advances in the ability for robots to understand nonobvious relationships but the insight of a human in understanding how to clean data can sometimes be the most effective approach.

Therefore, consider placing the responsibility to clean and maintain the data on a “data steward”. An MDM solution should be seen as the tool to allow them to do that, not the tool that does that for them. This is about placing correct expectations on the technical solution and using the tech in the most appropriate way. Automate the easy stuff, don’t consider the data management as easy.

3. Beware of consumers driving your design

You want every system that has a need for master data to consume it from the MDM database. But there is often debate over who gets to mandate the integration and data standards. Asking consumers how they want to consume the master data means you’ll get many differing requirements. When designing how an MDM solution will integrate, get high level sponsorship stipulating that consuming systems will make parallel changes to fit the standard interface, not the other way around. This requires engagement early on with the consuming systems and a clear vision of how the integration will work so changes can be made in good time. It also requires a good understanding of the level of investment and change in an MDM implementation.

4. You probably don’t need as much master data as you think you do

First draft data models based on asking the business what they want to store can contain thousands of fields of information and many more pieces of related data fields. This calls into question what is master data and what is just data. Being clear and getting agreement on what is master data and what is associated information is important and relies on going back to what problem the MDM solution is trying to solve.

Often, stakeholders will be concerned about losing the ability to store data they someday may require, which will produce a large data model that is complex, difficult to maintain and has many empty fields. A better solution is to separate out master data and additional data, creating placeholders for relevant but not critical information but not including it in the core data set. Cutting out the need to store everything makes it easier to establish the match rules and drive to a core set of relevant information. Separate out the need to store the other stuff and do a separate project to store this is in a relational database.

To conclude, having a single source of the truth is an incredibly powerful asset. If you are on a data management journey or about to start on one, my key advice is not to consider it as solely an IT challenge but also an organisational one.

About the author

Matt Williamson is a senior IT management consultant at 3gamma, with 14 years of experience from a wide variety of leadership roles in IT programme delivery and IT service management.

Related Articles

Managing a successful service desk through the service management office

Governance, Operations

The service desk is the interface between the business and IT and is in many cases the only IT representative that an end user ever comes in contact…

Emerging technologies: A transformative force of the new digital economy

In the new digital economy, driven by emerging technologies transforming how business is being done, IT is moving from being a reactive cost-centre to become a proactive business partner. IT is no longer just about servers and networks – it’s about delivering customer value across multiple internal and external touch-points.

Understanding attitude is key to successful stakeholder management

Change Management, Risk Management

Projects lead to changes – something that many people instinctively are sceptical about or even hostile to. No matter how well planned a project is, there is always…

Managing Technology in an ever-changing environment: Using integration to create business and IT alignment

Innovation, Operations, Strategy & Architecture

Successful companies leverage technology integration to create a competitive advantage. Technology integration, strategically deployed and executed with discipline, can be an enabler of cost reductions, improved customer relationships and digital innovation.

Risky business: There is a time for playing it safe and a time for risky business

Change Management, Risk Management

All projects are exposed to varying degree of risk. To secure successful delivery, this risk must be appropriately managed. Proper risk management reduces the number of threats that could materialise into problems and minimises the effect of those that do occur. It also results in more opportunities being captured proactively and turned into positive benefits for the project.