Top 15 MDM / PIM Requirements in RFPs

A Request for Proposal (RFP) process for a Master Data Management (MDM) and/or Product Information Management (PIM) solution has a hard fact side as well as there are The Soft Sides of MDM and PIM RFPs.

The hard fact side is the detailed requirements a potential vendor has to answer to in what in most cases is the excel sheet the buying organization has prepared – often with the extensive help from a consultancy.

Here are what I have seen as the most frequently included topics for the hard facts in such RFPs:

  • MDM and PIM: Does the solution have functionality for hierarchy management?
  • MDM and PIM: Does the solution have workflow management included?
  • MDM and PIM: Does the solution support versioning of master data / product information?
  • MDM and PIM: Does the solution allow to tailor the data model in a flexible way?
  • MDM and PIM: Does the solution handle master data / product information in multiple languages / character sets / script systems?
  • MDM and PIM: Does the solution have capabilities for (high speed) batch import / export and real-time integration (APIs)?
  • MDM and PIM: Does the solution have capabilities within data governance / data stewardship?
  • MDM and PIM: Does the solution integrate with “a specific application”? – most commonly SAP, MS CRM/ERPs, SalesForce?
  • MDM: Does the solution handle multiple domains, for example customer, vendor/supplier, employee, product and asset?
  • MDM: Does the solution provide data matching / deduplication functionality and formation of golden records?
  • MDM: Does the solution have integration with third-party data providers for example business directories (Dun & Bradstreet / National registries) and address verification services?
  • MDM: Does the solution underpin compliance rules as for example data privacy and data protection regulations as in GDPR / other regimes?
  • PIM: Does the solution support product classification and attribution standards as eClass, ETIM (or other industry specific / national standards)?
  • PIM: Does the solution support publishing to popular marketplaces (form of outgoing Product Data Syndication)?
  • PIM: Does the solution have a functionality to ease collection of product information from suppliers (incoming Product Data Syndication)?

Learn more about how I can help in the blog page about MDM / PIM Tool Selection Consultancy.

MDM PIM RFP Wordle

Human Errors and Data Quality

Every time there is a survey about what causes poor data quality the most ticked answer is human error. This is also the case in the Profisee 2019 State of Data Management Report where 58% of the respondents said that human error is among the most prevalent causes of poor data quality within their organization.

This topic was also examined some years ago in the post called The Internet of Things and the Fat-Finger Syndrome.

Errare humanum estEven the Romans new this as Seneca the Younger said that “errare humanum est” which translates to “to err is human”. He also added “but to persist in error is diabolical”.

So, how can we not persist in having human errors in data then? Here are three main approaches:

  • Better humans: There is a whip called Data Governance. In a data governance regime you define data policies and data standards. You build an organizational structure with a data governance council (or any better name), have data stewards and data custodians (or any better title). You set up a business glossary. And then you carry on with a data governance framework.
  • Machines: Robotic Processing Automation (RPA) has, besides operational efficiency, the advantage of that machines, unlike humans, do not make mistakes when they are tired and bored.
  • Data Sharing: Human errors typically occur when typing in data. However, most data are already typed in somewhere. Instead of retyping data, and thereby potentially introduce your misspelling or other mistake, you can connect to data that is already digitalized and validated. This is especially doable for master data as examined in the article about Master Data Share.

IoT and Business Ecosystem Wide MDM

Two of the disruptive trends in Master Data Management (MDM) are the intersection of Internet of Things (IoT) and MDM and business ecosystem wide MDM (aka multienterprise MDM).

These two trends will go hand in hand.

IoT and Ecosystem Wide MDM

The latest MDM market report from Forrester (the other analyst firm) was mentioned in the post Toward the Third Generation of MDM.

In here Forrester says: “As first-generation MDM technologies become outdated and less effective, improved second generation and third-generation features will dictate which providers lead the pack. Vendors that can provide internet-of-things (IoT) capabilities, ecosystem capabilities, and data context position themselves to successfully deliver added business value to their customers.”

This saying is close to me in my current job as co-founder and CTO at Product Data Lake as told in the post Adding Things to Product Data Lake.

In business ecosystem wide MDM business partners collaborate around master data. This is a prerequisite for handling asset master data involved in IoT as there are many parties involved included manufacturers of smart devices, operators of these devices, maintainers of the devices, owners of the devices and the data subjects these devices gather data about.

In the same way forward looking solution providers involved with MDM must collaborate as pondered in the post Linked Product Data Quality.

Artificial Intelligence (AI) and Multienterprise MDM

The previous post on this blog was called Machine Learning, Artificial Intelligence and Data Quality. In here the it was examined how Artificial Intelligence (AI) is impacted by data quality and how data quality can impact AI.

Master Data Management (MDM) will play a crucial role in sustaining the needed data quality for AI and with the rise of digital transformation encompassing business ecosystems we will also see an increasing need for ecosystem wide MDM – also called multienterprise MDM.

Right now, I am working with a service called Product Data Lake where we strive to utilize AI including using Machine Learning (ML) to understand and map data standards and exchange formats used within product information exchange between trading partners.

The challenge in this area is that we have many different classification systems in play as told in the post Five Product Classification Standards. Besides the industry and cross sector standards we still have many homegrown standards as well.

Some of these standards (as eClass and ETIM) also covers standards for the attributes needed for a given product classification, but still, we have plenty of homegrown standards (at no standards) for attribute requirements as well.

Add to that the different preferences for exchange methods and we got a chaotic system where human intervention makes Sisyphus look like a lucky man. Therefore, we have great expectations about introducing machine learning and artificial intelligence in this space.

AI ML PDL

Next week, I will elaborate on the multienterprise MDM and artificial theme on the Master Data Management Summit Europe in London.

Data Matching and Real-World Alignment

Data matching is a sub discipline within data quality management. Data matching is about establishing a link between data elements and entities, that does not have the same value, but are referring to the same real-world construct.

The most common scenario for data matching is deduplication of customer data records held across an enterprise. In this case we often see a gap between what we technically try to do and the desired business outcome from deduplication. In my experience, this misalignment has something to do with real-world alignment.

Data Matching and Real World Alignment

What we technically do is basically to find a similarity between data records that typically has been pre-processed with some form of standardization. This is often not enough.

Location Intelligence

Deduplication and other forms of data matching with customer master data revolves around names and addresses.

Standardization and verification of addresses is very common element in data quality / data matching tools. Often such at tool will use a service either from its same brand or a third-party service. Unfortunately, no single service is often enough. This is because:

  • Most services are biased towards a certain geography. They may for example be quite good for addresses in The United States but very poor compared to local services for other geographies. This is especially true for geographies with multiple languages in play as exemplified in the post The Art in Data Matching.
  • There is much more to an address than the postal format. In deduplication it is for example useful to know if the address is a single-family house or a high-rise building, a nursing home, a campus or other building with lots of units.
  • Timeliness of address reference data is underestimated. I recently heard from a leader in the Gartner Quadrant for Data Quality Tools that a quarterly refresh is fine. It is not, as told in the post Location Data Quality for MDM.

Identity Resolution

The overlaps and similarities between data matching and identity resolution was discussed in the post Deduplication vs Identity Resolution.

In summary, the capability to tell if two data records represent the same real-world entity will eventually involve identity resolution. And as this is very poorly supported by data quality tools around, we see that a lot of manual work will be involved if the business processes that relies on the data matching cannot tolerate too may, or in some cases any, false positives – or false negatives.

Hierarchy Management

Even telling that a true positive match is true in all circumstances is hard. The predominant examples of this challenge are:

  • Is a match between what seems to be an individual person and what seems to be the household where the person lives a true match?
  • Is a match between what seems to be a person in a private role and what seems to be the same person in a business role a true match? This is especially tricky with sole proprietors working from home like farmers, dentists, free lance consultants and more.
  • Is a match between two sister companies on the same address a true match? Or two departments within the same company?

We often realize that the answer to the questions are different depending on the business processes where the result of the data matching will be used.

The solution is not simple. The data matching functionality must, if we want automated and broadly usable results, be quite sophisticated in order to take advantage of what is available in the real-world. The data model where we hold the result of the data matching must be quite complex if we want to reflect the real-world.

The Recent Coupling on the MDM Market

When it has been about mergers and acquisitions on the Master Data Management (MDM) solution market, there have until recently not been so much going around since 2012. Rather we have seen people leaving the established vendors and formed or joined new companies.

But, three months ago Tibco was coupled with Orchestra.

Then on Valentine’s day 2019 Symphony Technology Group Acquired PIM and MDM Provider EnterWorks with the aim of coupling their offerings with the ones from WinShuttle. WinShuttle has been more a data management generalist company with focus on ERP data – not at least in SAP. This merger ties into the trend of extending MDM platforms to other kinds of data than traditional master data. It will also make an alternative to SAPs own MDM and data governance offering called MDG.

Fourteen days later there was a new coupling as reported in the post MDM Market News: Informatica acquires AllSight. This must also be seen as a step in the trend of providing an extended MDM platform with Artificial Intelligence (AI) capabilities. Also, Informatica is here going against the new MDM solution provider Reltio, who has been successful in promoting their big data extended MDM platform.

Both Enterworks and AllSight (and Reltio too) are listed on The Disruptive Master Data Management List.

MDM Coupling

 

1,000 Blog Posts and More to Come

number_1000I just realized that this post will be number 1,000 published on this blog. So, let me not say something new but just recap a little bit on what it has been all about in the last nearly 10 years of running a blog on some nerdy stuff.

Data quality has been the main theme. When writing about data quality one will not avoid touching Master Data Management (MDM). In fact, the most applied category used here on this site, with 464 and counting entries, is Master Data.

The second most applied category on this blog is, with 219 entries, Data Architecture.

The most applied data quality activity around is data matching. As this is also where I started my data quality venture, there has been 192 posts about Data Matching.

The newest category relates to Product Information Management (PIM) and is, with 20 posts at the moment, about Product Data Syndication.

Even though that data quality is a serious subject, you must not forget to have fun. 66 posts, including a yearly April Fools post, has been categorized as Supposed to be a Joke.

Thanks to all who are reading this blog and not least to all who from time to time takes time to make a comment, like and share.