My first blog post on Social PIM (Social Product Information Management) was over 4 years ago.
Since then Product Data Lake has been launched. Product Data Lake resembles a social network as you connect with your trading partners from the real world in order to collaborate on getting complete and accurate product information from the manufacturer to the point-of-sales.
I would love to see you, my blog readers, become involved. The options are:
The 2016 Magic Quadrant for Data Quality Tools by Gartner is out. One way to have a free read is downloading the report from Informatica, who is the most-top-right vendor in the tool vendor positioning.
Apart from the vendor positioning the report as always contains valuable opinions and observations about the market and how these tools are used to achieve business objectives.
Interenterprise data sharing is the last mentioned scenario besides BI and analytics (analytical scenarios), MDM (operational scenarios), information governance programs, ongoing operations and data migrations.
Another observation is that 90% of the reference customers surveyed for this Magic Quadrant consider party data a priority while the percentage of respondents prioritizing the product data domain was 47%.
My take on this difference is that it relates to interenterprise data sharing. Parties are per definition external to you and if your count of business partners (and B2C customers) exceeds some thousands (that’s the 90%), you need some of kind of tool to cope with data quality for the master data involved. If your product data are internal to you, you can manage data quality without profiling, parsing, matching and other core capabilities of a data quality tool. If your product data are part of a cross company supply chain, and your count of products exceeds some thousands (that’s the 47%), you probably have issues with product data quality.
In my eyes, the capabilities of a data quality tool will also have to be balanced differently for product data as examined in the post Multi-Domain MDM and Data Quality Dimensions.
This blog post revolves around how Master Data Management (MDM) and Product Information Management (PIM) can be the foundation of a better shopping experience and how to do this within driving digital transformation, being agile, and streamlining internal and external collaboration and workflows.
I agree with that. My only concern around the means mentioned relates to the section about how great customer experience starts with great supplier product data. The proposed approach for that is a self-service supplier data portal.
From what I have experienced, the concept of a supplier data portal for product data has limited chances of success. The problem for you as retailer or other form of downstream trading partner is your supplier. They will eventually have to deal with hundreds of supplier portals with different format and structure by the choice of their downstream trading partners, whereof you are just one. If you are a big one to them, it might work. Else probably not.
In the same way, your supplier could offer their customer data portal, build with their choice of format and structure. If they are a big one to you, you might go with that. Else, you probably would object to dealing with hundreds of different upstream data portals for you to go-to.
My Christmas present to you – suppliers, retailers, other supply chain nodes / PIM-MDM solution vendors – is a free trial / ambassadorship on Product Data Lake.
Product Data Lake is a cloud service for sharing product data in business ecosystems. Product Data Lake ensures:
Completeness of product information by enabling trading partners to exchange product data in a uniform way
Timeliness of product information by connecting trading partners in a process driven way
Conformity of product information by encompassing various international standards for product information
Consistency of product information by allowing upstream trading partners and downstream trading partners to interact with in-house structure of product information
Accuracy of product information by ensuring transparency of product information across the supply chain
Within Product Information Management (PIM) there is a growing awareness about that sharing product information between trading partners is a very important issue.
So, how do we do that? We could do that, on a global scale, by using:
1,234,567,890 spreadsheets
2,345,678 customer data portals
901,234 supplier data portals
Spreadsheets is the most common mean to exchange product information between trading partners today. The typical scenario is that a receiver of product information, being a downstream distributor, retailer or large end user, will have a spreadsheet for each product group that is sent to be filled by each supplier each time a new range of products is to be on-boarded (and potentially each time you need a new piece of information). As a provider of product information, being a manufacturer or upstream distributor, you will receive a different spreadsheet to be filled from each trading partner each time you are to deliver a new range of products (and potentially each time they need a new piece of information).
Customer data portals is a concept a provider of product information may have, plan to have or dream about. The idea is that each downstream trading partner can go to your customer data portal, structured in your way and format, when they need product information from you. Your trading partner will then only have to deal with your customer data portal – and the 1,234 other customer data portals in their supplier range.
Supplier data portals is a concept a receiver of product information may have, plan to have or dream about. The idea is that each upstream trading partner can go to your supplier data portal, structured in your way and format, when they have to deliver product information to you. Your trading partner will then only have to deal with your supplier data portal – and the 567 other supplier data portals in their business-to-business customer range.
Product Data Lake is the sound alternative to the above options. Hailstorms of spreadsheets does not work. If everyone has either a passive customer data portal or a passive supplier data portal, no one will exchange anything. The solution is that you as a provider of product information will push your data in your structure and format into Product Data Lake each time you have a new product or a new piece of product information. As a receiver you will set up pull requests, that will give you data in your structure and format each time you have a new range of products, need a new piece of information or each time your trading partner has a new piece of information.
Master Data Management (MDM) is increasingly being about supporting systems of engagement in addition to the traditional role of supporting systems of record. This topic was first examined on this blog back in 2012 in the post called Social MDM and Systems of Engagement.
The best known systems of engagement are social networks where the leaders are Facebook for engagement with persons in the private sphere and LinkedIn for engagement with people working in or for one or several companies.
But what about engagement between companies? Though you can argue that all (soft) engagement is neither business-to-consumer (B2C) nor business-to-business (B2B) but human-to-human (H2H), there are some hard engagement going on between companies.
One of the most important ones is exchange of product information between manufacturers, distributors, resellers and large end users of product information. And that is not going very well today. Either it is based on fluffy emailing of spreadsheets or using rigid data pools and portals. So there are definitely room for improvement here.
At Product Data Lake we have introduced a system of engagement for companies when it comes to the crucial task of exchanging product information between trading partners. Read more about that in the post What a PIM-2-PIM Solution Looks Like.
When looking out of the windows from Product Data Lake global headquarters (well, that is also our home office) we see our neighbour, which is the global headquarters of Maersk, a major worldwide operating shipping company.
In all humbleness we do very parallel business. Maersk is good at moving goods. We are going to move data about the goods. Product data or product information if you like.
The reason of being for a shipping company is that it would be very ineffective for each manufacturer of goods, if they should arrange and carry out the transportation of their manufactured goods to each distributor around the world. Furthermore, it would be equally ineffective, if each distributor should arrange and carry out the transportation of their range of goods to each reseller or large end buyer.
Until now, this ineffectiveness has unfortunately been the case when it comes to exchanging data about the goods. Manufacturers are asked by their distributors to provide product information in a different way for each – most often meaning in a different spreadsheet. And the same craziness repeats itself when it comes to exchanging data between distributors, resellers and large end users of product information.
This question was raised on this blog back in January this year in the post Tough Questions About MDM.
Since then the use of the term blockchain has been used more and more in general and related to Master Data Management (MDM). As you know, we love new fancy terms in our else boring industry.
However, there are good reasons to consider using the blockchain approach when it comes to master data. A blockchain approach can be coined as centralized consensus, which can be seen as opposite to centralized registry. After the MDM discipline has been around for more than a decade, most practitioners agree that the single source of truth is not practically achievable within a given organization of a certain size. Moreover, in the age of business ecosystems, it will be even harder to achieve that between trading partners.
This way of thinking is at the backbone of the MDM venture called Product Data Lake I’m working with right now. Yes, we love buzzwords. As if cloud computing, social network thinking, big data architecture and preparing for Internet of Things wasn’t enough, we can add blockchain approach as a predicate too.
In Product Data Lake this approach is used to establish consensus about the information and digital assets related to a given product and each instance of that product (physical asset or thing) where it makes sense. If you are interested in how that develops, why not follow Product Data Lake on LinkedIn.
The importance of having a viable Product Information Management (PIM) solution has become well understood for companies who participates in supply chains.
The next step towards excellence in PIM is to handle product information exchange (product data syndication) in close collaboration with your trading partners. Product Data Lake is the solution for that. Here upstream providers of product information (manufacturers and upstream distributors) and downstream receivers of product information (downstream distributors and retailers) connect their choice of in-house PIM solution or other product master data solution as PLM (Product Lifecycle Management) or ERP.
The PIM-2-PIM solution resembles a social network where you request and accept partnerships with your trading partners from the real world.
After connecting the next to set up is how your product attributes and digital asset types links with the one used by your trading partner. In Product Data Lake we encompass the use of these different scenarios (in prioritized order):
You and your trading partner uses the same standard in the same version
You and your trading partners uses the same standard in different versions
You and your trading partner uses different standards
You and/or your trading partners don’t use a public standard
Then it is time to link your common products. This can be done automatically if you both use a GTIN (or the older implementations as EAN number or UPC) as explained in the post Connecting Product Information. Alternatively, model numbers can be used for matching or, as a last option, the linking can be done in the interactive user interface.
Now you and your trading partner are set to start automating the process of sharing product information. In Product Data Lake upstream providers of product information can push new products, attribute values and digital assets from the in-house PIM solution to a hot folder, where from the information is uploaded by Product Data Lake. Downstream receivers can set up pull requests, where the linked product information is downloaded, so it is ready to be consumed by the in-house PIM solution.
This process can now be repeated with all your other trading partners, where you reuse the elements that are common between trading partners and build new linking where required.
If you have any questions, please contact me here:
In my eyes, this trend will have a huge impact on how data management platforms should be delivered in the future. Until now much of the methodology and technology for data management platforms have been limited to how these things are handled within the corporate walls. We will need a new breed of data management platforms build for business ecosystems.
Such platforms will have the characteristics of other new approaches to handling data. They will resemble social networks where you request and accept connections. They will embrace data as big data and data lakes, where every purpose of data consumption are not cut in stone before collecting data. These platforms will predominately be based in the cloud.
Right now I am working with putting such a data management service up in the cloud. The aim is to support product data sharing for business ecosystems. I will welcome you, and your trading partners, as subscriber to the service. If you help trading partners with Product Information Management (PIM) there is a place for you as ambassador. Anyway, please start with following Product Data Lake on LinkedIn.
Product Data Lake went live last month. Nevertheless, we are already planning the next big things in this cloud service for sharing product data. One of them is exactly things. Let me explain.
Product data is usually data about a product model, for example a certain brand and model of a pair of jeans, a certain brand and model of a drilling machine or a certain brand and model of a refrigerator. Handling product data on the model level within business ecosystems is hard enough and the initial reason of being for Product Data Lake.
However, we are increasingly required to handle data about each instance of a product model. Some use cases I have come across are:
Serialization, which is numbering and tracking of each physical product. We know that from having a serial number on our laptops and another example is how medicine packs now will be required to be serialized to prevent fraud as described in the post Spectre vs James Bond and the Unique Product Identifier.
Asset management. Asset is kind of the fourth domain in Master Data Management (MDM) besides party, product and location as touched in the post Where is the Asset. Also Gartner, the analyst firm, usually in theory (and also soon in practice in their magic quadrants) classifies product and asset together as thing opposite to party. Anyway, in asset management you handle each physical instance of the product model.
Internet of Things (IoT) is, according to Wikipedia, the internetworking of physical devices, vehicles (also referred to as “connected devices” and “smart devices”), buildings and other items—embedded with electronics, software, sensors, actuators, and network connectivity that enable these objects to collect and exchange data.
Fulfilling the promise of IoT, and the connected term Industry 4.0, certainly requires common understood master data from the product model over serialization and asset management as reported in the post Data Quality 3.0 as a stepping-stone on the path to Industry 4.0.