Considering success factors for data warehouse consolidation

0

The last thing you want to result from a data warehouse consolidation project is the creation of yet another siloed data asset that must be populated and managed with respect to the requirements of the downstream users.

To really benefit from a consolidation project, your newly-created consolidated warehouse should replace the data marts and warehouses that are used to populate it, and those systems should be scheduled for retirement. That said, it is worth suggesting some success criteria to specifically address the types of issues I raised in my previous post. Some of those success factors include:

  • Inclusiveness of the target data model: An inclusive data model must accommodate the union of the data attributes that exist across the data warehouses targeted for consolidation.
  • Elimination of duplicates: Eliminating duplicated records helps to reduce the storage footprint and will improve the degree of coherence for downstream applications.
  • Consistency in deduplication: It is extremely important to ensure consistency when attempting to merge a pair of records that is perceived to represent the same real-world entity.
  • Data validation and standardization: One of the failures associated with the proliferation of data warehouses is inconsistent application of data validation and standardization rules to data that is delivered to the data warehouse, and any consolidation effort must ensure consistency in applying data quality rules.
  • Synchronization: Issues with currency and timeliness of data delivery to different data warehouses should be resolved as a result of the consolidation.
  • Semantic consistency: This is a very common issue that is frequently ignored, and it often creates inconsistencies that can impact all downstream users. Similarly-named data elements in different data models are often presumed to mean the same thing. But subtle definition differences may lead to more obtuse inconsistencies when data sets are merged. The consolidation effort must employ semantic metadata and insist of strict semantic consistency as part of the consolidation process.

It is valuable to clearly define the expectations for data warehouse consolidation using criteria such as these, and then make sure that the right processes and tools in place to comply with the defined success criteria.

Tags
Share

About Author

David Loshin

President, Knowledge Integrity, Inc.

David Loshin, president of Knowledge Integrity, Inc., is a recognized thought leader and expert consultant in the areas of data quality, master data management and business intelligence. David is a prolific author regarding data management best practices, via the expert channel at b-eye-network.com and numerous books, white papers, and web seminars on a variety of data management best practices. His book, Business Intelligence: The Savvy Manager’s Guide (June 2003) has been hailed as a resource allowing readers to “gain an understanding of business intelligence, business management disciplines, data warehousing and how all of the pieces work together.” His book, Master Data Management, has been endorsed by data management industry leaders, and his valuable MDM insights can be reviewed at mdmbook.com . David is also the author of The Practitioner’s Guide to Data Quality Improvement. He can be reached at loshin@knowledge-integrity.com.

Leave A Reply

Back to Top