20-09-2017 Door: Barry Devlin

Building a Data Warehouse

Deel dit bericht

Over this series of four posts, I explore the keys to a successful data warehouse. Last time, I started with design - a reasonable place to begin! The topic of this post is build, with operation and maintenance to follow.

Even with a beautiful design model in your mind’s eye, the question of how to build the data warehouse raises its ugly head! Ugly because no matter how lovely the model, implementation is always hobbled by the less than perfect reality of the data source systems. In the words of an old Irish joke in reply to a request for directions: “if I wanted to go there, I wouldn’t start from here.” Since the earliest days, builders of data warehouses have struggled with missing data in source systems, poorly defined data structures, incorrect content, and missing relationships, to name but a few. Implementation, therefore, becomes a delicate balancing act between the vision of the model and the constraints of the sources.

In simplistic terms, the process comes down to the following steps:

1. Understand the data sources: often described as data archeology, this step presents major challenges, especially for legacy systems, which—even if originally well documented—have usually been “bent to fit” emerging and urgent requirements. Modern big data sources may be equally challenging as a result of poor or absent documentation.
2. Compare the data available to the data warehouse model and define appropriate transformations to convert the former to the latter.
3. Where transformations are too difficult, modify the data warehouse model to accommodate the reality of the data sources. Changing the data sources—which would be the right answer when they are in error—is usually impossible for reasons of cost, politics, or both.
4. Test performance of load/update processes and check ability of modified model to deliver the data needed by the business.
5. If successful, declare victory. Otherwise, rinse and repeat.

Traditionally, the output of the above process would be encoded in a script or program and run—typically overnight in batch—to populate the warehouse. Any changes in requirements or, more problematically, in the source systems (beyond the control of the data warehouse developers) required a round trip back through steps 1 to 5, followed by code update. The approach is manual, time-consuming, and error-prone.

The solution over the years has been to automate the process in a series of approaches: ETL (extract, transform, load) tools, data integration systems, and latterly, data warehouse automation (DWA). In essence, each step on this journey depicts an increasing level of automation, with DWA designed to address the entire process of design, build, operation, and maintenance.

In the transition from design to build, the combination of a well-structured data model and a DWA tool such as WhereScape RED offers a particularly powerful approach to automation. This is because the data model provides an integrated starting set of metadata that describes the target tables in both business terms and technical implementation. This is particularly true in case of the Data Vault model, which has been designed and optimized from the start for data warehousing.

Consider, for example, the business need to analyze orders by value and geographical source. To the business person, order seems a simple, straightforward concept. In modeling terms, of course, it consists of a rather complex combination of entities, including product and person/customer. The structure to be built is equally intricate in terms of tables and the relationships between to them. The Data Vault model provides a database template for that structure, mapping directly from the business entities to a best practice set of data elements—from tables and columns through to relationships to indexes.

A DWA tool automates the transformation of the data structures of the various sources to the optimized model of the Data Vault and populates the target tables with the appropriate data, creating necessary indexes, and cleansing and combining source data to create the basis for the analysis needed by the business. WhereScape Data Vault Express provides the underlying templates to automatically and quickly build all the required structures (tables, indexes, etc.) and processes (ETL code) without manual programming and optimized for the chosen implementation platform, such as Teradata, Oracle, Microsoft, etc.

But, it’s about more than automating programming. In the future, Data Vault Express plans to address further build-time elements, including the methodology and best delivery practices defined by the Data Vault community, to avoid design errors and support proper auditing and management of the warehouse environment. That leads us to part three of this series.

This blog is part of a series and has been published on WhereScape.com

See part 1.

Barry Devlin

Dr. Barry Devlin behoort tot de autoriteiten op het gebied van business insight en is een van de grondleggers van datawarehousing. Met meer dan 30 jaar ervaring in IT, waarvan 20 jaar bij IBM als "Distinguished Engineer", is hij een alom gewaardeerd consultant en docent, en auteur van het standaardwerk “Data Warehouse – from Architecture to Implementation.” Zijn meest recente boek “Business unIntelligence: Insight and Innovation Beyond Analytics and Big Data” verscheen in 2013. Hij is oprichter van 9sight Consulting, gespecialiseerd in de menselijke, organisatorische en IT-implicaties, en het ontwerp van oplossingen voor diepgaande inzichten in de business. Barry sprak op ons jaarlijkse congres DW&BI Summit

Alle blogs van deze auteur

Partners