Posted by Michele Goetz on September 7, 2012
We last spoke about how to reboot our thinking on master data to provide a more flexible and useful structure when working with big data. In the structured data world, having a model to work from provides comfort. However, there is an element of comfort and control that has to be given up with big data, and that is our definition and the underlying premise for data quality.
Current thinking: Persistence of cleansed data.For years data quality efforts have focused on finding and correcting bad data. We used the word “cleansing” to represent the removal of what we didn’t want, exterminating it like it was an infestation of bugs or rats. Knowing what your data is, what it should look like, and how to transform it into submission defined the data quality handbook. Whole practices were stood up to track data quality issues, establish workflows and teams to clean the data, and then reports were produced to show what was done. Accomplishment was the progress and maintenance of the number of duplicates, complete records, last update, conformance to standards, etc. Our reports may also be tied to our personal goals. Now comes big data — how do we cleanse and tame that beast?
Reboot: Disposability of data quality transformation. The answer to the above question is, maybe you don’t. The nature of big data doesn’t allow itself to traditional data quality practices. The volume may be too large for processing. The volatility and velocity of data change too frequently to manage. The variety of data, both in scale and visibility, is ambiguous.
Your data quality efforts need to be defined more as profiling and standards versus cleansing. This is better aligned to how big data is managed and processed. While on the surface, big data processing is batch in nature, it would seem obvious to institute data quality rules the way they have always been done. But the answer is to be more service-oriented, invoking data quality rules that provide improved standardization and sourcing during processing versus fundamentally changing the data. In addition, data quality rules are invoked in a customized fashion based on customer service calls from big data processing.
Why this also makes sense is that when you do decide to persist sourced big data into your internal infrastructure, you have pre-aligned the data to existing policies for integration and business rules for improved mapping and cleansing that would need to persist. In essence you treat big data as a reference source, not a primary source. When have you looked to persist your data quality rules on reference data from a third party?
So, think about data quality in the context of supporting preprocessing with Hadoop and MapR through profiling and standards, not cleansing.
Reboot: Data quality and acceptable risk
Search Forrester's Blogs
Save Money On Your Next Software Negotiation
Work with our software negotiation experts to save 10–20% on your next contract »
Lead BT Transformation
Develop customer-obsessed strategies to drive growth »
Forrester's CX Index
Predict how actions to improve CX will affect revenue performance.
Measure the customer experiences that matter most »
- Alex Cullen (42)
- Brian Hopkins (42)
- Charlie Dai (31)
- Cheryl McKinnon (11)
- Clay Richardson (42)
- Craig Le Clair (58)
- Diego Lo Giudice (1)
- Ellen Carney (1)
- Gene Leganza (24)
- Gordon Barnett (3)
- Henry Peyret (10)
- Leslie Owens (10)
- Michele Goetz (49)
- Pamela Heiligenthal (1)
- Sharyn Leaver (3)
- Skip Snow (2)