I’ve been presenting research on big data and data governance for the past several months where I show a slide of a businesswoman doing a backbend to access data in her laptop. The point I make is that data management has to be hyper-flexible to meet a wider range of analytic and consumption demands than ever before. Translated, you need to cross-train for data management to have cross-fit data.
The challenge is that traditional data management takes a one-size fits-all approach. Data systems are purpose built. If organizations want to reuse a finance warehouse for marketing and sales purposes, it often isn’t a match and a new warehouse is built. If you want to get out of this cycle and go from data couch potato to data athlete, a cross-fit data training program should focus on:
Context first. Understanding how data is used and will provide value drives platform design. Context indicates more than where data is sourced from and where it will be delivered. Context answers: operations or analytics, structured or unstructured, persistent or disposable? These guide decisions around performance, scale, sourcing, cost, and governance.
Data governance zones. Command and control data governance creates a culture of “no” that stifles innovation and can cause the business to go around IT for data needs. The solution is to create policies and processes that give permission as well as mitigate risk. Loosen quality and security standards in projects and scenarios that are in contained environments. Tighten rules and create gates when called for by regulation, where there are ethical conflicts, or when data quality or access exposes the business to significant financial risk.
We recently attended Amdocs' customer event in Singapore. Amdocs is gradually adjusting its strategy to reflect one of the most fundamental changes in the ICT industry today: Increasingly, business line managers, think the marketing or sales officer, are the ones influencing sourcing decisions. Traditional decision-makers, CTOs and CIOs, are no longer the sole ICT decision-makers. Amdocs is addressing this shift by:
Strengthening its customer experience portfolio.Successful telcos will try to regain lost relevance through improved customer experience. Marketing, portfolio product development, and sales are therefore growing in importance for telcos. Amdocs’ integrated customer experience offering, CES 9, provides telcos with a multichannel experience; proactive care; and self-service tools.
Betting big on big data/analytics.Amdocs is leveraging big data/analytics to provide real-time, predictive, and prescriptive insights to telcos about their customers’ behaviour. Communications-industry-specific converged charging and billing solutions as well as other catalogue solutions give Amdocs the opportunity to provide more value to telcos than some of the other players.
I attended Google’s annual atmosphere road show recently, an event aimed at presenting solutions for business customers. The main points I took away were:
Google’s “mosaic” approach to portfolio development offers tremendous potential. Google has comprehensive offerings covering communications and collaboration solutions (Gmail, Google Plus), contextualized services (Maps, Compute Engine), application development (App Engine), discovery and archiving (Search, Vault), and access tools to information and entertainment (Nexus range, Chromebook/Chromebox).
Google’s approach to innovation sets an industry benchmark. Google is going for 10x innovation, rather than the typical industry approach of pursuing 10% incremental improvements. Compared with its peers, this “moonshot” approach is unorthodox. However, moonshot innovation constitutes a cornerstone of Google’s competitive advantage. It requires Google’s team to think outside established norms. One part of its innovation drive encourages staff to spend 20% of their work time outside their day-to-day tasks. Google is a rare species of company in that it does not see failure if experiments don’t work out. Google cuts the losses, looks at the lessons learned — and employees move on to new projects.
The Obama 2012 campaign famously used big data predictive analytics to influence individual voters. They hired more than 50 analytics experts, including data scientists, to predict which voters will be positively persuaded by political campaign contact such as a call, door knock, flyer, or TV ad. Uplift modeling (aka persuasion modeling) is one of the hottest forms of predictive analytics, for obvious reasons — most organizations wish to persuade people to to do something such as buy! In this special episode of Forrester TechnoPolitics, Mike interviews Eric Siegel, Ph.D., author of Predictive Analytics, to find out: 1) What exactly is uplift modeling? and 2) How did the Obama 2012 campaign use it to persuade voters? (< 4 minutes)
Where customer experience and analytics meet, in real time
For a while now, I’ve been using Hailo as a European poster child for innovation in the context of big data analytics. Due to the level of interest generated by this example, and the number of questions I’ve received along the way about Hailo, its technology and business model, etc., I decided to put together this blog post rather than write loads of separate emails.
Ironically, I’ve not actually been able to use Hailo myself (much as I would like to), as I have neither an iOS or Android-based smartphone. I have, however, met lots of people who’re using Hailo as customers, and I’ve also spoken to taxi drivers about it. I have yet to meet anybody who isn’t a fan.
For those of you who don’t know Hailo, it’s an app that allows you to hail a registered cab from your smartphone; as it was started in London, it’s often also called “the black cab app.” With the company founders being three London cabbies (black cab drivers), the entire service has been uniquely focused around the needs of the two main participants in a taxi ride: the customer and the driver.
Notes from the TechAmerica Europe seminar in Brussels, March 27, 2013
This may not be the most timely event write-up ever produced, but in light of all the discussions I’ve had on the same themes during the past few weeks, I thought I’d share my notes anyway.
The purpose of the event was to peel away some of the hype layers around the “big data” discussion, and — from a European perspective — take a look at the opportunities as well as challenges brought by the increasing amounts of data that is available, and the technologies that enable its exploitation. As was to be expected, an ever-present subtext was the potential of having laws and regulations put in place which — while well-intentioned — can ultimately stifle innovation and even act against consumer interests. And speaking of innovation: Another theme running through several of the discussions was the seeming lack of technology-driven innovation in Europe, in particular when considered in the context of an economic environment in dire need of every stimulus it can get.
The scene was set by John Boswell, senior VP, chief legal officer, and corporate secretary at SAS, who provided a neat summary of the technology developments (cheap storage, unprecedented access to compute power, pervasive connectivity) giving rise to countless opportunities related to the availability, sharing and exploitation of ever-increasing amounts of data. He also outlined the threats posed to companies, governments, and individuals by those who with more sinister intent when it comes to data exploitation, be it for ideological, financial, or political reasons. Clearly, those threats require mitigation, but John also made the point that “regulatory overlays” can also hinder progress, through limiting or even preventing altogether the free flow of data.
Why all the fervor about big data? The answer is that it provides deep insights and predictive models that can dramatically improve business outcomes. But you need a data scientist to get there. There’s a lot of mythology about what a data scientist is and isn’t. In this episode of TechnoPolitics, Mike Gualtieri explains what a data scientist is, what skills they need, and how to hire one. You may also be interested in What Is Hadoop.
About Forrester Instant Insight
Navigating the fast changing world of business technology is a constant challenge. Forrester Instant Insight aims to provide simple, complete answers to some popular questions. Our goal: You will watch the video and be enlightened in 5-minutes or less.
This Forrester Instant Insight was produced by Mike Gualtieri and edited by Lindsay Gualtieri
In advance of next week’s Forrester’s European Business Technology Forums in London on June 10 and 11, we had an opportunity to speak with Greg Swimer about information management and how Unilever delivers real-time data to its employees. Greg Swimer is a global IT leader at Unilever, responsible for delivering new information management, business intelligence, reporting, consolidation, analytics, and master data solutions to more than 20,000 users across all of Unilever’s businesses globally.
1) What are the two forces you and the Unilever team are balancing with your “Data At Your Fingertips” vision?
Putting the data at Unilever’s fingertips means working on two complementary aspects of information management. One aspect is to build an analytics powerhouse with the capacity to handle big data, providing users with the technological power to analyse that data in order to gain greater insight and drive better decision-making. The other aspect is the importance of simplifying and standardizing that data so that it’s accessible enough to understand and act upon. We want to create a simplified landscape, one that allows better decisions, in real time, where there is a common language and a great experience for users.
2) What keys to success have you uncovered in your efforts?
How many of you suffer from at least mild “cyberchondria"? Do you run to the computer to Google your latest ailments? Are you often convinced that the headache you have is the first sign of some terminal illness you’ve been reading about?
Well, Symcat takes a new approach to Internet-assisted self-diagnosis. It provides not only the symptoms but the probability of getting the disease, using CDC data to rank results by the likelihood of the different conditions. It then allows users to further filter results by typing in information such as their gender, the duration of their symptoms and medical history. No, that headache you’ve had all week is likely not spinal stenosis or even viral pharyngitis. But if you’ve had a fall or a blow to the head you might want to consider a concussion.
As Symcat puts it, they “use data to help you feel better.” Never underestimate the palliative effects of peace of mind.
I had the chance to ask Craig Monsen, MD, co-founder and CEO of Symcat, a few questions about how they got their start with the business and their innovation with open data.
What was the genesis of Symcat? Can you describe the "ah-ha" moment of determining the need for Symcat?
There are multiple maturity models and associated assessments for Data Governance on the market. Some are from software vendors, or from consulting companies, which use these as the basis for selling services. Others are from professional groups like the one from the Data Governance Council.
They are all good – but frankly not adequate for the data economy many companies are entering into. I think it is useful to reshuffle some too well established ideas...
Maturity models in general are attractive because:
- Using a maturity model is nearly a ‘no-brainer’ exercise. You run an assessment and determine your current maturity level. Then you can make a list of the actions which will drive you to the next level. You do not need to ask your business for advice, nor involve too many people for interviews.
- Most data governance maturity models are modeled on the very well known CMMI. That means that they are similar at least in terms of structure/levels. So the debate between the advantages of one vs another is limited to its level of detail.
But as firms move into the data economy – with what this means for their sourcing, analyzing and leveraging data, I think that today’s maturity models for data governance are becoming less relevant – and even an impediment: