Yesterday Intel had a major press and analyst event in San Francisco to talk about their vision for the future of the data center, anchored on what has become in many eyes the virtuous cycle of future infrastructure demand – mobile devices and “the Internet of things” driving cloud resource consumption, which in turn spews out big data which spawns storage and the requirement for yet more computing to analyze it. As usual with these kinds of events from Intel, it was long on serious vision, and strong on strategic positioning but a bit parsimonious on actual future product information with a couple of interesting exceptions.
Content and Core Topics:
No major surprises on the underlying demand-side drivers. The the proliferation of mobile device, the impending Internet of Things and the mountains of big data that they generate will combine to continue to increase demand for cloud-resident infrastructure, particularly servers and storage, both of which present Intel with an opportunity to sell semiconductors. Needless to say, Intel laced their presentations with frequent reminders about who was the king of semiconductor manufacturingJ
Initial business intelligence (BI) ployment efforts are often difficult to predict and may dwarf the investment you made in BI platform software. The effort and costs associated with professional services, whether you use internal staff or hire contractors, depend not only on the complexity of business requirements like metrics, measures, reports, dashboards, and alerts, but also on the number of data sources you are integrating, the complexity of your data integration processes, and logical and physical data modeling. At the very least Forrester recommends considering the following components and their complexity to estimate development, system integration and deployment effort:
I’ve been presenting research on big data and data governance for the past several months where I show a slide of a businesswoman doing a backbend to access data in her laptop. The point I make is that data management has to be hyper-flexible to meet a wider range of analytic and consumption demands than ever before. Translated, you need to cross-train for data management to have cross-fit data.
The challenge is that traditional data management takes a one-size fits-all approach. Data systems are purpose built. If organizations want to reuse a finance warehouse for marketing and sales purposes, it often isn’t a match and a new warehouse is built. If you want to get out of this cycle and go from data couch potato to data athlete, a cross-fit data training program should focus on:
Context first. Understanding how data is used and will provide value drives platform design. Context indicates more than where data is sourced from and where it will be delivered. Context answers: operations or analytics, structured or unstructured, persistent or disposable? These guide decisions around performance, scale, sourcing, cost, and governance.
Data governance zones. Command and control data governance creates a culture of “no” that stifles innovation and can cause the business to go around IT for data needs. The solution is to create policies and processes that give permission as well as mitigate risk. Loosen quality and security standards in projects and scenarios that are in contained environments. Tighten rules and create gates when called for by regulation, where there are ethical conflicts, or when data quality or access exposes the business to significant financial risk.
We recently attended Amdocs' customer event in Singapore. Amdocs is gradually adjusting its strategy to reflect one of the most fundamental changes in the ICT industry today: Increasingly, business line managers, think the marketing or sales officer, are the ones influencing sourcing decisions. Traditional decision-makers, CTOs and CIOs, are no longer the sole ICT decision-makers. Amdocs is addressing this shift by:
Strengthening its customer experience portfolio.Successful telcos will try to regain lost relevance through improved customer experience. Marketing, portfolio product development, and sales are therefore growing in importance for telcos. Amdocs’ integrated customer experience offering, CES 9, provides telcos with a multichannel experience; proactive care; and self-service tools.
Betting big on big data/analytics.Amdocs is leveraging big data/analytics to provide real-time, predictive, and prescriptive insights to telcos about their customers’ behaviour. Communications-industry-specific converged charging and billing solutions as well as other catalogue solutions give Amdocs the opportunity to provide more value to telcos than some of the other players.
I attended Google’s annual atmosphere road show recently, an event aimed at presenting solutions for business customers. The main points I took away were:
Google’s “mosaic” approach to portfolio development offers tremendous potential. Google has comprehensive offerings covering communications and collaboration solutions (Gmail, Google Plus), contextualized services (Maps, Compute Engine), application development (App Engine), discovery and archiving (Search, Vault), and access tools to information and entertainment (Nexus range, Chromebook/Chromebox).
Google’s approach to innovation sets an industry benchmark. Google is going for 10x innovation, rather than the typical industry approach of pursuing 10% incremental improvements. Compared with its peers, this “moonshot” approach is unorthodox. However, moonshot innovation constitutes a cornerstone of Google’s competitive advantage. It requires Google’s team to think outside established norms. One part of its innovation drive encourages staff to spend 20% of their work time outside their day-to-day tasks. Google is a rare species of company in that it does not see failure if experiments don’t work out. Google cuts the losses, looks at the lessons learned — and employees move on to new projects.
The Obama 2012 campaign famously used big data predictive analytics to influence individual voters. They hired more than 50 analytics experts, including data scientists, to predict which voters will be positively persuaded by political campaign contact such as a call, door knock, flyer, or TV ad. Uplift modeling (aka persuasion modeling) is one of the hottest forms of predictive analytics, for obvious reasons — most organizations wish to persuade people to to do something such as buy! In this special episode of Forrester TechnoPolitics, Mike interviews Eric Siegel, Ph.D., author of Predictive Analytics, to find out: 1) What exactly is uplift modeling? and 2) How did the Obama 2012 campaign use it to persuade voters? (< 4 minutes)
Where customer experience and analytics meet, in real time
For a while now, I’ve been using Hailo as a European poster child for innovation in the context of big data analytics. Due to the level of interest generated by this example, and the number of questions I’ve received along the way about Hailo, its technology and business model, etc., I decided to put together this blog post rather than write loads of separate emails.
Ironically, I’ve not actually been able to use Hailo myself (much as I would like to), as I have neither an iOS or Android-based smartphone. I have, however, met lots of people who’re using Hailo as customers, and I’ve also spoken to taxi drivers about it. I have yet to meet anybody who isn’t a fan.
For those of you who don’t know Hailo, it’s an app that allows you to hail a registered cab from your smartphone; as it was started in London, it’s often also called “the black cab app.” With the company founders being three London cabbies (black cab drivers), the entire service has been uniquely focused around the needs of the two main participants in a taxi ride: the customer and the driver.
Notes from the TechAmerica Europe seminar in Brussels, March 27, 2013
This may not be the most timely event write-up ever produced, but in light of all the discussions I’ve had on the same themes during the past few weeks, I thought I’d share my notes anyway.
The purpose of the event was to peel away some of the hype layers around the “big data” discussion, and — from a European perspective — take a look at the opportunities as well as challenges brought by the increasing amounts of data that is available, and the technologies that enable its exploitation. As was to be expected, an ever-present subtext was the potential of having laws and regulations put in place which — while well-intentioned — can ultimately stifle innovation and even act against consumer interests. And speaking of innovation: Another theme running through several of the discussions was the seeming lack of technology-driven innovation in Europe, in particular when considered in the context of an economic environment in dire need of every stimulus it can get.
The scene was set by John Boswell, senior VP, chief legal officer, and corporate secretary at SAS, who provided a neat summary of the technology developments (cheap storage, unprecedented access to compute power, pervasive connectivity) giving rise to countless opportunities related to the availability, sharing and exploitation of ever-increasing amounts of data. He also outlined the threats posed to companies, governments, and individuals by those who with more sinister intent when it comes to data exploitation, be it for ideological, financial, or political reasons. Clearly, those threats require mitigation, but John also made the point that “regulatory overlays” can also hinder progress, through limiting or even preventing altogether the free flow of data.
Why all the fervor about big data? The answer is that it provides deep insights and predictive models that can dramatically improve business outcomes. But you need a data scientist to get there. There’s a lot of mythology about what a data scientist is and isn’t. In this episode of TechnoPolitics, Mike Gualtieri explains what a data scientist is, what skills they need, and how to hire one. You may also be interested in What Is Hadoop.
About Forrester Instant Insight
Navigating the fast changing world of business technology is a constant challenge. Forrester Instant Insight aims to provide simple, complete answers to some popular questions. Our goal: You will watch the video and be enlightened in 5-minutes or less.
This Forrester Instant Insight was produced by Mike Gualtieri and edited by Lindsay Gualtieri
In advance of next week’s Forrester’s European Business Technology Forums in London on June 10 and 11, we had an opportunity to speak with Greg Swimer about information management and how Unilever delivers real-time data to its employees. Greg Swimer is a global IT leader at Unilever, responsible for delivering new information management, business intelligence, reporting, consolidation, analytics, and master data solutions to more than 20,000 users across all of Unilever’s businesses globally.
1) What are the two forces you and the Unilever team are balancing with your “Data At Your Fingertips” vision?
Putting the data at Unilever’s fingertips means working on two complementary aspects of information management. One aspect is to build an analytics powerhouse with the capacity to handle big data, providing users with the technological power to analyse that data in order to gain greater insight and drive better decision-making. The other aspect is the importance of simplifying and standardizing that data so that it’s accessible enough to understand and act upon. We want to create a simplified landscape, one that allows better decisions, in real time, where there is a common language and a great experience for users.
2) What keys to success have you uncovered in your efforts?