Services budgets represent 10% of annual IT operating and capital budgets[i], but Forrester sees considerable evidence that the influence of these IT Services vendors is proportionally higher — and growing dramatically. While there are several reasons for the rising importance of your services partners, at the most fundamental level Forrester sees that:
Business professionals need immediate access to tech-enabled innovation. Most strategic business initiatives now have an underlying technology component. Service providers come to the table with the tech savvy, vertical market expertise, and best practices to make these initiatives work.
IT professionals can’t keep pace with business demand. The volume and complexity of technology demands from business professionals means that traditional IT organizations have difficulty keeping pace. They too need to work with the best mix of IT service providers to meet the demands of their business. Effective supplier management is quickly becoming the most essential skill in IT organizations.
On July 11, 2012, SingTel launched its PowerON Compute cloud service in Hong Kong. While certainly interesting on its own, I believe this announcement is particularly noteworthy as a harbinger of things to come.
Some key points to consider:
As a hybrid offering, PowerON Compute is a dynamic infrastructure services solution hosted in SingTel’s data centers in Singapore, Australia, and now Hong Kong. The computing resources (e.g., CPU, memory, storage) can be accessed either via a public Internet connection or a private secured network.
This announcement confirms the findings of my February 2012 report, “Sizing the Cloud Markets in Asia Pacific”: that market demand for cloud-based computing resources in Asia Pacific (AP) will rapidly shift from infrastructure-as-a-service (IaaS) to dynamic infrastructure services.
The poorly kept secret that is the Google Nexus 7 tablet was just announced amid much developer applause and excitement. The device is everything it was rumored to be and the specs — something that only developers care about, of course — were impressive, including the 12 core GPU that will make the Nexus 7 a gaming haven. True, it's just another in a long line of tablets, albeit a $199 one that competes directly with Amazon's Kindle Fire and undercuts the secondary market for the iPad.
But as a competitor to the iPad, Nexus 7 isn't worth the digital ink I'm consuming right now.
But Google isn't just selling a device. Instead, the company wants to create a content platform strategy that ties together all of its ragtag content and app experiences into a single customer relationship. Because the power of the platform is the only power that will matter (see my recent post for more information on platform power). It's unfortunate that consumers barely know what Google Play is because it was originally called Android Market, but the shift to the Google Play name a few months back and the debut of a device that is, according to its designers, "made for Google Play," show that Google understands what will matter in the future. Not connections, not devices. But experiences. The newly announced Nexus 7, as a device, is from its inception subservient to the experiences — some of them truly awesome — that Google's Play platform can provide through it.
Earlier this week Dell joined arch-competitor HP in endorsing ARM as a potential platform for scale-out workloads by announcing “Copper,” an ARM-based version of its PowerEdge-C dense server product line. Dell’s announcement and positioning, while a little less high-profile than HP’s February announcement, is intended to serve the same purpose — to enable an ARM ecosystem by providing a platform for exploring ARM workloads and to gain a visible presence in the event that it begins to take off.
Dell’s platform is based on a four-core Marvell ARM V7 SOC implementation, which it claims is somewhat higher performance than the Calxeda part, although drawing more power, at 15W per node (including RAM and local disk). The server uses the PowerEdge-C form factor of 12 vertically mounted server modules in a 3U enclosure, each with four server nodes on them for a total of 48 servers/192 cores in a 3U enclosure. In a departure from other PowerEdge-C products, the Copper server has integrated L2 network connectivity spanning all servers, so that the unit will be able to serve as a low-cost test bed for clustered applications without external switches.
Dell is offering this server to selected customers, not as a GA product, along with open source versions of the LAMP stack, Crowbar, and Hadoop. Currently Cannonical is supplying Ubuntu for ARM servers, and Dell is actively working with other partners. Dell expects to see OpenStack available for demos in May, and there is an active Fedora project underway as well.
Just three months after SAP acquired SuccessFactors, a cloud leader for human capital management solutions, for $3.4 billion, it has now announced the acquisition of Ariba, a cloud leader for eProcurement solutions, for another $4.3 billion. Now, $7.7 billion is a lot of money to spend in a short amount of time on two companies that hardly make any profit. But it’s all for the cloud, which means it’s for the future business opportunity in cloud computing services. So far, so good; SAP has invested and acquired quite a number of cloud companies over the past years: Frictionless, Clear Standards, Crossgate, etc. The difference in this most recent acquisition is the big overlap with existing solutions and internal R&D.
Following the first wave of cloud acquisitions, SAP was sitting amid a zoo of cloud solutions, all based on different platforms: ePurchasing, CRM-OnDemand, BI-OnDemand, Carbon Impact, ByDesign, Streamwork . . . They all used very different technology, resulting in big integration and scale challenges behind the scenes. The market welcomed with open arms SAP’s announcement 1.5 years ago that it would consolidate its cloud strategy on the new NetWeaver platform for both ABAP- and Java-based cloud solutions.
Through a combination of analyst briefings and customer events, Cisco has ramped up outbound communication and marketing of its collaboration strategy in Asia Pacific over the past several months. The foundation remains video (TelePresence), webconferencing (WebEx), and IP telephony, areas where Cisco is a leader. But Cisco understands that to drive growth and expand its customer footprint within enterprise accounts, it must move further up the stack and increasingly compete with both traditional collaboration vendors like Microsoft and IBM and cloud-based alternatives like Google and salesforce.com.
While the strategy still plays to the company’s core networking strength, I question whether Cisco can position itself as a “go-to” vendor in the traditional collaboration space. As our research shows, senior IT and business decision-makers in Asia Pacific don’t currently equate Cisco with collaboration.
To address this challenge, Cisco is pursuing multiple initiatives/approaches:
Leveraging its core strengths. Cisco is focused on expanding from existing unified communications (UC) initiatives within customer accounts by leveraging the combination of networking and video to drive value. Cisco is pushing “control” via intelligent networking capabilities (e.g., security, identity management, authentication, access), all delivered through Cisco networking hardware. Simultaneously, Cisco is pushing “flexibility” via device- and platform-independent collaboration capabilities like content, video, instant messaging, and social computing.
Sound familiar? Executives across the globe feel peer and competitive pressure to “get to yes” on private cloud. This burden falls on IT to provide a cloud solution — oh, and by the way, we need it by the end of the year. With this clock ticking, it’s hard to think about private cloud strategically. In fact, why not to just cloudwash your virtual environment and buy your team time? Many enterprises (yes, even those presenting at events) have gone down this road. And some vendors will suggest this as a short-term fix. DON’T DO IT.
You’re cutting yourself short on what you could achieve with this environment while losing credibility with the business and your peers. Sound overdramatic? The consumerization of IT is forcing IT to connect with the business or risk circumvention. For many, the existing relationship isn't great. And each future interaction could either improve or worsen that relationship. Promising the business a cloud delivered within your own data center, and then failing to provide basic functionality of a cloud will just make future initiatives and interactions even harder. In the meantime, the business will continue to circumvent your department. If you're going to invest the resources/time to build this environment and rope in rogue cloud users — make sure you get to cloud.
SaaS vendors must collect customer insights for innovation and compliance.
As of the end of last year, about 30% of companies from our Forrsights Software Survey, Q4 2011, were using some software-as-a-service (SaaS) solution; that number will grow to 45% by the end of 2012 and 60% by the end of 2013. The public cloud market for SaaS is the biggest and fastest-growing of all of the cloud markets ($33 billion in 2012, growing to $78 billion by the end of 2015).
However, most of this growth is based on the cannibalization of the on-premises software market; software companies need to build their cloud strategy or risk getting stuck in the much slower-growing traditional application market and falling behind the competition. This is no easy task, however. Implementing a cloud strategy involves a lot of changes for a software company in terms of products, processes, and people.
A successful SaaS strategy requires an open architecture (note: multitenancy is not a prerequisite for a SaaS solution from a definition point of view but is highly recommended for vendors for better scale) and a flexible business model that includes the appropriate sales incentive structure that will bring the momentum to the street. For the purposes of this post, I’d like to highlight the challenge that software vendors need to solve for sustainable growth in the SaaS market: maintaining and increasing customer insights.
The other day I visited Colt’s London HQ and saw how the telco is revamping its approach to developing more customer-centric and Agile solutions (Colt consciously avoids the “cloud” terminology). By now, most telcos managed to jump onto the cloud bandwagon by launching cloud-based services. The challenge, from an end user perspective, is that these solutions all seem very similar. Customers can get storage, server capacity, unified communications, etc., from most telcos. All telcos underline the value-added nature of end-to-end network QoS and security that they can ensure (check out our report, "Telcos As Cloud Rainmakers"). Indeed, telcos have some right to feel that they have achieved some progress regarding their cloud offerings — although it took Amazon to show them the opportunity.
But most telco cloud offerings suffer from the fact that telcos develop cloud solutions in the traditional sense through their traditional product factories. This approach tends to follow rather than slow product innovation cycles. Moreover, it produces products that, once developed, are pushed to the customer as a standard offering. All customisation costs extra.
The reality of cloud demand is that each customer is different. Most customers want some form of customisation. Most customers want some form of hybrid cloud, a private part for core apps, as well as access to the open Internet to, for instance, exchange views and information with end customers via Twitter or for crowd sourcing with suppliers. Similarly, most customers want a mix of fixed and virtual assets and a blend of self-service and managed service solutions as the chart indicates.
I said last year that this would happen sometime in the first half of this year, but for some reason my colleagues and clients have kept asking me exactly when we would see a real ARM server running a real OS. How about now?
To copy from Calxeda’s most recent blog post:
“This week, Calxeda is showing a live Calxeda cluster running Ubuntu 12.04 LTS on real EnergyCore hardware at the Ubuntu Developer and Cloud Summit events in Oakland, CA. … This is the real deal; quad-core, w/ 4MB cache, secure management engine, and Calxeda’s fabric all up and running.”
This is a significant milestone for many reasons. It proves that Calxeda can indeed deliver a working server based on its scalable fabric architecture, although having HP signing up as a partner meant that this was essentially a non-issue, but still, proof is good. It also establishes that at least one Linux distribution provider, in this case Ubuntu, is willing to provide a real supported distribution. My guess is that Red Hat and Centos will jump on the bus fairly soon as well.
Most importantly, we can get on with the important work of characterizing real benchmarks on real systems with real OS support. HP’s discovery centers will certainly play a part in this process as well, and I am willing to bet that by the end of the summer we will have some compelling data on whether the ARM server will deliver on its performance and energy efficiency promises. It’s not a slam dunk guaranteed win – Intel has been steadily ratcheting up its energy efficiency, and the latest generation of x86 server from HP, IBM, Dell, and others show promise of much better throughput per watt than their predecessors. Add to that the demonstration of a Xeon-based system by Sea Micro (ironically now owned by AMD) that delivered Xeon CPUs at a 10 W per CPU power overhead, an unheard of efficiency.