After considerable speculation and anticipation, VMware has finally announced vSphere 5 as part of a major cloud infrastructure launch, including vCloud Director 1.5, SRM 5 and vShield 5. From our first impressions, it is both well worth the wait and merits immediate serious consideration as an enterprise virtualization platform, particularly for existing VMware customers.
The list of features is voluminous, with at least 100 improvements, large and small, but among the features, several stand out as particularly significant as I&O professionals continue their efforts to virtualize the data center, primarily dealing with and support for both larger VMs and physical host systems, and also with the improved manageability of storage and improvements Site Recovery Manager (SRM), the remote-site HA components:
Replication improvements for Site Recovery Manager, allowing replication without SANs
Distributed Resource Scheduling (DRS) for Storage
Support for up to 1 TB of memory per VM
Support for 32 vCPUs per VM
Support for up to 160 Logical CPUs and 2 TB or RAM
New GUI to configure multicore vCPUs
Storage driven storage delivery based on the VMware-Aware Storage APIs
Improved version of the Cluster File System, VMFS5
Storage APIs – Array Integration: Thin Provisioning enabling reclaiming blocks of a thin provisioned LUN on the array when a virtual disk is deleted
Swap to SSD
2TB+ LUN support
Storage vMotion snapshot support
vNetwork Distributed Switch improvements providing improved visibility in VM traffic
vCenter Server Appliance
vCenter Solutions Manager, providing a consistent interface to configure and monitor vCenter-integrated solutions developed by VMware and third parties
Revamped VMware High Availability (HA) with Fault Domain Manager
Cloud computing continues to be hyped. By now, almost every ICT hardware, software, and services company has some form of cloud strategy — even if it’s just a cloud label on a traditional hosting offering — to ride this wave. This misleading vendor “cloud washing” and the complex diversity of the cloud market in general make cloud one of the most popular and yet most misunderstood topics today (for a comprehensive taxonomy of the cloud computing market, see this Forrester blog post).
Software-as-a-service (SaaS) is the largest and most strongly growing cloud computing market; its total market size in 2011 is $21.2 billion, and this will explode to $78.4 billion by the end of 2015, according to our recently published sizing of the cloud market. But SaaS consists of many different submarkets: Historically, customer relationship management (CRM), human capital management (HCM) — in the form of “lightweight” modules like talent management rather than payroll — eProcurement, and collaboration software have the highest SaaS adoption rates, but highly integrated software applications that process the most sensitive business data, such as enterprise resource planning (ERP), are the lantern-bearers of SaaS adoption today.
What is it that you think makes one tech company stand out from another? “My product is better than your product”? Not anymore. “My salespeople are better than your salespeople”? Possibly. “My channel is better than your channel”. You’re getting warmer. How about, “My marketing machine is better than your marketing machine”?
For example, 41% of customers identify “the vendor’s (not including its salespeople’s) ability to understand our business problem”, compared with only 21% who identified “the vendor’s salesperson’s ability to understand our business problem” as the most important vendor action factor when selecting a tech vendor. Marketing is clearly the difference-maker.
But cloud computing changes everything. The implications of cloud computing go far beyond its technology delivery/consumption model. It seems I get questions from tech marketers about all things cloud these days. A few examples:
“How can I use the cloud more effectively to market our solutions?” (Answer: It’s not what you read in USA Today about Facebook and Twitter. According to the results of our 2011 B2B Social Technographics® survey, discussion forums and professional social networking sites (read: not consumer social sites) outpace Facebook and Twitter ten-fold as information sources for informing businesses’ technology purchase decisions.)
The drum continues to beat for converged infrastructure products, and Dell has given it the latest thump with the introduction of vStart, a pre-integrated environment for VMware. Best thought of as a competitor to VCE, the integrated VMware, Cisco and EMC virtualization stack, vStart combines:
A lot of tech vendors – and channel partners – are struggling over what channel partners’ play in the cloud services demand chain is going to be. Technology is decreasingly delivered/consumed in the form of on-premise installation (a function performed by and the original raison d’être of channel partners), and increasingly delivered as-a-service by a service provider. In the software sector, that service provider is typically (but not always) the software vendor (think: salesforce.com).
And, in most cases, for good reason. Software has bugs. Early versions of software can be unstable and unpredictable. In the classic channel-partner-sells-and-installs-software model, the product (the software) remains in the control of the software vendor, i.e., the vendor assumes the risk of customers’ unmet expectations. The license is between the vendor and the customer, and the vendor is on the hook for providing bug fixes and tier-2 and -3 support.
As much as many channel partners would like to act as application hosters (and many of them do – approximately 15% of software is delivered via a hosting model today, and 20% of channel partners today have a hosting business [see “Channel Models In The Era Of Cloud”]), when it comes to early-version or mission-critical software, vendors simply can’t risk putting the as-a-service service level/performance responsibility in the hands of channel partners. Service failures, over which the vendor would have no control, would result in egg (or worse!) on the vendor’s brand, not the channel partner’s. Until tech vendors’ partner programs mature to the point where they can certify partners’ data centers, those vendors are going to be reticent to hand over the data center reins to partners.
Cloud infrastructure-as-a-service (IaaS) is a hot market. Amazon Web Services, now five years old, drives a lot of attention and customer volume, but the vendor strategists at enterprise-facing providers such as IBM, HP, AT&T and Verizon have been building and delivering IaaS offerings. As I’ve studied the market, I’ve heard wildly different types of requirements from buyers and quite a range of offerings from service providers. Yet much of the industry dialogue is about one central idea of what IaaS is – think that’s wrong headed. I found that there were really two buyer types: 1) informal buyers outside of the IT operations/data center manager organizations, such as engineers, scientists, marketing executives, and developers, and 2) formal buyers, the IT operations and data center managers responsible for operating applications and maintaining infrastructure.
With this idea in mind, I set out to test the views of IT infrastructure buyers in the Forrsights Hardware Survey, Q3 2010 and learned that:
After 2+ years of cloud hype, only 6% of enterprises IT infrastructure respondents report using IaaS, with another 7% planning to implement by Q3, 2012. After flat adoption from 2008 to 2009, this represents an approximate doubling from 2009, off a very small base.
Almost two thirds of IT infrastructure buyers themselves don’t believe they are the primary buyer of cloud IaaS! We asked them which groups in their company are using or most interested in cloud IaaS. Only 36% of IT infrastructure buyers listed themselves, while 7% didn’t know. The rest, 58% said that IT developers, Web site owners, business unit owners of batch compute intensive apps, and other business unit developers were more interested in using IaaS than themselves.
Calxeda, one of the most visible stealth mode startups in the industry, has finally given us an initial peek at the first iteration of its server plans, and they both meet our inflated expectations from this ARM server startup and validate some of the initial claims of ARM proponents.
While still holding their actual delivery dates and details of specifications close to their vest, Calxeda did reveal the following cards from their hand:
The first reference design, which will be provided to OEM partners as well as delivered directly to selected end users and developers, will be based on an ARM Cortex A9 quad-core SOC design.
The SOC, as Calxeda will demonstrate with one of its reference designs, will enable OEMs to design servers as dense as 120 ARM quad-core nodes (480 cores) in a 2U enclosure, with an average consumption of about 5 watts per node (1.25 watts per core) including DRAM.
While not forthcoming with details about the performance, topology or protocols, the SOC will contain an embedded fabric for the individual quad-core SOC servers to communicate with each other.
Most significantly for prospective users, Calxeda is claiming, and has some convincing models to back up these claims, that they will provide a performance advantage of 5X to 10X the performance/watt and (even higher when price is factored in for a metric of performance/watt/$) of any products they expect to see when they bring the product to market.
Intel, despite a popular tendency to associate a dominant market position with indifference to competitive threats, has not been sitting still waiting for the ARM server phenomenon to engulf them in a wave of ultra-low-power servers. Intel is fiercely competitive, and it would be silly for any new entrants to assume that Intel will ignore a threat to the heart of a high-growth segment.
In 2009, Intel released a microserver specification for compact low-power servers, and along with competitor AMD, it has been aggressive in driving down the power envelope of its mainstream multicore x86 server products. Recent momentum behind ARM-based servers has heated this potential competition up, however, and Intel has taken the fight deeper into the low-power realm with the recent introduction of the N570, a an existing embedded low-power processor, as a server CPU aimed squarely at emerging ultra-low-power and dense servers. The N570, a dual-core Atom processor, is being currently used by a single server partner, ultra-dense server manufacturer SeaMicro (see Little Servers For Big Applications At Intel Developer Forum), and will allow them to deliver their current 512 Atom cores with half the number of CPU components and some power savings.
Technically, the N570 is a dual-core Atom CPU with 64 bit arithmetic, a differentiator against ARM, and the same 32-bit (4 GB) physical memory limitations as current ARM designs, and it should have a power dissipation of between 8 and 10 watts.
SAP Has Managed A Turnaround After Léo Apotheker’s Departure
In February 2010, after Léo Apotheker resigned as CEO of SAP, I wrote a blog post with 10 predictions for the company for the remaining year. Although the new leadership mentioned again and again that this step would not have any influence on the company’s strategy, it was clear that further changes would follow, as it doesn’t make any sense to simply replace the CEO and leave everything else as is when problems were obviously growing bigger for the company.
I predicted that the SAP leadership change was just the starting point, the visible tip of an iceberg, with further changes to come. Today, one year later, I want to review these predictions and shed some light on 2010, which has become the “Turnaround Year For SAP.”
The 10 SAP Predictions For 2010 And Their Results (7 proved true / 3 proved wrong)
Only a few weeks to go before Forrester’s US EA Forum 2011 in San Francisco in February! I’ll be presenting a number of sessions, including the opening kickoff, where I’ll paint a picture of where I see EA going in the next decade. As Alex Cullen mentioned, I’ll examine three distinct scenarios where EA rises in importance, EA crashes and burns, or EA becomes marginalized.
But the most fun I’ve had preparing for this year’s event is putting together a new track: “Key Technology Trends That Will Change Your Business.” In the past, we’ve focused this conference on the practice of EA and used our big IT Forum conference in the spring to talk about technology strategies, but this year I’ve had the opportunity to put together five sessions that drill down into the technology trends that we think will have significant impact in your environment, with a particular focus on impacting business outcomes. Herewith is a quick summary of the sessions in this track: