Background — High Performance Attached Processors Handicapped By Architecture
The application of high-performance accelerators, notably GPUs, GPGPUs (APUs in AMD terminology) to a variety of computing problems has blossomed over the last decade, resulting in ever more affordable compute power for both horizon and mundane problems, along with growing revenue streams for a growing industry ecosystem. Adding heat to an already active mix, Intel’s Xeon Phi accelerators, the most recent addition to the GPU ecosystem, have the potential to speed adoption even further due to hoped-for synergies generated by the immense universe of x86 code that could potentially run on the Xeon Phi cores.
However, despite any potential synergies, GPUs (I will use this term generically to refer to all forms of these attached accelerators as they currently exist in the market) suffer from a fundamental architectural problem — they are very distant, in terms of latency, from the main scalar system memory and are not part of the coherent memory domain. This in turn has major impacts on performance, cost, design of the GPUs, and the structure of the algorithms:
Performance — The latency for memory accesses generally dictated by PCIe latencies, which while much improved over previous generations, are a factor of 100 or more longer than latency from coherent cache or local scalar CPU memory. While clever design and programming, such as overlapping and buffering multiple transfers can hide the latency in a series of transfers, it is difficult to hide the latency for an initial block of data. Even AMD’s integrated APUs, in which the GPU elements are on a common die, do not share a common memory space, and explicit transfers are made in and out of the APU memory.
Dell just picked up Enstratius for an undisclosed amount today, making the cloud management vendor the latest well-known cloud controller to get snapped up by a big infrastructure or OS vendor. Dell will add Enstratius cloud management capabilities to its existing management suite for converged and cloudy infrastructure, which includes element manager and configuration automator Active System Manager (ASM, the re-named assets acquired with Gale Technologies in November), Quest Foglight performance monitoring, and (maybe) what’s still around from Scalent and DynamicOps.
This is a good move for Dell, but it doesn’t exactly clarify where all these management capabilities will fall out. The current ASM product seems to be a combo of code from the original Scalent acquisition upgraded with the GaleForce product; regardless of what’s in it, though, what it does is discover, configure and deploy physical and virtual converged infrastructure components. A private cloud automation platform, basically. Like all private cloud management stacks, it does rapid template-based provisioning and workflow orchestration. But it doesn’t provision apps or provision to public or open-source cloud stacks. That’s where Enstratius comes in.
I’ve noticed a growing trend among Asia Pacific organizations over the past 6-12 months: complete IT resistance to SaaS has steadily given way to more pragmatic discussions, even if IT has come to the table grudgingly. Over the next two years I expect this trend to accelerate. In fact, I believe that many SaaS solutions, particularly those that cross business and functional boundaries, will be rapidly subsumed within the broader IT portfolio, even if they were originally sourced outside IT.
Many SaaS vendors report already seeing more IT involvement in procurement, requirements definition, RFP creation, and negotiations. The clear procurement guidelines published by the IT department of the Australian Government Information Management Office (AGIMO) is one high profile example. Don’t get me wrong, in most instances business decision-makers will still lead, particularly in identifying the required business processes and determining how best to consume SaaS-based services. But IT decision-makers are getting more involved, particularly around integration.
Some areas to consider as you look to work more closely with business decision-makers to evaluate and negotiate SaaS and other public cloud deals:
No one would claim that the US tech market is booming. With Europe still mired in recession and debt problems, US economic growth looking soft, and business and consumer worries about the US government raising tax rates and cutting Federal spending, it is not surprising that businesses and governments are being cautious in their purchases of technology goods and services. But we think the fear is overblown. Forrester's forecast for the US tech market in 2013 and 2014 -- published today as "US Tech Market Outlook For 2013 And 2014: Better Times Ahead" -- projects a 6.2% rise in 2013 and a 6.8% growth in 2014 in US business and government purchases of computer equipment, communications equipment, software, IT consulting and systems integration services, and IT outsourcing. Adding in slow growing telecommunications services pulls growth down to 5.7% in 2013 and 6.1% in 2014. That may not be a boom, but it is certainly not a bust.
While CIOs are cautious in their tech buying -- and in the case of the Federal government, actually cutting back -- that caution has and will show up mostly in reduced spending on computer and communications equipment (with the exception of tablets). CIOs will be most aggressive in software, especially for SaaS apps, analytics, and mobile apps. IT outsourcing will see good growth in 2013 as the result of 2012 selection decisions, while IT consulting and systems integration will come on strong in 2014. Business and government purchases of telecommunications services will continue to grow at a slower rate than the overall tech market.
As businesses get serious about the cloud, developers are bringing more business-critical transaction data to cloud-resident web and mobile apps. Indeed, web and mobile apps that drive systems of engagement (how you interact with your customers and partners) are the reason why many companies look to the cloud in the first place. Public clouds offer the speed and agility developers want, plus the development tools they need. Once you’ve built a killer web or mobile app in the cloud and it’s in production, driving real revenue, who’s responsible for making sure it performs?
It’s a team effort. Developers have to think about performance management as they build, and IT operations teams need to design application monitoring and management into their cloud deployment processes up front. Why? Because there’s no time to do it later. You won’t have time to implement a new app monitoring solution for each new cloud app before you need to get it out to users. And once it’s out there, you need to be tracking user experience immediately.
In traditional IT, one of the reasons we could get away with limited insight into application performance was because we usually overprovisioned resources to make sure we didn’t have to worry about it. It’s easier to have excess capacity than to solve tricky performance problems – problems you might only see once in a while.
At the OpenStack Summit in Portland last week, the open-source cloud platform got real, to echo Forrester’s cloud team predictions for 2013. At the busy gathering attended by over 2,400, suits mingled effortlessly with hoodies and deep-tech design committee meetings were sandwiched between marquee-name customers sharing success stories. Three core themes drove the show, as outlined by Jonathan Bryce in the opening keynote: the OpenStack technology platform has matured, the ecosystem is vibrant, and the global user footprint now includes enterprise customers doing real business.
The show followed on the heels of the Grizzly release, the 7th release of the OpenStack platform. Along with stronger support for VMware and Microsoft hypervisors, Grizzly widens block storage options and includes 10+ new enterprise storage platform drivers and workload-based scheduling. A wide range of new network plugins expand the platform’s software-defined networking options and a sexier Dashboard to access, provision and automate resources.
HP today announced the Moonshot 1500 server, their first official volume product in the Project Moonshot server product family (the initial Redstone, a Calxeda ARM-based server, was only available in limited quantities as a development system), and it represents both a significant product today and a major stake in the ground for future products, both from HP and eventually from competitors. It’s initial attractions – an extreme density low power x86 server platform for a variety of low-to-midrange CPU workloads – hides the fact that it is probably a blueprint for both a family of future products from HP as well as similar products from other vendors.
Geek Stuff – What was Announced
The Moonshot 1500 is a 4.3U enclosure that can contain up to 45 plug-in server cartridges, each one a complete server node with a dual-core Intel Atom 1200 CPU, up to 8 GB of memory and a single disk or SSD device, up to 1 TB, and the servers share common power supplies and cooling. But beyond the density, the real attraction of the MS1500 is its scalable fabric and CPU-agnostic architecture. Embedded in the chassis are multiple fabrics for storage, management and network giving the MS1500 (my acronym, not an official HP label) some of the advantages of a blade server without the advanced management capabilities. At initial shipment, only the network and management fabric will be enabled by the system firmware, with each chassis having up two Gb Ethernet switches (technically they can be configured with one, but nobody will do so), allowing the 45 servers to share uplinks to the enterprise network.
Digital capability – social, mobile, cloud, data & analytics – disrupts business models, introduces new competitive threats, and places new demands on your business. Highlighting this fact: Forrester’s 2012 “Digital Readiness Assessment” survey found that 65% of global executives say they are “excited about the changes that digital tools and experiences will bring” to their company.
While most people know these digital trends are coming, however, far fewer know how to purchase these cutting-edge digital capabilities. What companies will you rely on? Where are the new risks? What are the pricing models? In the survey mentioned above, only 32% of the same sample agreed that their organization “has policies and business practices in place to adapt” to those digital changes.
This is important, since developing the breadth of digital capabilities your company needs cannot all be done in-house. To succeed, your company will need to access the strengths of its supplier ecosystem, maximize value from strategic partners, and leverage emerging supplier models.
This is a tremendous opportunity for sourcing and vendor management professionals to increase the strategic value they provide to their business. But to do this, you’ll need to balance your traditional cost-cutting goals with demands for business expectations for growth, innovation, and value.
A year and a half ago I broke up with Blackberry and started dating iPhone. It was a clean but cruel breakup: AT&T cancelled my T-Mobile contract on my behalf, the equivalent of getting dumped by your girlfriend’s new boyfriend.
This year I’ve been cheating on my laptop with my iPad. But it’s an on-again, off-again relationship. While I tell my iPad it’s the only one, I keep going back to my laptop. When I travel, my iPad is with me meeting clients. Meanwhile my laptop is in the hotel room surfing the online menu for a turkey club.
The iPad beats my laptop on size, weight, connectivity, and battery life. It also improves the human element when I’m having a face-to-face conversation but need to take notes. These are all critically important to me when I'm out of the office visiting clients or at an event.
But my laptop wins when I need to perform other important activities. For example, the larger screen really helps to write and edit research reports (John Rakowski, you’ll have your edits soon!). Or when I need to approve expenses behind the VPN or access files on my hard drive that I haven’t stored in Google Drive (yes, Forrester sanctioned).
Now that I've had a few months of compare both devices, I come back to outcomes . . .
For the vast majority of Forrester customers who I have not had the pleasure of meeting, my name is Henry Baltazar and I'm the new analyst covering Storage for the I&O team. I've covered the Storage industry for over 15 years and spent the first 9 years of my career as a Technical Analyst at eWEEK/PCWeek Labs, where I was responsible for benchmarking storage systems, servers and Network Operating Systems.
During my lab days, I tested hundreds of different products and was fortunate to witness the development and maturation of a number of key innovations such as data deduplication, WAN optimization and scale-out storage. In the technology space "Better, Faster, Cheaper - Pick Two" used to be the design goal for many innovators, and I've seen many technologies struggle to attain two, let alone three of these goals, especially in the first few product iterations. For example, while iSCSI was able to challenge Fibre Channel on the basis of being cheaper - despite being around for over a decade many storage professionals are still not convinced that iSCSI is faster or better.
Looking at storage technologies today, relative to processors and networking, storage has not held up its end of the bargain. Storage needs to improve in all three vectors to either push innovation forward, or avoid being viewed as a bottleneck in the infrastructure. At Forrester I will be looking at a number of areas of innovation which should drive enterprise storage capabilities to new heights including: