August 13, 2007
If I had to boil what I observed at last week’s Next Generation Data Center conference into one thought, it would be this: Although technologies to virtualize, optimize and automate datacenters do exist and are mature, organizations are still leery about making the transformation, despite coveting the associated improvements in performance, flexibility and overall efficiency.
As evidence that these technologies are mature, one really need look no further than the event’s opening keynotes, in which Amazon CTO Werner Vogels and eBay distinguished research scientist Paul Strong discussed the Web-scale datacenters being operated by their respective employers. Strong, for his part, delved a little deeper into the nitty gritty details (see here and here for more on this), all the while, however, stressing the importance of building a datacenter that: (1) runs processes driven by SLAs; (2) operates as a value center rather than a cost center; and (3) enables the rolling out of new utilities, platforms, etc. To achieve this next-generation datacenter, he said, many technologies can and should be considered, including (but certainly not limited to) grid computing, utility computing, real-time solutions and virtualization.
Now, it’s unlikely that most organizations have the resources (or the need -- said Strong, “If we don’t keep up, our business is gone.”) to develop and/or manage a datacenter like eBay’s -- a highly automated and virtualized environment consisting of several thousand blades servers -- but that doesn’t mean companies with less demand on their infrastructures can’t learn some lessons from the online auction leader. For starters, said Strong, automation is the key to efficiency and, in point he made sure to drive home, it is important to “manage relationships, not things.” This advice should be particularly relevant as today’s average datacenters continue to evolve toward advanced models like that of eBay. After all, when you’re staring down thousands of physical machines (and likely significantly more virtual ones) you can’t possibly expect to manage each one individually.
As for Vogels, whose presentation kicked off the doubleheader, he targeted his comments toward those companies who aren’t too keen on managing their own resources, and he used the opportunity to push Amazon’s stable of Web services. For most companies, he estimates, 70 percent of time and money expenditures go toward “heavy lifting” operations like maintenance, load balancing or software management, among others, none of which offer much in terms of innovation or helping differentiate your company from competitors. Unless you’re in an industry where having a customized, highly efficient datacenter directly translates into dollars, he suggested, it might be a waste of resources on all levels. “At [Amazon’s] scale, datacenters matter,” he stated. “They don’t for everyone.”
Following his statement that “I hate datacenters,” Vogels cited the recent power outage at a popular San Francisco datacenter -- which led to temporary shutdowns of several Web 2.0 leaders, including Craigslist, Second Life and Netflix -- as one example of what can go wrong. Power-wise, he added, even if you have generators to ensure you keep running, you still need batteries to handle the gap in time between the power going out and the generators kicking in. Still in the realm of possible physical issues, Vogels noted that datacenter managers still need to worry about issues such as sufficient cooling and how to handle a fire or other disasters. And that’s not even addressing business-side concerns, such as whether one datacenter is enough, or how you’re going to push out enough bandwidth to handle demand. Often times, he said, companies need to overprovision to handle peak loads or in case they become successful.
His solution? Utilize services such as Amazon’s Elastic Compute Cloud (EC2), as well its other Web services offerings, to handle your computing needs, paying only for what you actually use. Vogels talked about this notion as the “push versus pull” model of resource management, where “pushing” refers to the old-style method of preemptively pushing resources toward problems and “pulling” refers to the more progressive concept of pulling in resources from a centralized pool as needed -- just like Amazon itself does. For organizations that don’t necessarily benefit from managing their own datacenters, he said, this utility model will allow them to get the computing resources they need elsewhere, thus freeing up time and money to spend on innovation.
However, while Vogels’ solution to datacenter woes might sound logical and easy enough to incorporate, I wouldn’t hold my breath waiting for this idea to gain mainstream acceptance, much less adoption. Why, you ask? Because organizations are having a difficult enough time taking advantage of next-generation tools within their own walls -- something far less scary than the notion of relying on someone else to make sure their applications get the attention they deserve. This was made abundantly clear during a presentation by OGF President Mark Linesch, who was simply trying to lay out the business case for and current status of grid computing technologies.
During Linesch’s presentation, an audience member (who actually has some firsthand grid experience under his belt) asked how he is supposed to sell the idea of grid or virtualization to his applications developers, some of whom still oppose running their applications on distributed or virtualized platforms. (In fact, this gentleman just denied a request for 60-plus servers to develop and test a new application, instead preferring the work to be done on existing, virtually partitioned machines.) Linesch, backed up by Ravi Subramaniam, who has plenty of insight to offer after years of managing Intel’s in-house grid, gave really the only answer one can give in this situation, regardless of how frustrating it might be to someone desperately seeking a cost-efficient, dynamic IT platform for his/her organization: You have to start slow, showing success in one area at a time -- perhaps with just one application -- and illustrate how that translates into other areas. Not exactly the best way to show off a technology’s full range of capabilities, but not exactly uncommon, either.
This point was hammered home when I sat down to speak with Jay Fry and Ken Oestreich of Cassatt, vice president of marketing and director of product management and marketing, respectively. With its capabilities in areas like capacity on-demand, application virtualization, service level automation, utility metering, etc., Cassatt’s Collage software certainly falls under the “next-generation” umbrella, but customers aren’t always ready to experience it in full force from the get-go. In fact, customers have been known to ask for a pared down version of Collage, something Cassatt might have to do in order to show them -- one step at a time -- that its software is for real.
I was happy to hear, though, that Cassatt is making inroads on another front: the battle to ease customers’ minds about the cultural and organization changes that come along with the technical changes of a shared IT platform. Gone are the siloed applications and their siloed personnel. Gone are the days of server hugging. Gone are the days of low utilization and high overhead. While these all sound like great things, that kind of change apparently can be quite foreboding for IT departments, which is why many are hesitant to cross over into the promised land. Well, Cassatt and consultancy partner BearingPoint have been walking customers through this process, which they believe needs to be done in parallel, in their New York-based customer experience center, and the reaction has been very positive thus far. You can read more about the Cassatt/BearingPoint partnership here, and you can expect to see more about Cassatt’s take on utility computing in the weeks to come.
Speaking of application virtualization and its associated functionalities, the topic came up in a panel discussion featuring three distinct virtualization users and, wouldn’t you know, it seems to be a little much for them at this point. When the topic of “virtualization 2.0” was brought up, which was defined as including the grid-like abilities (e.g., high availability, SLA management, scalability, etc.) often associated with application virtualization solutions, the response was not overly positive. While Brian Harris, president and founder of Virtual Ngenuity, stated that he believes these functionalities are currently driving business decisions, two of his fellow panelists showed that this might not be entirely the case. Richard Robinson, chief operations officer for Department of Telecommunications and Information Services, City and County of San Francisco, commented that while his department is working toward these goals, they are not there yet and might not be for quite a while. After all, he noted, the “if it’s not broken, don’t fix it” axiom carries a lot of weight in the local government sector. Sudip Chahal, a senior architect in Intel’s IT Strategy, Architecture and Innovation organization, espoused his belief that this type of architecture isn’t well-suited for traditional business applications -- a view not shared by audience member Dave Pearson of Oracle or, I would assume, most of the distributed IT community.
Of course, there was more going on at LinuxWorld/NGDC than just vendors showing off to and discussing with skeptical end-users their cutting-edge technologies. For example, some cool news also came out of the show, such as Appistry adding power-saving functionality to its Enterprise Application Fabric; ServePath announcing high-performance hosting via its virtualized GoGrid service; EnterpriseDB challenging Oracle with GridSQL; and IBM tackling your mountains of widely dispersed data with its grid- and virtualization-powered Information Server Blade, which Big Blue says has demonstrated significant improvements in batch process performance, hardware price performance and budget expenditure. If you’re still hungry for more after reading these announcements, don’t fret, as we’ll have more on all of them -- as well as a look at the growing grid hosting business -- in the weeks and months to come.
Outside of NGDC news, be sure to check these very noteworthy items: “GigaSpaces Powers Sun's Market Data Solution”; “NCAR Adds Resources to TeraGrid”; “Imense Using Grid to Become 'Google of Image Searching'”; “Trigence Intros Optimized App Virtualization Software”; “Sun Releases Fastest Commodity Microprocessor”; and “Layered Tech Announces Super Grid.”
Comments about GRIDtoday are welcomed and encouraged. Write to me, Derrick Harris, at firstname.lastname@example.org.
Posted by Derrick Harris - August 13, 2007 @ 11:28 AM, Pacific Daylight Time
Derrick Harris is the Editor of On-Demand Enterprise
No Recent Blog Comments
The ever-growing complexity of scientific and engineering problems continues to pose new computational challenges. Thus, we present a novel federation model that enables end-users with the ability to aggregate heterogeneous resource scale problems. The feasibility of this federation model has been proven, in the context of the UberCloud HPC Experiment, by gathering the most comprehensive information to date on the effects of pillars on microfluid channel flow.
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
Frank Ding, engineering analysis & technical computing manager at Simpson Strong-Tie, discussed the advantages of utilizing the cloud for occasional scientific computing, identified the obstacles to doing so, and proposed workarounds to some of those obstacles.
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/02/2012 | AMD | Developers today are just beginning to explore the potential of heterogeneous computing, but the potential for this new paradigm is huge. This brief article reviews how the technology might impact a range of application development areas, including client experiences and cloud-based data management. As platforms like OpenCL continue to evolve, the benefits of heterogeneous computing will become even more accessible. Use this quick article to jump-start your own thinking on heterogeneous computing.