July 03, 2012
Google is officially in the Infrastructure-as-a-Service (IaaS) business. Its newly-launched utility computing service, Google Compute Engine, lets users run workloads inside big G's massive datacenters, just like Amazon does with its Elastic Compute Cloud (EC2). Urs Hölzle, Google's vice president of engineering and longtime employee, revealed the news last Thursday at the company's annual developers conference. The compute service joins Google's App Engine PaaS, which launched in 2008 and now supports over one million active apps and receives 7.5 billion hits per day.
Mirroring the Microsoft news from earlier in the month, Google simultaneously downshifted and expanded its cloud stack when it added an IaaS offering. The only surprise is that it didn't happen sooner. Google's uber-scale datacenters are perfect for a public infrastructure cloud of this ilk. Google said that it was the company's success with App Engine and customer demand that led to the debut of an IaaS product.
While comparisons can be made to HP's Cloud Service, Microsoft's revamped Azure offering, and Amazon-runnerup Rackspace, the company with the biggest X on its back is none other than lead IaaS purveyor Amazon, which kicked off its EC2 service in 2006, two years ahead of Google's App Engine. Google, though, is one of the few companies with the infrastructure scale and technical know-how to pose a threat to AWS dominance. Utility computing is a volume business, and Google is well-positioned to support that volume. With hindsight, you could say the Web-era poster child has spent the last decade preparing for this moment.
The new service lets users run Linux Virtual Machines (VMs) on the same infrastructure that powers Google. Currently, 1, 2, 4 and 8 virtual core VMs are available with 3.75 GB RAM per virtual core. Already partners are lining up. RightScale, Puppet Labs, Opscode, MapR, CliQr and Numerate have all announced integration with Google Compute Engine.
Google is emphasizing three qualities essential for a successful IaaS play: scale, performance and value. Google Product Manager Craig McLuckie spelled out how the company will drive each of these areas. As to value, McLuckie makes the bold claim that Google Compute Engine gives users 50 percent more compute power dollar-for-dollar than the competition, but does not say how he arrived at this figure.
While direct oranges-to-oranges comparisons are difficult to generate, that doesn't stop analysts from trying. Google's n1-standard-8-d Machine Type furnishes an 8-core VM with 30GB of RAM and 22 "Cloud Engine Units" at $1.16 per hour. The closest Amazon EC2 Instance Type is the High-Memory Double Extra Large Instance (m2.2xlarge) with 4 virtual cores, 34.2GB of RAM and 13 "Compute Units" at $.90 per hour. Keep in mind the Amazon setup offers 1.7x less CPU, and when you line them up based on $/unit/hour, Google comes out ahead (cheaper): $0.053 versus $0.069. Note that the GCEU (Google Cloud Engine Unit) is approximately equal to the EC2 Compute Unit, and Amazon prices are based on their least expensive region (US East - Virginia).
Circular representation of a human genome
While the public cloud can be used for a range of workloads, Google has suggested several primary use cases, including batch processing jobs, like video transcoding and rendering; data processing and analysis using Hadoop; and high-performance computing. Speaking of HPC, the service allows users to launch enormous compute clusters on the order of tens of thousands of cores and higher. In fact, during his keynote, Hölzle said that every single core could be put to a specific task — 771,886 cores at the time, and that number is set to grow.
To illustrate the power and scale of Google Compute Engine, the VP demonstrated the Cancer Regulome Explorer application used by the Institute for Systems Biology. While the app would normally take about 10 minutes to locate one match, using 10,000 Google processor cores (1,250 8-core VMs) allows a match to be made in mere seconds. Using the same number of cores, the entire computation can be completed in an hour rather than 15 hours.
"That is how infrastructure-as-a-service is supposed to work," Hölzle said.
Taking it a step further, the VP showed that allocating about 600,000 cores to the same genomics app allowed for the discovery of multiple matches every few seconds. In order to take advantage of this many cores, the code was modified ahead of time using Compute Engine's ExaCycle technology. Further details of Google's work with the Institute for System Biology are outlined in this case study.
There's no Google equivalent yet of an Amazon HPC instance type, but it's likely only a matter of time. While Google's offerings are less evolved than the Amazon cloud portfolio, they are not that far apart considering their relative ages.
Google Compute Engine joins the suite of services that comprise the Google Cloud Platform, which also includes Google App Engine, Google BigQuery, Google Cloud Storage, Google Cloud SQL and the Prediction API. These roughly mirror Amazon's Web Services. For example, Google Cloud Storage is similar to Amazon's Simple Storage Service (S3).
Google Compute Engine is currently available in "limited preview," accessible here.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 10, 2013 |
Australian visual effects company, Animal Logic, is considering a move to the public cloud.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
May 08, 2013 |
For engineers looking to leverage high-performance computing, the accessibility of a cloud-based approach is a powerful draw, but there are costs that may not be readily apparent.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/02/2012 | AMD | Developers today are just beginning to explore the potential of heterogeneous computing, but the potential for this new paradigm is huge. This brief article reviews how the technology might impact a range of application development areas, including client experiences and cloud-based data management. As platforms like OpenCL continue to evolve, the benefits of heterogeneous computing will become even more accessible. Use this quick article to jump-start your own thinking on heterogeneous computing.