June 28, 2012
Combination results in price performance breakthrough for Hadoop
SAN FRANCISCO, June 28 — MapR Technologies, the provider of the open, enterprise-grade distribution for Apache Hadoop, today announced that MapR will make its distribution for Hadoop available on Google Compute Engine. The combination of the new Google service and the MapR distribution enables customers to quickly provision large MapR clusters on demand and to take advantage of the flexibility, agility and scalability of a cloud-based solution.
Google pioneered MapReduce for their internal search framework, which inspired the community development of Hadoop. Now, through MapR's distribution for Hadoop, organizations are able to benefit from Google's highly available, dependable, and scalable infrastructure for Big Data analytics.
MapR also demonstrated a significant price/performance breakthrough on stage at Google I/O completing a one terabyte (TB) TeraSort job in 1 minute 20 seconds. This result was achieved on a Google Compute Engine cluster with 1256 nodes, 1256 disks and 5024 cores at a cost of $16. This result compares favorably with the existing world record of one minute two seconds that was set with a physical cluster with more than four times the disks, twice as many cores, 200 more servers, and at a cost of more than $5 million.
"Off-premise, on-demand computing is an important part of the future for Hadoop," said John Schroeder, CEO and co-founder of MapR Technologies. "MapR is solidifying that future by partnering with Google and leveraging their cost-effective, high performance and scale-out infrastructure."
The integration of MapR with Google Compute Engine includes a menu of MapR compute configurations where customers can easily store, manage and analyze large volumes of data cost effectively in the cloud. Customers have the flexibility within Google Compute Engine to pay on demand and quickly and easily spin up 1,000+ node clusters.
"To help customers get the most out of our cloud platform products," explains Eric Morse, head of sales and business development, for Google's cloud platform, "we work closely with technology companies that provide powerful complementary solutions integrated with our platform."
Google Compute Engine partners include technology firms, such as software vendors, platform companies, and management and tools vendors, and early collaboration with services firms, such as systems integrators, developers, and IT consultants. These partners offer complementary services, solutions, and technologies that have been integrated to provide customers with powerful new solutions using Google Compute Engine. MapR on Google Compute Engine will be available as a free private beta for a select number of customers. Interested customers can learn more and request access www.mapr.com/google and to learn more about the Google Compute Engine, go to http://googleenterprise.blogspot.com/.
The availability of MapR on the Google Compute Engine will be available as a free private beta for a select number of customers. Interested customers should review and fill out the nomination form.
About MapR Technologies
MapR delivers on the promise of Hadoop, making Big Data management and analysis a reality for more business users. The award-winning MapR Distribution brings unprecedented dependability, speed, and ease-of-use to Hadoop. Providing unparalleled data protection and business continuity, MapR enables customers to harness the power of Big Data analytics. The company is headquartered in San Jose, Calif. Investors include Lightspeed Venture Partners, NEA and Redpoint Ventures. Download the latest MapR Distribution for Apache Hadoop. Connect with MapR onFacebook,Linkedin, and Twitter.
Source: MapR Technologies
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
Frank Ding, engineering analysis & technical computing manager at Simpson Strong-Tie, discussed the advantages of utilizing the cloud for occasional scientific computing, identified the obstacles to doing so, and proposed workarounds to some of those obstacles.
The private industry least likely to adopt public cloud services for data storage are financial institutions. Holding the most sensitive and heavily-regulated of data types, personal financial information, banks and similar institutions are mostly moving towards private cloud services – and doing so at great cost.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 10, 2013 |
Australian visual effects company, Animal Logic, is considering a move to the public cloud.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/02/2012 | AMD | Developers today are just beginning to explore the potential of heterogeneous computing, but the potential for this new paradigm is huge. This brief article reviews how the technology might impact a range of application development areas, including client experiences and cloud-based data management. As platforms like OpenCL continue to evolve, the benefits of heterogeneous computing will become even more accessible. Use this quick article to jump-start your own thinking on heterogeneous computing.