August 09, 2011
For those who have considered tapping Amazon Web Services for public cloud services but have been hesitant due to security concerns, there might be a new on-ramp to the public cloud.
Amazon announced this week that it would use Equinix datacenters as a private networking hub linked directly to EC2, which offers a more secure way to connect to the cloud by removing the most often-cited “fear factor”—the public internet.
In essence, this means that there are special facilities located in the EC2 infrastructure that have been contracted for by a private company that remain separated from the multi-tenant cloud.
According to Adam Selipsky, VP of Amazon Web Services, “We have heard consistently over time that companies don’t want to use the public Internet for workloads involving compliance-sensitive data.” Aside from helping to alleviate some of these concerns, the concept of a virtual private cloud means that customers will have more control over issues like latency and the amount of bandwidth they use as well as the overall reliability of the services.
He adds that this complements another feature AWS added earlier last week called the Identify and Access Management service, which permits existing identity management systems to be used directly by the AWS account holder. In other words, this gives a company the ability to decide who can use the account and to offer multiple levels of access for accepted users.
Instead of using a VPN or the Internet to move data, users are now able to create their own private network that feeds to the high-speed, private link to the closest AWS datacenter.
As the primary partner for its new Direct Connect service, Amazon chose Equinix, which operates 90 high-performance datacenters on networking hubs in 37 markets around the world.
As Information Week reported this week, “The first operative connection is in Equinix’s Ashburn, Virginia facility, linked to EC2’s U.S. East datacenter in Northern Virginia. By going through Ashburn, customers will have their data and workloads transported into EC2 over all private connections, avoiding any use of the Internet."
Full story at InformationWeek
The ever-growing complexity of scientific and engineering problems continues to pose new computational challenges. Thus, we present a novel federation model that enables end-users with the ability to aggregate heterogeneous resource scale problems. The feasibility of this federation model has been proven, in the context of the UberCloud HPC Experiment, by gathering the most comprehensive information to date on the effects of pillars on microfluid channel flow.
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
Frank Ding, engineering analysis & technical computing manager at Simpson Strong-Tie, discussed the advantages of utilizing the cloud for occasional scientific computing, identified the obstacles to doing so, and proposed workarounds to some of those obstacles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/02/2012 | AMD | Developers today are just beginning to explore the potential of heterogeneous computing, but the potential for this new paradigm is huge. This brief article reviews how the technology might impact a range of application development areas, including client experiences and cloud-based data management. As platforms like OpenCL continue to evolve, the benefits of heterogeneous computing will become even more accessible. Use this quick article to jump-start your own thinking on heterogeneous computing.