November 01, 2012
SEATTLE, Nov. 1 — Symform is using humor and prizes to highlight the benefits of its distributed, decentralized and secure cloud storage network with its recently launched 'Byte Me' promotion. The campaign features two fictional IT managers trekking the globe in search of the perfect location to build their next data center. The two characters, Gus and Benny, are shown in the cartoon series visiting the Arctic, farm country, the high desert, and other locations where many cloud providers are building massive data centers today. Users who enter the promotion by downloading the Symform software have a chance to win a network attached storage (NAS) appliance, and one in four entrants receives a 'Byte Me' t-shirt.
"We are not anti-datacenter, but we do want people to think about all the excess capacity that sits wasted in data centers, networks and computing devices around the world that could be utilized, as well as illustrate how we literally cannot build enough data centers to store all the digital data being created," said Margaret Dawson, vice president of product management and marketing at Symform. "We speak seriously about the data center behind every cloud and how it behooves all of us to start thinking about distributed and decentralized approaches to IT and cloud computing, and this is just a way of having some fun with the topic and giving a chance for users to win great prizes."
The topic of data center costs and energy utilization has been a popular topic lately, highlighted in both a comprehensive report released by Greenpeace as well as a recent New York Times article titled, "Power, Pollution and the Internet." The Greenpeace report suggested data centers used for cloud computing account for about 2 percent of the world's carbon footprint, while the NYT article claimed data centers worldwide use about 30 billion watts of electricity, roughly equivalent to the output of 30 nuclear power plants. The article also cites Hank Seader, managing principal for research and education at the Uptime Institute, who says using the cloud "just changes where the applications are running. It all goes to a data center somewhere."
Symform offers the industry's first free and unlimited cloud backup service with their 'Bytes or Bucks' pricing, which enables customers to pay with 'Bytes' by contributing excess local drive space in exchange for secure cloud backup. Alternatively, customers can pay with 'Bucks' at the low cost of $0.15 per GB/month. With its patented technology and quickly-growing user base, Symform is building the world's largest virtual datacenter by leveraging contributed local storage space to power a global crowdsourced cloud storage network.
Today, Symform has customers in more than 150 countries, who are storing more than 5.5 billion data fragments across the cloud network.
The Symform Cloud Storage Network is a better way to store and backup all of your data. As the world's first distributed and crowdsourced data backup solution, Symform enables users to pay with bytes instead of bucks. Every business on the network contributes excess local drive space to the grid in exchange for secure, fast and reliable cloud data backup. Before data leaves the source device, it is encrypted and shredded, redundancy added, and then geo-distributed across the global network. With its proprietary and patented technology, Symform is building the world's largest virtual datacenter using existing Internet infrastructure.
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
Frank Ding, engineering analysis & technical computing manager at Simpson Strong-Tie, discussed the advantages of utilizing the cloud for occasional scientific computing, identified the obstacles to doing so, and proposed workarounds to some of those obstacles.
The private industry least likely to adopt public cloud services for data storage are financial institutions. Holding the most sensitive and heavily-regulated of data types, personal financial information, banks and similar institutions are mostly moving towards private cloud services – and doing so at great cost.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 10, 2013 |
Australian visual effects company, Animal Logic, is considering a move to the public cloud.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/02/2012 | AMD | Developers today are just beginning to explore the potential of heterogeneous computing, but the potential for this new paradigm is huge. This brief article reviews how the technology might impact a range of application development areas, including client experiences and cloud-based data management. As platforms like OpenCL continue to evolve, the benefits of heterogeneous computing will become even more accessible. Use this quick article to jump-start your own thinking on heterogeneous computing.