February 25, 2013
Central to the attraction of cloud computing is the "pay for only what you use" claim made by almost all service providers. But look a little deeper within the contracts and you will often find that they are not quite as attractive as first appears.
Hidden away in the small print of many cloud service contracts are a range of charges that customers may not 'spot' until it is too late; that is until very large bills for data processing or data storage have made their eyes bulge. The problem is compounded in that individual users, potentially junior grades, have the potential to rack up significant bills without any clear financial monitoring or control. So suppliers really need to be more transparent and customers need to be more demanding for price clarity, otherwise it will severely limit the growth of the cloud services business.
Here are some pricing examples:
None of the charges are particularly unfair. However, when these different pricing structures are combined and not transparent, it creates a very confusing picture for the customer who has the challenge of implementing controls over the use of the service if he is to deliver the benefits defined in his business case for cloud.
On a related point, data duplication can also affect cloud storage costs and should be taken seriously. It is not uncommon for multiple versions of a file to exist that are mostly identical and are not necessarily required, but are costing the user for their storage. A 2012 study from Johannes Gutenberg University Mainz, the Barcelona Supercomputing Center and University of Hamburg, on high performance computing (HPC) data sets, found typically 20 to 30 percent of this online data could be removed by applying data de-duplication techniques, peaking up to 70 percent for some data sets.
There is a growing concern amongst IT and finance managers over this issue and they are increasingly asking searching questions about fee structures of cloud services. This is filtering down to users; many are hearing the message from employers to take a cautious approach, to be wary of anything that sounds like it might carry the 'cloud services' badge. Naturally we expect users to be cautious and certainly don't want to discourage that level of wariness.
For smaller businesses particularly, the benefits of cloud (e.g., scalability, elasticity, ease-of-use, on-demand access, utility pricing, reduced CAPEX/OPEX, etc.) far outweigh the pricing 'hurdles' of using such a service, but caveat emptor certainly applies here
As with everything, the devil is in the detail. Right now, users should look for a supplier that will enable self-monitoring of their spend in real-time whilst they're using a cloud service, not when the bill is drops onto the doormat. Companies must set in-house policies to control usage by users – it's very easy to get carried away on a cloud service. Users must check the fine print of the contract and ask questions if it is to confusing to understand; look for a supplier that is prepared to explain the billing process, not rush you into signing a contract. Perhaps most importantly, pick a supplier, partner up and let them guide you through the process; plug and play isn't always cheapest in the long run.
About the Author
Joining OCF in October 2010, Jerry Dixon is Compute-on-Demand business development manager responsible for OCF's "enCORE" service. His role encompasses the recruitment of a network of academic and research server clusters to contribute server power to enCORE, such as the Science and Technology Facilities Council's (STFC) Hartree Centre. He works with application software vendors, develops the market and provides expert advice and consultancy to customers. Jerry's background is in managed IT services, having previously worked for Calyx UK and ServiceTec Limited.
Jun 19, 2013 |
Ruan Pethiyagoda, Cameron Boehmer, John S. Dvorak, and Tim Sze, trained at San Francisco’s Hack Reactor, an institute designed for intense fast paced learning of programming, put together a program based on the N-Queens algorithm designed by the University of Cambridge’s Martin Richards, and modified it to run in parallel across multiple machines.
Jun 17, 2013 |
With that in mind, Datapipe hopes to establish themselves as a green-savvy HPC cloud provider with their recently announced Stratosphere platform. Datapipe markets Stratosphere as a green HPC cloud service and in doing so partnering with Verne Global and their Icelandic datacenter, which is known for its propensity in green computing.
Jun 12, 2013 |
Cloud computing is gaining ground in utilization by mid-sized institutions who are looking to expand their experimental high performance computing resources. As such, IBM released what they call Redbooks, in part to assist institutions’ movement of high performance computing applications to the cloud.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/02/2012 | AMD | Developers today are just beginning to explore the potential of heterogeneous computing, but the potential for this new paradigm is huge. This brief article reviews how the technology might impact a range of application development areas, including client experiences and cloud-based data management. As platforms like OpenCL continue to evolve, the benefits of heterogeneous computing will become even more accessible. Use this quick article to jump-start your own thinking on heterogeneous computing.