December 10, 2007
SANTA CLARA, Calif., Dec. 4 -- Sun Microsystems Inc., today announced availability and pricing details for the Sun xVM Ops Center, a highly scalable datacenter automation tool for complete management of heterogeneous global IT environments. As part of the company's commitment to the open source community, Sun will also release the source code used to build Sun's xVM Ops Center to the OpenxVM.org community under the GNU General Public License version three (GPLv3).
Sun will provide ongoing code releases to OpenxVM.org, starting with the contribution of the Common Agent Container source code on Dec. 10. In addition, the first commercial release of Sun xVM Ops Center is expected to be available on Jan. 8.
A critical part of Sun's virtualization portfolio, Sun xVM Ops Center reduces datacenter management complexity by combining a range of lifecycle management functionality into an all-in-one tool. Sun xVM Ops Center helps simplify discovery, monitoring, operating system provisioning, comprehensive updates and patch management, firmware updates, and hardware management from power up to production in cross-platform Linux and Solaris Operating System-based x86 and SPARC environments. As a result, Sun xVM Ops Center allows enterprises to streamline their datacenter operations, helping to lower costs and more easily manage rapid growth, datacenter consolidation and compliance requirements.
"Business requirements that demand tremendous growth are creating more datacenter complexity, as administrators need to manage increasing volumes of data, applications and people on the network, as well as the rapid proliferation of servers and guest operating systems," said Steve Wilson, vice president of xVM for Sun Microsystems. "Through Sun xVM Ops Center's unique combination of asset discovery, inventory tracking, operating system provisioning, firmware updating and patch management capabilities, customers can automate a number of tasks on a global scale to improve their time to market and productivity to meet ever demanding service level agreements."
Designed to scale for demanding HPC and globally distributed IT environments, Sun xVM Ops Center helps customers to automate their datacenters and manage thousands of assets simultaneously. Through automated monitoring of assets on a network, and implementation of upgrades and patches for operating systems, Sun xVM Ops Center automates routine system administrator tasks in the datacenter. Sun xVM Ops Center addresses provisioning and update requirements to help improve predictability and keep Linux and Solaris instances current, thus improving the security and stability of systems. It gives the system administrator greater control to simulate updates and related dependencies before actually going forward with a change to enable more informed datacenter planning and minimize downtime.
A distribution of Sun xVM Ops Center will be made available via free download. Building on Sun's commitment to open standards and customer choice, Sun will continue to innovate the Sun xVM platform and collaborate with open source communities. The first of Sun's contributions will be the Common Agent Container (CAC) code to the OpenxVM.org community under GPLv3. The CAC is the heart of the management infrastructure for many of Sun's products, including the Sun xVM Ops Center. In addition, Sun plans to make the entire code base used by the xVM Ops Center available to the OpenxVM.org community in the first quarter of 2008.
Sun xVM is the company's open, comprehensive virtualization and management platform, which includes Sun xVM Ops Center and Sun xVM Server, an enterprise-grade bare-metal hypervisor. Sun has committed $2 billion in research and development for the xVM platform and Sun xVM Ops Center is the first product in the Sun xVM family of virtualization solutions to become available.
About Sun Microsystems Inc.
Sun Microsystems develops the technologies that power the global marketplace. Guided by a singular vision -- "The Network is the Computer" -- Sun drives network participation through shared innovation, community development and open source leadership. Sun can be found in more than 100 countries and on the Web at http://sun.com.
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
Frank Ding, engineering analysis & technical computing manager at Simpson Strong-Tie, discussed the advantages of utilizing the cloud for occasional scientific computing, identified the obstacles to doing so, and proposed workarounds to some of those obstacles.
The private industry least likely to adopt public cloud services for data storage are financial institutions. Holding the most sensitive and heavily-regulated of data types, personal financial information, banks and similar institutions are mostly moving towards private cloud services – and doing so at great cost.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 10, 2013 |
Australian visual effects company, Animal Logic, is considering a move to the public cloud.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
May 08, 2013 |
For engineers looking to leverage high-performance computing, the accessibility of a cloud-based approach is a powerful draw, but there are costs that may not be readily apparent.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/02/2012 | AMD | Developers today are just beginning to explore the potential of heterogeneous computing, but the potential for this new paradigm is huge. This brief article reviews how the technology might impact a range of application development areas, including client experiences and cloud-based data management. As platforms like OpenCL continue to evolve, the benefits of heterogeneous computing will become even more accessible. Use this quick article to jump-start your own thinking on heterogeneous computing.