March 08, 2013
LUXEMBOURG, Grand Duchy of Luxembourg, March 8 — NoMachine, a leading provider of remote access and hosted desktop delivery solutions for Linux and other platforms, announced that it has established its new headquarters in Luxembourg, with further offices planned for Riga and Poland.
The new headquarters will be a strategic centre for NoMachine and will have a central role in supporting its international growth. This is the latest opening of a number of new sites as NoMachine establishes itself across Europe. An R&D site in Georgia was opened in January 2012 with the aim of expanding development activities in Tbilisi.
Originating as an early player in open source technology, NoMachine offers remote access solutions for accessing files and applications on Linux as well as Windows, Mac and other platforms. These solutions include the top-ranked NX Server for Linux, known world-wide for its terminal server and application delivery capabilities used by many of the biggest organizations to provide employees with cross-platform remote access across continents. Its upcoming product “NoMachine” for Windows and Mac, a desktop sharing, multimedia streaming and file-transfer tool will expand that portfolio to consumers looking for an easy-to-use product that offers extensive functionality and speed: one of the main traits of all of NoMachine's remote computing software is its exclusive compression technology bringing a local-desktop experience to the user even though their desktop might be thousands of miles away.
Gaspare Tripi, General Manager of NoMachine Sàrl said: “NoMachine is very excited to be expanding not just its customer base through new product offerings, but also its offices. Putting NoMachine's global headquarters in Luxembourg was an obvious choice due to its central position between NoMachine's offices already in place, and the country’s successful track record in technology start-ups.”
NoMachine is the creator of NoMachine (NX) software, an enterprise-class solution for secure remote access, multi-platform application delivery, and hosted desktop deployment. NoMachine revolutionizes the way users access their computing resources across the Internet to make desktop access as easy and widespread as Web browsing. Regardless of location or the network environment, NoMachine's software provides real-time performance in remote sessions, automatically tuning itself to network bandwidth and latency parameters. It's upcoming version 4 will extend its renowned remote desktop capabilities for companies with Linux environments to the consumer, offering Windows and Mac users an intuitive remote access and desktop sharing solution with multimedia support, record and playback, file-transfer, bidirectional printing and much more.
The ever-growing complexity of scientific and engineering problems continues to pose new computational challenges. Thus, we present a novel federation model that enables end-users with the ability to aggregate heterogeneous resource scale problems. The feasibility of this federation model has been proven, in the context of the UberCloud HPC Experiment, by gathering the most comprehensive information to date on the effects of pillars on microfluid channel flow.
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
Frank Ding, engineering analysis & technical computing manager at Simpson Strong-Tie, discussed the advantages of utilizing the cloud for occasional scientific computing, identified the obstacles to doing so, and proposed workarounds to some of those obstacles.
May 23, 2013 |
he study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 10, 2013 |
Australian visual effects company, Animal Logic, is considering a move to the public cloud.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/02/2012 | AMD | Developers today are just beginning to explore the potential of heterogeneous computing, but the potential for this new paradigm is huge. This brief article reviews how the technology might impact a range of application development areas, including client experiences and cloud-based data management. As platforms like OpenCL continue to evolve, the benefits of heterogeneous computing will become even more accessible. Use this quick article to jump-start your own thinking on heterogeneous computing.