Welcome!

IoT User Interface Authors: Elizabeth White, Pat Romanski, John Basso, Ram Sonagara, Liz McMillan

News Feed Item

New Editions of Moab HPC Suite Enhance HPC System Efficiency

Adaptive Computing, the largest provider of private cloud management and High-Performance Computing (HPC) workload management software, today announced the release of two new editions of its Moab HPC Suite: Application Portal Edition and Remote Visualization Edition. These two new solutions are designed to leverage next-generation access models to simplify the collection and interpretation of data, improving the time it takes to achieve meaningful results.

Moab HPC Suite – Application Portal Edition

With an integrated NICE EnginFrame application portal, Moab streamlines the process of accessing job information by making it all available from a single point – applications, data, resources and job submissions – thus keeping costs to a minimum throughout the design and research processes. It supports the most common ISV and open-source applications used in a variety of disciplines, including manufacturing, energy, life science, government and education, without requiring users to undergo specialized HPC training. Data is also stored efficiently to minimize the need for file transfers.

It also makes it easier for users to collaborate on jobs, while maintaining security through standard browser access. This allows users both inside and outside the organization to access the data, which is encrypted for transmission. Access can be granted based on user groups, applications or resource parameters.

This edition of Moab also reduces costs by improving the efficiency of accessing and sharing the centralized pool of resources. Its advanced scheduling policies allow maximum HPC resource utilization by more users. Application licenses are also pooled to reduce costs and improve service to users. In addition, policies and usage budgets are enforced to ensure that service level agreements are met and project priorities are maintained. This enables optimized use of the centralized HPC storage capacity.

Moab HPC Suite – Remote Visualization Edition

The new Remote Visualization Edition of Moab allows organizations to take advantage of the ability to consolidate workstations into centralized resources, saving money and making specialized computing resources available to more users. This reduces management costs by keeping overhead low, as well as keeping network congestion at a minimum by removing the need to transmit full data sets. The result is less data transfer time lags for users. Less data transmission in turn reduces energy usage and storage costs as storage is centralized. This approach also minimizes security risks, keeping data sets closer to the centralized computing resources. Its shared pool of resources also decreases software licensing costs.

The Remote Visualization Edition also provides guaranteed shared visualization resource access for users, with automated policies and budgets that enforce priorities and service levels. Moab’s allocating and scheduling policies enable workload packing, allowing optimized utilization of GPUs and other accelerators, as well as multiple Linux user sessions on a single machine. Changing demands within workloads can also be better managed by Moab’s ability to dynamically re-provision the operating system as needed.

Moab HPC Suite – Remote Visualization Edition also includes an integrated NICE EnginFrame Views portal, as well as NICE Desktop Cloud Visualization, allowing secure remote access to cloud sessions with ideal compression, in Linux, Windows or virtual environments.

Moab HPC Suite – Application Portal Edition and Moab HPC Suite – Remove Visualization Edition are available in November from Adaptive Computing as well as through our partner, HP, and HP’s reseller and integrator partners.

Supporting Quotes

“In order for our customers to get the most out of our cluster technology, it’s important that they be able to schedule and allocate resources as effectively as possible,” said Jan Wallenberg, CEO of Go Virtual. “That’s where Moab comes in. Its efficiency makes a significant difference in the user experience and overall performance of the system.”

“We’re excited to see NICE EnginFrame and Desktop Cloud Visualization included in these latest versions of Moab,” said Andrea Rodolico, CTO of NICE. “Integrating our leading-edge HPC portal and remote 3D visualization technology into the framework of Moab makes it an even more capable product, enabling greater usability and productivity for HPC users.”

“These latest editions of Moab demonstrate our continued commitment to improving the user experience of Moab as well as the back-end functionality,” noted Michael Jackson, president of Adaptive Computing. “By integrating the latest technology from other industry leaders into our solutions, we are making HPC systems run more effectively, which means manufacturers and researchers can more quickly bring their discoveries to the world.”

About Adaptive Computing

Adaptive Computing is the largest provider of High-Performance Computing (HPC) workload management software and manages the world’s largest cloud computing environment with Moab, a self-optimizing dynamic cloud management solution and HPC workload management system. Moab®, a patented multidimensional intelligence engine, delivers policy-based governance, allowing customers to consolidate and virtualize resources, allocate and manage applications, optimize service levels and reduce operational costs. Adaptive Computing offers a portfolio of Moab cloud management and Moab HPC workload management products and services that accelerate, automate, and self-optimize IT workloads, resources, and services in large, complex heterogeneous computing environments such as HPC, data centers and cloud. Our products act as a brain on top of existing and future diverse infrastructure and middleware to enable it to self-optimize and deliver higher ROI to the business with its:

Moab Cloud Suite for self-optimizing cloud management

Moab HPC Suite for self-optimizing HPC workload management

For more information, call (801) 717-3700 or visit www.adaptivecomputing.com.

NOTE TO EDITORS: If you would like additional information on Adaptive Computing and its products, please visit the Adaptive Computing Newsroom at http://www.adaptivecomputing.com/category/news/. All prices noted are in U.S. dollars and are valid only in the United States.

More Stories By Business Wire

Copyright © 2009 Business Wire. All rights reserved. Republication or redistribution of Business Wire content is expressly prohibited without the prior written consent of Business Wire. Business Wire shall not be liable for any errors or delays in the content, or for any actions taken in reliance thereon.

@CloudExpo Stories
Basho Technologies has announced the latest release of Basho Riak TS, version 1.3. Riak TS is an enterprise-grade NoSQL database optimized for Internet of Things (IoT). The open source version enables developers to download the software for free and use it in production as well as make contributions to the code and develop applications around Riak TS. Enhancements to Riak TS make it quick, easy and cost-effective to spin up an instance to test new ideas and build IoT applications. In addition to...
Predictive analytics tools monitor, report, and troubleshoot in order to make proactive decisions about the health, performance, and utilization of storage. Most enterprises combine cloud and on-premise storage, resulting in blended environments of physical, virtual, cloud, and other platforms, which justifies more sophisticated storage analytics. In his session at 18th Cloud Expo, Peter McCallum, Vice President of Datacenter Solutions at FalconStor, discussed using predictive analytics to mon...
Organizations planning enterprise data center consolidation and modernization projects are faced with a challenging, costly reality. Requirements to deploy modern, cloud-native applications simultaneously with traditional client/server applications are almost impossible to achieve with hardware-centric enterprise infrastructure. Compute and network infrastructure are fast moving down a software-defined path, but storage has been a laggard. Until now.
Internet of @ThingsExpo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with the 19th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world and ThingsExpo Silicon Valley Call for Papers is now open.
IoT is rapidly changing the way enterprises are using data to improve business decision-making. In order to derive business value, organizations must unlock insights from the data gathered and then act on these. In their session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, and Peter Shashkin, Head of Development Department at EastBanc Technologies, discussed how one organization leveraged IoT, cloud technology and data analysis to improve customer experiences and effi...
SYS-CON Events announced today that Isomorphic Software will exhibit at DevOps Summit at 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Isomorphic Software provides the SmartClient HTML5/AJAX platform, the most advanced technology for building rich, cutting-edge enterprise web applications for desktop and mobile. SmartClient combines the productivity and performance of traditional desktop software with the simp...
"We've discovered that after shows 80% if leads that people get, 80% of the conversations end up on the show floor, meaning people forget about it, people forget who they talk to, people forget that there are actual business opportunities to be had here so we try to help out and keep the conversations going," explained Jeff Mesnik, Founder and President of ContentMX, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
"When you think about the data center today, there's constant evolution, The evolution of the data center and the needs of the consumer of technology change, and they change constantly," stated Matt Kalmenson, VP of Sales, Service and Cloud Providers at Veeam Software, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Let’s face it, embracing new storage technologies, capabilities and upgrading to new hardware often adds complexity and increases costs. In his session at 18th Cloud Expo, Seth Oxenhorn, Vice President of Business Development & Alliances at FalconStor, discussed how a truly heterogeneous software-defined storage approach can add value to legacy platforms and heterogeneous environments. The result reduces complexity, significantly lowers cost, and provides IT organizations with improved efficienc...
With 15% of enterprises adopting a hybrid IT strategy, you need to set a plan to integrate hybrid cloud throughout your infrastructure. In his session at 18th Cloud Expo, Steven Dreher, Director of Solutions Architecture at Green House Data, discussed how to plan for shifting resource requirements, overcome challenges, and implement hybrid IT alongside your existing data center assets. Highlights included anticipating workload, cost and resource calculations, integrating services on both sides...
In his session at @DevOpsSummit at 19th Cloud Expo, Yoseph Reuveni, Director of Software Engineering at Jet.com, will discuss Jet.com's journey into containerizing Microsoft-based technologies like C# and F# into Docker. He will talk about lessons learned and challenges faced, the Mono framework tryout and how they deployed everything into Azure cloud. Yoseph Reuveni is a technology leader with unique experience developing and running high throughput (over 1M tps) distributed systems with extre...
Manufacturers are embracing the Industrial Internet the same way consumers are leveraging Fitbits – to improve overall health and wellness. Both can provide consistent measurement, visibility, and suggest performance improvements customized to help reach goals. Fitbit users can view real-time data and make adjustments to increase their activity. In his session at @ThingsExpo, Mark Bernardo Professional Services Leader, Americas, at GE Digital, discussed how leveraging the Industrial Internet a...
Big Data engines are powering a lot of service businesses right now. Data is collected from users from wearable technologies, web behaviors, purchase behavior as well as several arbitrary data points we’d never think of. The demand for faster and bigger engines to crunch and serve up the data to services is growing exponentially. You see a LOT of correlation between “Cloud” and “Big Data” but on Big Data and “Hybrid,” where hybrid hosting is the sanest approach to the Big Data Infrastructure pro...
In his session at 18th Cloud Expo, Sagi Brody, Chief Technology Officer at Webair Internet Development Inc., and Logan Best, Infrastructure & Network Engineer at Webair, focused on real world deployments of DDoS mitigation strategies in every layer of the network. He gave an overview of methods to prevent these attacks and best practices on how to provide protection in complex cloud platforms. He also outlined what we have found in our experience managing and running thousands of Linux and Unix ...
Cloud analytics is dramatically altering business intelligence. Some businesses will capitalize on these promising new technologies and gain key insights that’ll help them gain competitive advantage. And others won’t. Whether you’re a business leader, an IT manager, or an analyst, we want to help you and the people you need to influence with a free copy of “Cloud Analytics for Dummies,” the essential guide to this explosive new space for business intelligence.
"My role is working with customers, helping them go through this digital transformation. I spend a lot of time talking to banks, big industries, manufacturers working through how they are integrating and transforming their IT platforms and moving them forward," explained William Morrish, General Manager Product Sales at Interoute, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
A critical component of any IoT project is what to do with all the data being generated. This data needs to be captured, processed, structured, and stored in a way to facilitate different kinds of queries. Traditional data warehouse and analytical systems are mature technologies that can be used to handle certain kinds of queries, but they are not always well suited to many problems, particularly when there is a need for real-time insights.
Choosing the right cloud for your workloads is a balancing act that can cost your organization time, money and aggravation - unless you get it right the first time. Economics, speed, performance, accessibility, administrative needs and security all play a vital role in dictating your approach to the cloud. Without knowing the right questions to ask, you could wind up paying for capacity you'll never need or underestimating the resources required to run your applications.
Enterprise networks are complex. Moreover, they were designed and deployed to meet a specific set of business requirements at a specific point in time. But, the adoption of cloud services, new business applications and intensifying security policies, among other factors, require IT organizations to continuously deploy configuration changes. Therefore, enterprises are looking for better ways to automate the management of their networks while still leveraging existing capabilities, optimizing perf...
"Software-defined storage is a big problem in this industry because so many people have different definitions as they see fit to use it," stated Peter McCallum, VP of Datacenter Solutions at FalconStor Software, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.