Click here to close now.

Welcome!

Linux Authors: Mike Kavis, Ian Khan, VictorOps Blog, Bart Copeland, Carmen Gonzalez

Related Topics: Linux

Linux: Article

Case Study: Biocomputing Solutions Using Linux Networx Cluster System

Tulane University is home to the Center for Computational Science (CCS), a unique facility

Tulane University is home to the Center for Computational Science (CCS), a unique facility designed to provide computational resources for research projects across many disciplines. The Center provides an infrastructure for investigators interested in computational science to exchange ideas, produce research, and establish new collaborations.

The Challenge
One of these collaborative efforts involves a team of researchers performing computational simulations of multi-scale models in biological systems. This research is funded through individual research grants from the National Institute of Health (NIH), National Science Foundation (NSF), and National Aeronautics and Space Administration (NASA), as well as a Center grant from the NIH entitled "Biocomputing: Integrating Molecular/Organ-Level Function."

"Since computationally demanding simulations of 3D models of bioelectric phenomena, fluid-structure interactions, and molecular dynamics simulation are important to these research projects, it was important for the CCS to obtain a scalable high-performance computing system that could be easily shared among several different research groups," said Dr. Donald Gaver, director of the CCS.

For example, Dr. Natalia Trayanova, professor of biomedical engineering, studies cardiac defibrillation - the application of a strong electric shock to the heart to restore cardiac rhythm and prevent sudden cardiac death. Despite the critical role that defibrillation therapy plays in saving human life, it is only known what happens on the surface of the heart during these electric shocks. By using computer models, Dr. Trayanova is able to produce a 3D simulation of a heart to better understand what happens during a shock within the cardiac walls. This improved understanding is expected to lead to new advancements and optimization of the clinical procedure of cardiac defibrillation.

"Computer models and 3D simulations are imperative to my research to better understand how electric shock affects the heart," said Dr. Trayanova. "Therefore, it is crucial that the computing system we use be capable of handling compute-intensive workloads and simulations."

Tulane's existing biomedical server was a shared memory machine, but the cost of upgrading and maintaining the system was already quite costly - Tulane knew expanding this existing system would be too expensive.

"To achieve successful, accurate results in a timely manner for our researchers' projects, we needed a computing system that was fast, reliable, but affordable," said Rene Salmon, senior systems analyst for CCS. "Since the outcome of so many projects would depend on the reliability and ease-of-use of the system, we had to seriously consider what type of computing system would benefit our users the most."

Investigating Linux Clusters
Tulane started looking for computing alternatives that could fit within their budget and provide the speed and reliability that was crucial for ensuring successful research results. After investigating several alternatives, Tulane became increasingly interested in Linux clusters. This distributed computing platform seemed to provide the price/performance ratio Tulane was interested in achieving. However, the researchers at the CCS were worried about the difficulty of migrating to a Linux cluster.

"With a Linux cluster we could afford many more CPUs, which allowed us to run much larger simulations and get faster results at a fraction of the cost," said Salmon. "However, we were concerned about what would be involved in migrating our codes to a new system. Luckily, our researches were willing to adapt their codes to run on a Linux cluster."

The scalability of Linux clusters was another attractive feature for Tulane. As researchers received more grant money from various sources, they could add more compute nodes to the cluster and increase its computing power.

"The great thing about the scalability of Linux clusters is that it allows our researchers to pool their resources so they can have access to more CPUs and computing power," said Salmon.

Managing the Linux Cluster
Since the Linux cluster would be running multiple programs for different users, it was imperative that the computing system Tulane chose be easy to use and manage. Tulane was drawn to the comprehensive management tools offered by Linux Networx, especially the Icebox management appliance and LinuxBIOS.

"Since we knew the cluster would be running multiple programs for several projects simultaneously, how we would manage the cluster became a big concern," said Salmon. "After looking at Linux Networx's management tools, we were very impressed with the capabilities of Icebox. The serial terminal server and a remote-controlled power distribution were also critical for ensuring a productive, scalable system."

Linux Networx's expertise with LinuxBIOS, an open source BIOS alternative, was another management feature that interested Tulane. LinuxBIOS performs the same basic functions as commercial BIOS only 10-20 times faster. LinuxBIOS initializes the hardware, checks for valid memory, and begins loading the operating system in about three seconds. Most commercial BIOS require about 30-60 seconds to perform the same tasks. In addition, LinuxBIOS can be configured and accessed from within the Linux operating system. This means changes to the BIOS can be made remotely to a single node or to all the nodes in a cluster system.

"LinuxBIOS was important to us for scalability reasons. We really liked the fact that we could make changes to the BIOS on the entire cluster from Linux with just a few commands," said Salmon.

With expertise in LinuxBIOS, comprehensive management tools, and powerful computing architecture, Linux Networx became Tulane's choice for their first cluster implementation.

"We wanted a Linux cluster vendor that could provide us with an easy-to-manage cluster that could scale and grow easily. Linux Networx did this for us," said Salmon.

Implementing a Linux Networx Cluster System
Starting with a 20-processor Evolocity system, Tulane researchers quickly noticed the benefits of Linux Networx cluster technology and have scaled up their cluster four times as additional grant money has been received. Their Evolocity cluster now totals over 80 AMD Opteron processors.

"The scalability of the Evolocity cluster has proved to be an essential feature as we've quadrupled the cluster's power since we first installed it," said Salmon. "Scaling the cluster has also been a fairly easy process as programs and codes were easily migrated."

Results
Tulane researchers quickly noticed the benefits and potential of Linux cluster technology. In the past, researchers were limited to running small simulations, now they can run much larger, complex simulations than was possible before.

"Running large, detailed simulations is essential to our research efforts as it allows us to better understand biomedical functions," said Dr. Trayanova. "Additionally, a quick turnaround time for running simulations is essential as there are several research groups that need time on the cluster to complete their simulations."

Most important, with the power and capabilities of the Linux Networx system, Tulane researchers can improve the quality of their research and achieve much more focused and valid results than they were able to accomplish previously.

"By completing these jobs in a timely manner, researchers can get results faster, which helps advance the entire mission of the biocomputing program and allows us to do better research," said Dr. Trayanova.

More Stories By Mark R. Hinkle

Mark Hinkle is the Senior Director, Open Soure Solutions at Citrix. He also is along-time open source expert and advocate. He is a co-founder of both the Open Source Management Consortium and the Desktop Linux Consortium. He has served as Editor-in-Chief for both LinuxWorld Magazine and Enterprise Open Source Magazine. Hinkle is also the author of the book, "Windows to Linux Business Desktop Migration" (Thomson, 2006). His blog on open source, technology, and new media can be found at http://www.socializedsoftware.com.

Comments (2) View Comments

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


Most Recent Comments
ITSG News Desk 11/07/05 11:13:54 AM EST

Tulane University is home to the Center for Computational Science (CCS), a unique facility designed to provide computational resources for research projects across many disciplines. The Center provides an infrastructure for investigators interested in computational science to exchange ideas, produce research, and establish new collaborations.

LinuxWorld News Desk 11/07/05 11:00:56 AM EST

Tulane University is home to the Center for Computational Science (CCS), a unique facility designed to provide computational resources for research projects across many disciplines. The Center provides an infrastructure for investigators interested in computational science to exchange ideas, produce research, and establish new collaborations.

@ThingsExpo Stories
The industrial software market has treated data with the mentality of “collect everything now, worry about how to use it later.” We now find ourselves buried in data, with the pervasive connectivity of the (Industrial) Internet of Things only piling on more numbers. There’s too much data and not enough information. In his session at @ThingsExpo, Bob Gates, Global Marketing Director, GE’s Intelligent Platforms business, to discuss how realizing the power of IoT, software developers are now focused on understanding how industrial data can create intelligence for industrial operations. Imagine ...
Operational Hadoop and the Lambda Architecture for Streaming Data Apache Hadoop is emerging as a distributed platform for handling large and fast incoming streams of data. Predictive maintenance, supply chain optimization, and Internet-of-Things analysis are examples where Hadoop provides the scalable storage, processing, and analytics platform to gain meaningful insights from granular data that is typically only valuable from a large-scale, aggregate view. One architecture useful for capturing and analyzing streaming data is the Lambda Architecture, representing a model of how to analyze rea...
SYS-CON Events announced today that Vitria Technology, Inc. will exhibit at SYS-CON’s @ThingsExpo, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. Vitria will showcase the company’s new IoT Analytics Platform through live demonstrations at booth #330. Vitria’s IoT Analytics Platform, fully integrated and powered by an operational intelligence engine, enables customers to rapidly build and operationalize advanced analytics to deliver timely business outcomes for use cases across the industrial, enterprise, and consumer segments.
The explosion of connected devices / sensors is creating an ever-expanding set of new and valuable data. In parallel the emerging capability of Big Data technologies to store, access, analyze, and react to this data is producing changes in business models under the umbrella of the Internet of Things (IoT). In particular within the Insurance industry, IoT appears positioned to enable deep changes by altering relationships between insurers, distributors, and the insured. In his session at @ThingsExpo, Michael Sick, a Senior Manager and Big Data Architect within Ernst and Young's Financial Servi...
SYS-CON Events announced today that Open Data Centers (ODC), a carrier-neutral colocation provider, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place June 9-11, 2015, at the Javits Center in New York City, NY. Open Data Centers is a carrier-neutral data center operator in New Jersey and New York City offering alternative connectivity options for carriers, service providers and enterprise customers.
The explosion of connected devices / sensors is creating an ever-expanding set of new and valuable data. In parallel the emerging capability of Big Data technologies to store, access, analyze, and react to this data is producing changes in business models under the umbrella of the Internet of Things (IoT). In particular within the Insurance industry, IoT appears positioned to enable deep changes by altering relationships between insurers, distributors, and the insured. In his session at @ThingsExpo, Michael Sick, a Senior Manager and Big Data Architect within Ernst and Young's Financial Servi...
PubNub on Monday has announced that it is partnering with IBM to bring its sophisticated real-time data streaming and messaging capabilities to Bluemix, IBM’s cloud development platform. “Today’s app and connected devices require an always-on connection, but building a secure, scalable solution from the ground up is time consuming, resource intensive, and error-prone,” said Todd Greene, CEO of PubNub. “PubNub enables web, mobile and IoT developers building apps on IBM Bluemix to quickly add scalable realtime functionality with minimal effort and cost.”
Sensor-enabled things are becoming more commonplace, precursors to a larger and more complex framework that most consider the ultimate promise of the IoT: things connecting, interacting, sharing, storing, and over time perhaps learning and predicting based on habits, behaviors, location, preferences, purchases and more. In his session at @ThingsExpo, Tom Wesselman, Director of Communications Ecosystem Architecture at Plantronics, will examine the still nascent IoT as it is coalescing, including what it is today, what it might ultimately be, the role of wearable tech, and technology gaps stil...
With several hundred implementations of IoT-enabled solutions in the past 12 months alone, this session will focus on experience over the art of the possible. Many can only imagine the most advanced telematics platform ever deployed, supporting millions of customers, producing tens of thousands events or GBs per trip, and hundreds of TBs per month. With the ability to support a billion sensor events per second, over 30PB of warm data for analytics, and hundreds of PBs for an data analytics archive, in his session at @ThingsExpo, Jim Kaskade, Vice President and General Manager, Big Data & Ana...
In the consumer IoT, everything is new, and the IT world of bits and bytes holds sway. But industrial and commercial realms encompass operational technology (OT) that has been around for 25 or 50 years. This grittier, pre-IP, more hands-on world has much to gain from Industrial IoT (IIoT) applications and principles. But adding sensors and wireless connectivity won’t work in environments that demand unwavering reliability and performance. In his session at @ThingsExpo, Ron Sege, CEO of Echelon, will discuss how as enterprise IT embraces other IoT-related technology trends, enterprises with i...
When it comes to the Internet of Things, hooking up will get you only so far. If you want customers to commit, you need to go beyond simply connecting products. You need to use the devices themselves to transform how you engage with every customer and how you manage the entire product lifecycle. In his session at @ThingsExpo, Sean Lorenz, Technical Product Manager for Xively at LogMeIn, will show how “product relationship management” can help you leverage your connected devices and the data they generate about customer usage and product performance to deliver extremely compelling and reliabl...
The Internet of Things (IoT) is causing data centers to become radically decentralized and atomized within a new paradigm known as “fog computing.” To support IoT applications, such as connected cars and smart grids, data centers' core functions will be decentralized out to the network's edges and endpoints (aka “fogs”). As this trend takes hold, Big Data analytics platforms will focus on high-volume log analysis (aka “logs”) and rely heavily on cognitive-computing algorithms (aka “cogs”) to make sense of it all.
One of the biggest impacts of the Internet of Things is and will continue to be on data; specifically data volume, management and usage. Companies are scrambling to adapt to this new and unpredictable data reality with legacy infrastructure that cannot handle the speed and volume of data. In his session at @ThingsExpo, Don DeLoach, CEO and president of Infobright, will discuss how companies need to rethink their data infrastructure to participate in the IoT, including: Data storage: Understanding the kinds of data: structured, unstructured, big/small? Analytics: What kinds and how responsiv...
Since 2008 and for the first time in history, more than half of humans live in urban areas, urging cities to become “smart.” Today, cities can leverage the wide availability of smartphones combined with new technologies such as Beacons or NFC to connect their urban furniture and environment to create citizen-first services that improve transportation, way-finding and information delivery. In her session at @ThingsExpo, Laetitia Gazel-Anthoine, CEO of Connecthings, will focus on successful use cases.
Sensor-enabled things are becoming more commonplace, precursors to a larger and more complex framework that most consider the ultimate promise of the IoT: things connecting, interacting, sharing, storing, and over time perhaps learning and predicting based on habits, behaviors, location, preferences, purchases and more. In his session at @ThingsExpo, Tom Wesselman, Director of Communications Ecosystem Architecture at Plantronics, will examine the still nascent IoT as it is coalescing, including what it is today, what it might ultimately be, the role of wearable tech, and technology gaps stil...
The true value of the Internet of Things (IoT) lies not just in the data, but through the services that protect the data, perform the analysis and present findings in a usable way. With many IoT elements rooted in traditional IT components, Big Data and IoT isn’t just a play for enterprise. In fact, the IoT presents SMBs with the prospect of launching entirely new activities and exploring innovative areas. CompTIA research identifies several areas where IoT is expected to have the greatest impact.
Wearable devices have come of age. The primary applications of wearables so far have been "the Quantified Self" or the tracking of one's fitness and health status. We propose the evolution of wearables into social and emotional communication devices. Our BE(tm) sensor uses light to visualize the skin conductance response. Our sensors are very inexpensive and can be massively distributed to audiences or groups of any size, in order to gauge reactions to performances, video, or any kind of presentation. In her session at @ThingsExpo, Jocelyn Scheirer, CEO & Founder of Bionolux, will discuss ho...
SYS-CON Events announced today that GENBAND, a leading developer of real time communications software solutions, has been named “Silver Sponsor” of SYS-CON's WebRTC Summit, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. The GENBAND team will be on hand to demonstrate their newest product, Kandy. Kandy is a communications Platform-as-a-Service (PaaS) that enables companies to seamlessly integrate more human communications into their Web and mobile applications - creating more engaging experiences for their customers and boosting collaboration and productiv...
Roberto Medrano, Executive Vice President at SOA Software, had reached 30,000 page views on his home page - http://RobertoMedrano.SYS-CON.com/ - on the SYS-CON family of online magazines, which includes Cloud Computing Journal, Internet of Things Journal, Big Data Journal, and SOA World Magazine. He is a recognized executive in the information technology fields of SOA, internet security, governance, and compliance. He has extensive experience with both start-ups and large companies, having been involved at the beginning of four IT industries: EDA, Open Systems, Computer Security and now SOA.
From telemedicine to smart cars, digital homes and industrial monitoring, the explosive growth of IoT has created exciting new business opportunities for real time calls and messaging. In his session at @ThingsExpo, Ivelin Ivanov, CEO and Co-Founder of Telestax, shared some of the new revenue sources that IoT created for Restcomm – the open source telephony platform from Telestax. Ivelin Ivanov is a technology entrepreneur who founded Mobicents, an Open Source VoIP Platform, to help create, deploy, and manage applications integrating voice, video and data. He is the co-founder of TeleStax, a...