Welcome!

Open Source Authors: Pat Romanski, Elizabeth White, Carmen Gonzalez, Liz McMillan, Plutora Blog

News Feed Item

Texas A&M System Teams with IBM to Drive Computational Sciences Research through Big Data and Analytics

High performance computing (HPC) system will speed research to advance energy resource management, accelerate materials development, ensure the sustainability of food supplies, and improve animal health

COLLEGE STATION, Texas and ARMONK, N.Y., Jan. 29, 2014 /PRNewswire/ -- Texas A&M University System and IBM (NYSE: IBM) today announced an agreement that is the beginning of a broad research collaboration supported by one of the largest computational sciences infrastructure dedicated to advances in agriculture, geosciences and engineering.

(Logo:  http://photos.prnewswire.com/prnh/20090416/IBMLOGO )

The collaboration will leverage the power of big data analytics and high performance computing (HPC) systems for innovative solutions across a spectrum of challenges, such as improving extraction of Earth-based energy resources, facilitating the smart energy grid, accelerating materials development, improving disease identification and tracking in animals, and fostering better understanding and monitoring of our global food supplies.

"Combining the incredible intellectual and technological resources of Texas A&M University and IBM will further position Texas as a leader in identifying and solving some of the most complex challenges we face," Texas Gov. Rick Perry said. "The work that will be done here will change lives and potentially save lives not just in our state, but our nation and around the world."

IBM will provide the infrastructure for the joint research consisting of Blue Gene/Q technology, Power and System x servers, and General Parallel File Systems (GPFS) Storage Systems. A test of the Blue Gene/Q on campus found that it ran a material sciences problem that previously took weeks to solve and produced a solution in "a fraction of an hour" with much greater analytical depth.

"The Texas A&M System and IBM share a passion and a commitment to research that identifies practical solutions to global challenges," said Chancellor John Sharp, Texas A&M University System. "As the largest research university in the state, this agreement is a major step forward for the A&M System in research computing power. This brings together the best computer scientists and technology in the world to focus on issues so important to our role as a leading research institution and to our land-grant mission of serving the state while also providing resources to serve the greater good throughout the world."

IBM Research and the A&M System intend to align skills, assets and resources to pursue fundamental research, applied development, educational reach and sustainable commercial activities with projects that may include:

  • Sustainable Availability of Food: Efficiently providing sufficient food for a growing global population
  • Disease Spread Tracking, Modeling and Prediction: Early and accurate detection and prediction of infectious disease spread to allow the design, testing and manufacturing of medical countermeasures
  • Energy Resource Management: Responsibly explore, extract, and deliver energy resources
  • New Materials Development: Atomic-level modeling, design and testing of new materials for advanced applications in energy, aerospace, structural and defense applications

As a premier engineering research agency of Texas, Texas A&M Engineering Experiment Station (TEES), which conducts research to provide practical answers to critical state and national needs, will be heavily involved from the Texas A&M University System and according to Katherine Banks, Director of TEES and Vice Chancellor of Engineering, "This is a unique opportunity to meet the needs of engineering, geosciences and agriculture and life sciences researchers to expand in areas not feasible before with small-scale HPC systems."

"IBM and the Texas A&M System have crafted a unique collaboration that could apply computational science and big data analytics to some of the most daunting problems in agriculture, geosciences and engineering," said William LaFontaine, Vice President of High Performance Analytics and Cognitive Markets at IBM. "With the combined research capabilities of both institutions and ready access to state-of-the-art computing technology, we feel this collaboration could produce significant scientific insights leading to industry-changing solutions and material economic impact. We are extremely pleased to be engaged with such extraordinarily capable institutions in the A&M System and look forward to years of discovery and innovation."

TEES partners with academic institutions, governmental agencies, industries, and communities to solve problems to help improve the quality of life, promote economic development, and enhance the educational systems of Texas. It is intimately connected with the College of Engineering of Texas A&M University, which is undergoing an unprecedented growth to become a College with 25,000 students by the year 2025 and hire a new generation of faculty who will be addressing the Nation's needs for research and technology development. 

In support of the long-term research effort, IBM will supply to the A&M System cutting edge technical computing technologies, which will be cloud-enabled. The A&M System will deploy a research computing cloud that will comprise of IBM hardware and software including:

  • Blue Gene/Q: Serving as the foundation of the computing infrastructure, a Blue Gene/Q system consisting of two racks, with more than 2,000 compute nodes, will provide 418 teraflops (TF) of sustained performance for big data analytics, complex modeling, and simulation of molecular dynamics, protein folding and organ modeling.
  • Power Systems: A total of 75 PowerLinux 7R2 servers with POWER7+ microprocessors will be connected by 10GbE into a system optimized for big data and analytics and high performance computing. This complex includes IBM BigInsights and Platform Symphony software, IBM Platform LSF scheduler, and IBM General Parallel File System.
  • System x: The solution will contain an estimated 900 IBM System x dense hyperscale compute nodes as part of an IBM NeXtScale system. Some of the nodes will be managed by Platform Cluster Manager Advanced Edition (PCM-AE) as a University-wide HPC cloud while the others will be managed by Platform Cluster Manager Standard Edition (PCM-SE) and serve as a general purpose compute infrastructure for the geosciences and open source analytics initiatives.
  • Platform Computing: Platform Computing software will be used to manage and accelerate various computational workloads. Platform Symphony will drive big data and analytics, and Platform LSF will drive traditional HPC and technical computing workloads. Platform Computing will also power the creation of an HPC cloud, allowing users within the A&M System access to the system.
  • General Parallel File System (GPFS): Five IBM System x GPFS Storage Servers (GSS) will provide five petabytes (PB) of shared storage for use by the compute building blocks using high-speed networks. GPFS will also include an IBM FlashSystem 820 tier with 10 terabytes (TB) of flash storage, delivering performance to accelerate computation for use primarily by Texas A&M Agrilife Research, Geosciences and university HPC as a part of the research computing infrastructure.

Furthermore, IBM will work with researchers at the A&M System to assess new computing technologies that will be necessary to advance data-driven science discovery and innovation over the next several years.

About IBM
For more information on IBM Research visit www.research.ibm.com.
For more information on IBM Technical Computing visit www.ibm.com/systems/technicalcomputing/.

About the A&M System
The A&M System is one of the largest systems of higher education in the nation, with a budget of $3.5 billion. Through a statewide network of 11 universities, seven state agencies, two service units, a comprehensive health science center and a system administration office, the A&M System educates more than 125,000 students and makes more than 22 million additional educational contacts through service and outreach programs each year. Externally funded research expenditures exceed $780 million and help drive the state's economy.

Contact:
Ciri Haugh
617-693-2345
[email protected]

SOURCE IBM

More Stories By PR Newswire

Copyright © 2007 PR Newswire. All rights reserved. Republication or redistribution of PRNewswire content is expressly prohibited without the prior written consent of PRNewswire. PRNewswire shall not be liable for any errors or delays in the content, or for any actions taken in reliance thereon.

@ThingsExpo Stories
The Industrial Internet revolution is now underway, enabled by connected machines and billions of devices that communicate and collaborate. The massive amounts of Big Data requiring real-time analysis is flooding legacy IT systems and giving way to cloud environments that can handle the unpredictable workloads. Yet many barriers remain until we can fully realize the opportunities and benefits from the convergence of machines and devices with Big Data and the cloud, including interoperability, data security and privacy.
Wearable devices have come of age. The primary applications of wearables so far have been "the Quantified Self" or the tracking of one's fitness and health status. We propose the evolution of wearables into social and emotional communication devices. Our BE(tm) sensor uses light to visualize the skin conductance response. Our sensors are very inexpensive and can be massively distributed to audiences or groups of any size, in order to gauge reactions to performances, video, or any kind of presentation. In her session at @ThingsExpo, Jocelyn Scheirer, CEO & Founder of Bionolux, will discuss ho...
DevOps Summit 2015 New York, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that it is now accepting Keynote Proposals. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long development cycles that produce software that is obsolete at launch. DevOps may be disruptive, but it is essential.
The 3rd International Internet of @ThingsExpo, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that its Call for Papers is now open. The Internet of Things (IoT) is the biggest idea since the creation of the Worldwide Web more than 20 years ago.
Connected devices and the Internet of Things are getting significant momentum in 2014. In his session at Internet of @ThingsExpo, Jim Hunter, Chief Scientist & Technology Evangelist at Greenwave Systems, examined three key elements that together will drive mass adoption of the IoT before the end of 2015. The first element is the recent advent of robust open source protocols (like AllJoyn and WebRTC) that facilitate M2M communication. The second is broad availability of flexible, cost-effective storage designed to handle the massive surge in back-end data in a world where timely analytics is e...

ARMONK, N.Y., Nov. 20, 2014 /PRNewswire/ --  IBM (NYSE: IBM) today announced that it is bringing a greater level of control, security and flexibility to cloud-based application development and delivery with a single-tenant version of Bluemix, IBM's platform-as-a-service. The new platform enables developers to build ap...

Building low-cost wearable devices can enhance the quality of our lives. In his session at Internet of @ThingsExpo, Sai Yamanoor, Embedded Software Engineer at Altschool, provided an example of putting together a small keychain within a $50 budget that educates the user about the air quality in their surroundings. He also provided examples such as building a wearable device that provides transit or recreational information. He then reviewed the resources available to build wearable devices at home including open source hardware, the raw materials required and the options available to power s...
The Internet of Things promises to transform businesses (and lives), but navigating the business and technical path to success can be difficult to understand. In his session at @ThingsExpo, Sean Lorenz, Technical Product Manager for Xively at LogMeIn, demonstrated how to approach creating broadly successful connected customer solutions using real world business transformation studies including New England BioLabs and more.
Since 2008 and for the first time in history, more than half of humans live in urban areas, urging cities to become “smart.” Today, cities can leverage the wide availability of smartphones combined with new technologies such as Beacons or NFC to connect their urban furniture and environment to create citizen-first services that improve transportation, way-finding and information delivery. In her session at @ThingsExpo, Laetitia Gazel-Anthoine, CEO of Connecthings, will focus on successful use cases.
Enthusiasm for the Internet of Things has reached an all-time high. In 2013 alone, venture capitalists spent more than $1 billion dollars investing in the IoT space. With "smart" appliances and devices, IoT covers wearable smart devices, cloud services to hardware companies. Nest, a Google company, detects temperatures inside homes and automatically adjusts it by tracking its user's habit. These technologies are quickly developing and with it come challenges such as bridging infrastructure gaps, abiding by privacy concerns and making the concept a reality. These challenges can't be addressed w...
The Domain Name Service (DNS) is one of the most important components in networking infrastructure, enabling users and services to access applications by translating URLs (names) into IP addresses (numbers). Because every icon and URL and all embedded content on a website requires a DNS lookup loading complex sites necessitates hundreds of DNS queries. In addition, as more internet-enabled ‘Things' get connected, people will rely on DNS to name and find their fridges, toasters and toilets. According to a recent IDG Research Services Survey this rate of traffic will only grow. What's driving t...
The Internet of Things is a misnomer. That implies that everything is on the Internet, and that simply should not be - especially for things that are blurring the line between medical devices that stimulate like a pacemaker and quantified self-sensors like a pedometer or pulse tracker. The mesh of things that we manage must be segmented into zones of trust for sensing data, transmitting data, receiving command and control administrative changes, and peer-to-peer mesh messaging. In his session at @ThingsExpo, Ryan Bagnulo, Solution Architect / Software Engineer at SOA Software, focused on desi...
"For over 25 years we have been working with a lot of enterprise customers and we have seen how companies create applications. And now that we have moved to cloud computing, mobile, social and the Internet of Things, we see that the market needs a new way of creating applications," stated Jesse Shiah, CEO, President and Co-Founder of AgilePoint Inc., in this SYS-CON.tv interview at 15th Cloud Expo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
The Internet of Things is tied together with a thin strand that is known as time. Coincidentally, at the core of nearly all data analytics is a timestamp. When working with time series data there are a few core principles that everyone should consider, especially across datasets where time is the common boundary. In his session at Internet of @ThingsExpo, Jim Scott, Director of Enterprise Strategy & Architecture at MapR Technologies, discussed single-value, geo-spatial, and log time series data. By focusing on enterprise applications and the data center, he will use OpenTSDB as an example t...
The industrial software market has treated data with the mentality of “collect everything now, worry about how to use it later.” We now find ourselves buried in data, with the pervasive connectivity of the (Industrial) Internet of Things only piling on more numbers. There’s too much data and not enough information. In his session at @ThingsExpo, Bob Gates, Global Marketing Director, GE’s Intelligent Platforms business, to discuss how realizing the power of IoT, software developers are now focused on understanding how industrial data can create intelligence for industrial operations. Imagine ...
Cultural, regulatory, environmental, political and economic (CREPE) conditions over the past decade are creating cross-industry solution spaces that require processes and technologies from both the Internet of Things (IoT), and Data Management and Analytics (DMA). These solution spaces are evolving into Sensor Analytics Ecosystems (SAE) that represent significant new opportunities for organizations of all types. Public Utilities throughout the world, providing electricity, natural gas and water, are pursuing SmartGrid initiatives that represent one of the more mature examples of SAE. We have s...
There is no doubt that Big Data is here and getting bigger every day. Building a Big Data infrastructure today is no easy task. There are an enormous number of choices for database engines and technologies. To make things even more challenging, requirements are getting more sophisticated, and the standard paradigm of supporting historical analytics queries is often just one facet of what is needed. As Big Data growth continues, organizations are demanding real-time access to data, allowing immediate and actionable interpretation of events as they happen. Another aspect concerns how to deliver ...
Scott Jenson leads a project called The Physical Web within the Chrome team at Google. Project members are working to take the scalability and openness of the web and use it to talk to the exponentially exploding range of smart devices. Nearly every company today working on the IoT comes up with the same basic solution: use my server and you'll be fine. But if we really believe there will be trillions of these devices, that just can't scale. We need a system that is open a scalable and by using the URL as a basic building block, we open this up and get the same resilience that the web enjoys.
The Internet of Things will greatly expand the opportunities for data collection and new business models driven off of that data. In her session at @ThingsExpo, Esmeralda Swartz, CMO of MetraTech, discussed how for this to be effective you not only need to have infrastructure and operational models capable of utilizing this new phenomenon, but increasingly service providers will need to convince a skeptical public to participate. Get ready to show them the money!
The 3rd International Internet of @ThingsExpo, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that its Call for Papers is now open. The Internet of Things (IoT) is the biggest idea since the creation of the Worldwide Web more than 20 years ago.