Massive computing grid to probe the nature of the universe
Apr 4, 2003 — by LinuxDevices Staff — from the LinuxDevices Archive — views(PR excerpt) — IBM and the European Organization for Nuclear Research (CERN) announced that IBM is joining the CERN openlab for DataGrid applications to collaborate in creating a massive data-management system built on Grid computing. IBM will supply CERN with the system's initial 20 terabytes of high-performance disk storage, a cluster of six IBM eServer xSeries systems running Linux, and on-site… engineering support and services.
IBM's innovative storage virtualization and file management technologies will play a pivotal role in this collaboration, which aims to create a data file system far larger than exists today, to help scientists at the renowned particle physics research center understand some of the most fundamental questions about the nature of matter and the universe.
Conceived in IBM Research as “Storage Tank”, the new technology is designed to provide scalable, high-performance and highly available management of huge amounts of data using a single file namespace regardless of where or on what operating system the data reside. Storage Tank employs policy-based storage management to help lower its “total cost of ownership.” Clustering and specialized protocols that detect network failures enable very high reliability and availability.
IBM and CERN will work together to extend Storage Tank's capabilities so it can manage and provide access from any location worldwide to the unprecedented torrent of data — billions of gigabytes a year — that CERN's Large Hadron Collider (LHC) is expected to produce when it goes online in 2007. The LHC is the next-generation particle accelerator. It will recreate — on a tiny scale — conditions that existed shortly after the Big Bang, enabling researchers to answer outstanding questions about what the universe is made of and the laws that govern its behaviour.
The very same CERN community that invented the World Wide Web in 1990 is now developing a new application for the Internet — Grid computing — that will push technology limits with its data processing requirements for the LHC. CERN openlab is a collaboration between CERN and leading industrial partners, which aims to create and implement data-intensive Grid-computing technologies that will aid the LHC scientists. Because the same issues facing CERN are becoming increasingly important to the IT industry, the CERN openlab and its innovative partners — which include Enterasys Networks, HP, and Intel — are eager to explore new computing and data management solutions far beyond today's Internet-based computing.
By 2005, the CERN openlab collaboration with IBM is expected to be able to handle up to a petabyte (a million gigabytes) of data, which is equivalent to the information stored in 20 million four-draw filing cabinets full of paper, or 500 million floppy disks, or 1.5 million CD-ROMs.
This article was originally published on LinuxDevices.com and has been donated to the open source community by QuinStreet Inc. Please visit LinuxToday.com for up-to-date news and articles about Linux and open source.