Powerful New Supercomputer Analyzes Earthquakes
The Division of Geological and Planetary Sciences' new Geosciences Computational Facility will feature a 2,048-processor supercomputer, housed in the basement of the Seeley G. Mudd Building of Geophysics and Planetary Science on campus.
Computer hardware fills long rows of black racks in the facility, each contains about 35 compute nodes. Massive air conditioning units line an entire wall of the 20-by-80-foot room to re-circulate and chill the air. Miles of optical-fiber cables tie the processors together into a working cluster that went online in September.
The $5.8 million parallel computing project was made possible by gifts from Dell, Myricom, Intel, and the National Science Foundation.
According to Jeroen Tromp, McMillan Professor of Geophysics and director of the Institute's Seismology Lab, who spearheaded the project, "The other crucial ingredient was Caltech's investment in the infrastructure necessary to house the new machine," he says. Some 500 kilowatts of power and 90 tons of air conditioning are needed to operate and cool the hardware.
David Kewley, the project's systems administrator, explained that's enough kilowatts to power 350 average households.
Tromp's research group will share use of the cluster with other division professors and their research groups, while a job-scheduling system will make sure the facility runs at maximum possible capacity. Tromp, who came to Caltech in 2000 from Harvard, is known as one of the world's leading theoretical seismologists. Until now, he and his Institute colleagues have used a smaller version of the machine, popularly known as a Beowulf cluster. Helping revolutionize the field of earthquake study, Tromp has created 3-D simulations of seismic events. He and former Caltech postdoctoral scholar Dimitri Komatitsch designed a computer model that divides the earth into millions of elements. Each element can be divided into slices that represent the earth's geological features.
In simulations involving tens of millions of operations per second, the seismic waves are propagated from one slice to the next, as they speed up, slow down, and change direction according to the earth's characteristics. The model is analogous to a CAT scan of the earth, allowing scientists to track seismic wave paths. "Much like a medical doctor uses a CAT scan to make an image of the brain, seismologists use earthquake-generated waves to image the earth's interior," Tromp says, adding that the earthquake's location, origin time, and characteristics must also be determined.
Tromp will now be able to deliver better, more accurate models in less time. "We hope to use the new machine to do much more detailed mapping. In addition to improving the resolution of our images of the earth's interior, we will also quantitatively assess the devastating effects associated with earthquakes based upon numerical simulations of strong ground motion generated by hypothetical earthquakes."
"One novel way in which we are planning to use the new machine is for near real-time seismology," Tromp adds. "Every time an earthquake over magnitude 3.5 occurs anywhere in California we will routinely simulate the motions associated with the event. Scientific products that result from these simulations are 'synthetic' seismograms that can be compared to actual seismograms."
The "real" seismograms are recorded by the Southern California Seismic Network (SCSN), operated by the Seismo Lab in conjunction with the U.S. Geological Survey. Of interest to the general public, Tromp expects that the collaboration will produce synthetic ShakeMovies of recent quakes, and synthetic ShakeMaps which can be compared to real ShakeMaps derived from the data. "These products should be available within an hour after the earthquake," he says. The Seismology Lab Media Center will be renovated with a large video wall on which scientists can show the results of simulations and analysis.
The new generation of seismic knowledge may also help scientists, engineers, and others lessen the potentially catastrophic effects of earthquakes.
"Intel is proud to be a sponsor of this premier system for seismic research which will be used by researchers and scientists," said Les Karr, Intel Corporate Business Development Manager. "The project reflects Caltech's growing commitment, in both research and teaching, to a broadening range of problems in computational geoscience. It is also a reflection of the growing use of commercial, commodity computing systems to solve some of the world's toughest problems."
The Dell equipment consists of 1,024 dual Dell PowerEdge 1850 servers that were pre-assembled for easy implementation. Dell Services representatives came to campus to complete the installation.
"CITerra, as this new research tool is known on the TOP500 Supercomputer list, is a proud accomplishment both for Caltech and for Myricom," said Charles Seitz, founder and CEO of Myricom, and a former professor of computer science at Caltech. "The talented technical team of Myricom about half of whom are Caltech alumni/ae, are eager for people to know that the architecture, programming methods, and technology of cluster computing was pioneered at Caltech 20 years ago. Those of us at Myricom who have drawn so much inspiration from our Caltech years are delighted to give some of the results of our efforts back to Caltech."
About Myricom: Founded in 1994, Myricom, Inc. created Myrinet, the high-performance computing (HPC) interconnect technology used in thousands of computing clusters in more than 50 countries worldwide. With its next-generation Myri-10G solutions, Myricom is bridging the gap between the rigorous demands of traditional HPC applications and the growing need for affordable computing speed in mainstream enterprises. Privately held, Myricom achieved and has sustained profitability since 1995 with 42 consecutive profitable quarters through September 2005. Based in Arcadia, California, Myricom solutions are sold direct and through channels. Myrinet clusters are supplied by OEM computer companies including IBM, HP, Dell, and Sun, and by other leading cluster integrators worldwide.
About Intel: Intel, the world's largest chipmaker, is also a leading manufacturer of computer, networking, and communications products. Intel processors, platform architectures, interconnects, networking technology, software tools, and services power some of the fastest computers in the world at price points that have expanded high performance computing beyond the confines of elite supercomputer centers and into the broad community of customers in mainstream industries. Those industries span automotive, aerospace, electronics manufacturing, energy and oil and gas in addition to scientific, research and academic organizations.
About the National Science Foundation: The NSF is an independent federal agency created by Congress in 1950 "to promote the progress of science; to advance the national health, prosperity, and welfare; to secure the national defense..." With an annual budget of about $5.5 billion, it is the funding source for approximately 20 percent of all federally supported basic research conducted by America's colleges and universities. In many fields such as mathematics, computer science, and the social sciences, NSF is the major source of federal backing.