[NY Times] July 15, 2003 Teaching Computers to Work in Unison By STEVE LOHR
Computers do wondrous things, but computer science itself is largely a discipline of step-by-step progress as a steady stream of innovations in hardware, software and networking pile up. It is an engineering science whose frontiers are pushed ahead by people building new tools rendered in silicon and programming code rather than the breathtaking epiphanies and grand unifying theories of mathematics or physics. Yet computer science does have its revelatory moments, typically when several advances come together to create a new computing experience. One of those memorable episodes took place in December 1995 at a supercomputing conference in San Diego. For three days, a prototype project, called I-Way, linked more than a dozen big computer centers in the United States to work as if a single machine on computationally daunting simulations, like the collision of neutron stars and the movement of cloud patterns around the globe. There were glitches and bugs. Only about half of the 60 scientific computer simulations over the I-Way worked. But the participants recall those few days as the first glimpse of what many computer scientists now regard as the next big evolutionary step in the development of the Internet, known as grid computing. "It was the Woodstock of the grid - everyone not sleeping for three days, running around and engaged in a kind of scientific performance art," said Dr. Larry Smarr, director of the California Institute for Telecommunications and Information Technology, who was the program chairman for the conference. The idea of lashing computers together to tackle computing chores for users who tap in as needed - almost as if a utility - has been around since the 1960's. But to move the concept of distributed computing utilities, or grids, toward practical reality has taken years of continuous improvement in computer processing speeds, data storage and network capacity. Perhaps the biggest challenge, however, has been to design software able to juggle and link all the computing resources across far-flung sites, and deliver them on demand. The creation of this basic software - the DNA of grid computing - has been led by Dr. Ian Foster, a senior scientist at the Argonne National Laboratory and a professor of computer science at the University of Chicago, and Dr. Carl Kesselman, director of the center for grid technologies at the University of Southern California's Information Sciences Institute. They have worked together for more than a decade and, a year after the San Diego supercomputing conference, they founded the Globus Project to develop grid software. It is supported mainly by the government, with financing from the Department of Energy, the National Science Foundation, NASA and the Defense Advanced Research Projects Agency. There has been a flurry of grid projects in the last few years in the United States, Europe and Japan, most of them collaborations among scientific researchers at national laboratories and universities on projects like climate modeling, high-energy physics, genetic research, earthquake simulations and brain research. More recently, computer companies including IBM, Platform Computing, Sun Microsystems, Hewlett-Packard and Microsoft have become increasingly interested in grid technology, and some of the early commercial applications include financial risk analysis, oil exploration and drug research. This month, grid computing moved further toward the commercial mainstream when the Globus Project released new software tools that blend the grid standards with a programming technology called Web services, developed mainly in corporate labs, for automated computer-to-computer communications. Enthusiasm for grid computing is also broadening among scientists. A report this year by a National Science Foundation panel, "Revolutionizing Science and Engineering Through Cyberinfrastructure," called for new financing of $1 billion a year to make grid-style computing a routine tool of research. The long-term grid vision is that anyone with a desktop machine or hand-held computer can have the power of a supercomputer at his or her fingertips. And small groups with shared interests could find answers to computationally complex problems as never before. Imagine, for example, a handful of concerned citizens running their own simulation of the environmental impact of a proposed real-estate development in their community. They wouldn't need their own data center or consultants. They would describe what they want, and intelligent software would find the relevant data and summon the computing resources needed for the simulation. "The ultimate goal is a fundamental shift in how we go about solving human problems, and a new way of interacting with technology," Dr. Kesselman said. That grand vision, however, is years away, perhaps a decade or more. Dr. Smarr is the former director of the National Center for Supercomputing Applications at the University of Illinois, where Web browsing software later used by both Netscape Communications and Microsoft was developed in the 1990's. He compares the state of grid computing now to the Web in 1994, when groundbreaking work in a new technology had come from the elite science labs in the United States and Europe but before commercial investment had gathered momentum. The grid is widely regarded as the next stage for the Internet after the World Wide Web. The Web is the Internet's multimedia retrieval system, providing access to text, images, music and video. The promise of the grid is to add a problem-solving system. Computer scientists say the contribution of Dr. Foster and Dr. Kesselman to grid computing is roughly similar to that made by Tim Berners-Lee to the development of the Web. Mr. Berners-Lee, who is now the director of the World Wide Web Consortium at the Massachusetts Institute of Technology, came up with the software standards for addressing, linking and sharing documents over the Web: U.R.L.'s (uniform resource locators), HTTP (hypertext transfer protocol) and HTML (hypertext mark-up language). The heart of the grid problem is managing and linking computing resources. Dr. Foster and Dr. Kesselman, assisted by another software designer at the Argonne lab, Steve Tuecke, have devised basic grid standards with their own acronyms: GRAM (Globus resource allocation manager), M.D.S. (monitoring and discovery service), G.S.I. (grid security infrastructure) and GridFTP (grid file transfer protocol). The wisdom of their work, according to computer scientists, lies in its farsighted simplicity, designing a set of minimalist standards that others can build upon. It is the same design philosophy, they note, found in the original Internet and the Web. "If you look at the history of computer science, the people who have had the biggest impact are the ones who envisioned big systems and then came up with simple but smart mechanisms for building those systems," said Dr. Ken Kennedy, a computer science professor at Rice University. "That's what Ian Foster and Carl Kesselman have done." In 1998, after they developed some early working software, the Globus leaders had to decide the best way to proceed. After long discussions, they chose not to make Globus commercial. Instead, they opted for the open-source model, in which computer code is openly shared, allowing programmers to modify, improve and fix the software. The decision, Dr. Foster recalled, was both practical and ethical. "Our belief was that open source was the best way to maximize adoption," he said. "Globus is an infrastructure technology, and it is only going to be successful if everyone uses it. And if you're doing something that is primarily funded by the government, sharing the software seemed the most appropriate thing to do." Grid computing is a far bigger challenge than simpler forms of distributed computing. Today, most grid projects remain the province of supercomputing centers and university labs. The research centers are linked by network connections about 20 times as fast as the standard high-speed connections and are equipped with storage systems able to handle vast data files and high-performance computers. The Biomedical Informatics Research Network, begun in 2001 and supported by the National Institutes of Health, is a grid created to help scientists gain a better understanding of the way the brain works. One project, called the Brain Morphometry BIRN, involves pooling and processing magnetic resonance imaging data to look for early anatomical and functional precursors of Alzheimer's disease. That knowledge may then be used to tailor drugs to inhibit the onset of the disease. Researchers from Harvard, Duke, the University of North Carolina, Johns Hopkins, the University of California at Los Angeles, the University of California at San Diego, Massachusetts General Hospital and Brigham and Women's Hospital participate in the study. From their desktop computer, any of them can tap into data anywhere on the brain project's grid. "BIRN is a leading example of how you use this cyberinfrastructure to make team science happen to achieve stretch goals in research," said Dr. Mark Ellisman, the neuroscientist at the University of California at San Diego who led the design of the BIRN data grid. The collaborative computing tools for data sharing are also fostering a new style of research. "We're helping a scientific community to understand that it does more good to make information more generally accessible than squirreling it away," Dr. Ellisman said. Dr. Foster, the software tool maker, is encouraged by the applications built on his group's underlying technology. "Like nearly everything in computer science, the work we've done on the Globus software is incremental," he said. "But it is having an impact. There are thousands of people doing collaborative, computing-intensive work in a variety of fields that they could not do before."