http://www.time.com/time/business/article/0,8599,1610306,00.html


Researchers Explore Scrapping Internet

(NEW YORK)­Although it has already taken nearly four decades to get  
this far in building the Internet, some university researchers with  
the federal government's blessing want to scrap all that and start over.

The idea may seem unthinkable, even absurd, but many believe a "clean  
slate" approach is the only way to truly address security, mobility  
and other challenges that have cropped up since UCLA professor  
Leonard Kleinrock helped supervise the first exchange of meaningless  
test data between two machines on Sept. 2, 1969.

The Internet "works well in many situations but was designed for  
completely different assumptions," said Dipankar Raychaudhuri, a  
Rutgers University professor overseeing three clean-slate projects.  
"It's sort of a miracle that it continues to work well today."

No longer constrained by slow connections and computer processors and  
high costs for storage, researchers say the time has come to rethink  
the Internet's underlying architecture, a move that could mean  
replacing networking equipment and rewriting software on computers to  
better channel future traffic over the existing pipes.

Even Vinton Cerf, one of the Internet's founding fathers as co- developer of 
the key communications techniques, said the exercise was  
"generally healthy" because the current technology "does not satisfy  
all needs."

One challenge in any reconstruction, though, will be balancing the  
interests of various constituencies. The first time around,  
researchers were able to toil away in their labs quietly. Industry is  
playing a bigger role this time, and law enforcement is bound to make  
its needs for wiretapping known.

There's no evidence they are meddling yet, but once any research  
looks promising, "a number of people (will) want to be in the drawing  
room," said Jonathan Zittrain, a law professor affiliated with Oxford  
and Harvard universities. "They'll be wearing coats and ties and  
spilling out of the venue."

The National Science Foundation wants to build an experimental  
research network known as the Global Environment for Network  
Innovations, or GENI, and is funding several projects at universities  
and elsewhere through Future Internet Network Design, or FIND.

Rutgers, Stanford, Princeton, Carnegie Mellon and the Massachusetts  
Institute of Technology are among the universities pursuing  
individual projects. Other government agencies, including the Defense  
Department, have also been exploring the concept.

The European Union has also backed research on such initiatives,  
through a program known as Future Internet Research and  
Experimentation, or FIRE. Government officials and researchers met  
last month in Zurich to discuss early findings and goals.

A new network could run parallel with the current Internet and  
eventually replace it, or perhaps aspects of the research could go  
into a major overhaul of the existing architecture.

These clean-slate efforts are still in their early stages, though,  
and aren't expected to bear fruit for another 10 or 15 years ­  
assuming Congress comes through with funding.

Guru Parulkar, who will become executive director of Stanford's  
initiative after heading NSF's clean-slate programs, estimated that  
GENI alone could cost $350 million, while government, university and  
industry spending on the individual projects could collectively reach  
$300 million. Spending so far has been in the tens of millions of  
dollars.

And it could take billions of dollars to replace all the software and  
hardware deep in the legacy systems.

Clean-slate advocates say the cozy world of researchers in the 1970s  
and 1980s doesn't necessarily mesh with the realities and needs of  
the commercial Internet.

"The network is now mission critical for too many people, when in the  
(early days) it was just experimental," Zittrain said.

The Internet's early architects built the system on the principle of  
trust. Researchers largely knew one another, so they kept the shared  
network open and flexible ­ qualities that proved key to its rapid  
growth.

But spammers and hackers arrived as the network expanded and could  
roam freely because the Internet doesn't have built-in mechanisms for  
knowing with certainty who sent what.

The network's designers also assumed that computers are in fixed  
locations and always connected. That's no longer the case with the  
proliferation of laptops, personal digital assistants and other  
mobile devices, all hopping from one wireless access point to  
another, losing their signals here and there.

Engineers tacked on improvements to support mobility and improved  
security, but researchers say all that adds complexity, reduces  
performance and, in the case of security, amounts at most to bandages  
in a high-stakes game of cat and mouse.

Workarounds for mobile devices "can work quite well if a small  
fraction of the traffic is of that type," but could overwhelm  
computer processors and create security holes when 90 percent or more  
of the traffic is mobile, said Nick McKeown, co-director of  
Stanford's clean-slate program.

The Internet will continue to face new challenges as applications  
require guaranteed transmissions ­ not the "best effort" approach  
that works better for e-mail and other tasks with less time sensitivity.

Think of a doctor using teleconferencing to perform a surgery  
remotely, or a customer of an Internet-based phone service needing to  
make an emergency call. In such cases, even small delays in relaying  
data can be deadly.

And one day, sensors of all sorts will likely be Internet capable.  
Rather than create workarounds each time, clean-slate researchers  
want to redesign the system to easily accommodate any future  
technologies, said Larry Peterson, chairman of computer science at  
Princeton and head of the planning group for the NSF's GENI.

Even if the original designers had the benefit of hindsight, they  
might not have been able to incorporate these features from the get- go. 
Computers, for instance, were much slower then, possibly too weak  
for the computations needed for robust authentication.

"We made decisions based on a very different technical landscape,"  
said Bruce Davie, a fellow with network-equipment maker Cisco Systems  
Inc., which stands to gain from selling new products and  
incorporating research findings into its existing line.

"Now, we have the ability to do all sorts of things at very high  
speeds," he said. "Why don't we start thinking about how we take  
advantage of those things and not be constrained by the current  
legacy we have?"

Of course, a key question is how to make any transition ­ and  
researchers are largely punting for now. "Let's try to define where  
we think we should end up, what we think the Internet should look  
like in 15 years' time, and only then would we decide the path,"  
McKeown said. "We acknowledge it's going to be really hard but I  
think it will be a mistake to be deterred by that."

Kleinrock, the Internet pioneer at UCLA, questioned the need for a  
transition at all, but said such efforts are useful for their out-of- the-box 
thinking.
 "A thing called GENI will almost surely not become  
the Internet, but pieces of it might fold into the Internet as it  
advances," he said.

Think evolution, not revolution. Princeton already runs a smaller  
experimental network called PlanetLab, while Carnegie Mellon has a  
clean-slate project called 100 x 100.

These days, Carnegie Mellon professor Hui Zhang said he no longer  
feels like "the outcast of the community" as a champion of clean- slate 
designs. 
Construction on GENI could start by 2010 and take  
about five years to complete. Once operational, it should have a  
decade-long lifespan.

FIND, meanwhile, funded about two dozen projects last year and is  
evaluating a second round of grants for research that could  
ultimately be tested on GENI. These go beyond projects like Internet2  
and National LambdaRail, both of which focus on next-generation needs  
for speed.

Any redesign may incorporate mechanisms, known as virtualization, for  
multiple networks to operate over the same pipes, making further  
transitions much easier. Also possible are new structures for data  
packets and a replacement of Cerf's TCP/IP communications protocols.

"Almost every assumption going into the current design of the  
Internet is open to reconsideration and challenge," said Parulkar,  
the NSF official heading to Stanford. "Researchers may come up with  
wild ideas and very innovative ideas that may not have a lot to do  
with the current Internet."




---------------------------------------------------------------
             WWWhatsup NYC
http://pinstand.com - http://punkcast.com
--------------------------------------------------------------- 

_______________________________________________
Discuss mailing list
[email protected]
http://lists.isoc-ny.org/mailman/listinfo/discuss

Reply via email to