Absolutely, you can use map/reduce without HDFS. That is the standard debugging style, for one thing. For another, Nutch is all about accessing non-HDFS data (i.e. the web).
If I get your drift this time, though, I would expect that you will have severe problems with bandwidth if you have a lot of task nodes working on a conventional file store. One of the great virtues of map/reduce + HDFS is that most of the map inputs are read from local disk, as are the reduce inputs. Any system that doesn't coordinate work with storage this way is likely to suffer from lower througput due to network congestion. -----Original Message----- From: C G [mailto:[EMAIL PROTECTED] Sent: Thu 9/6/2007 5:54 PM To: hadoop-user@lucene.apache.org Subject: Re: Use HDFS as a long term storage solution? Actually my question is 'Can you use Map/Reduce without HDFS?" Ted Dunning <[EMAIL PROTECTED]> wrote: You can definitely use HDFS without map/reduce. It should be pretty easy to use it from a variety of languages as well, although it is unlikely that there are language bindings available off the shelf. ... On 9/6/07 1:04 PM, "C G" wrote: > Do you have to use HDFS with map/reduce? I don't fully understand how closely > bound map/reduce is to HDFS. >