Hadoop may not be what you want for storing lots and lots of files.

If you need to store >10^7 files or if you are storing lots of small (<40MB)
files, then you may prefer a solution like mogileFS.  It is engineered for a
very different purpose than hadoop, but may be more appropriate for what you
want.  It is also already intended for web-scale reliable applications so
there is a bit more that you can do for redundancy.

On the other hand, HDFS might be just what you need.


On 9/5/07 1:03 PM, "Dongsheng Wang" <[EMAIL PROTECTED]> wrote:

> 
> We are looking at using HDFS as a long term storage solution. We want to use
> it to stored lots of files. The file could be big and small, they are images,
> videos etc... We only write the files once, and may read them many times.
> Sounds like it is perfect to use HDFS.
> 
> The concern is that since it¹s been engineered to support MapReduce there may
> be  fundamental assumptions that the data being stored by HDFS is transient in
> nature. Obviously for our scalable storage solution zero data loss or
> corruption is a heavy requirement.
> 
> Is anybody using HDFS as a long term storage solution? Interested in any info.
> Thanks
> 
> - ds
> 
>        
> ---------------------------------
> Yahoo! oneSearch: Finally,  mobile search that gives answers, not web links. 

Reply via email to