Forgot to add, this JIRA details the latest security features that are being
worked on in Hadoop trunk: https://issues.apache.org/jira/browse/HADOOP-4487.
This document describes the current status and limitations of the
permissions mechanism:
http://hadoop.apache.org/core/docs/current/hdfs_permissions_guide.html.

On Sun, Feb 15, 2009 at 2:35 PM, Matei Zaharia <ma...@cloudera.com> wrote:

> I think it's safe to assume that Hadoop works like MapReduce/GFS at the
> level described in those papers. In particular, in HDFS, there is a master
> node containing metadata and a number of slave nodes (datanodes) containing
> blocks, as in GFS. Clients start by talking to the master to list
> directories, etc. When they want to read a region of some file, they tell
> the master the filename and offset, and they receive a list of block
> locations (datanodes). They then contact the individual datanodes to read
> the blocks. When clients write a file, they first obtain a new block ID and
> list of nodes to write it to from the master, then contact the datanodes to
> write it (actually, the datanodes pipeline the write as in GFS) and report
> when the write is complete. HDFS actually has some security mechanisms built
> in, authenticating users based on their Unix ID and providing Unix-like file
> permissions. I don't know much about how these are implemented, but they
> would be a good place to start looking.
>
> On Sun, Feb 15, 2009 at 1:36 PM, Amandeep Khurana <ama...@gmail.com>wrote:
>
>> Thanks Matie
>>
>> I had gone through the architecture document online. I am currently
>> working
>> on a project towards Security in Hadoop. I do know how the data moves
>> around
>> in the GFS but wasnt sure how much of that does HDFS follow and how
>> different it is from GFS. Can you throw some light on that?
>>
>> Security would also involve the Map Reduce jobs following the same
>> protocols. Thats why the question about how does the Hadoop framework
>> integrate with the HDFS, and how different is it from Map Reduce and GFS.
>> The GFS and Map Reduce papers give a good information on how those systems
>> are designed but there is nothing that concrete for Hadoop that I have
>> been
>> able to find.
>>
>> Amandeep
>>
>>
>> Amandeep Khurana
>> Computer Science Graduate Student
>> University of California, Santa Cruz
>>
>>
>> On Sun, Feb 15, 2009 at 12:07 PM, Matei Zaharia <ma...@cloudera.com>
>> wrote:
>>
>> > Hi Amandeep,
>> > Hadoop is definitely inspired by MapReduce/GFS and aims to provide those
>> > capabilities as an open-source project. HDFS is similar to GFS (large
>> > blocks, replication, etc); some notable things missing are read-write
>> > support in the middle of a file (unlikely to be provided because few
>> Hadoop
>> > applications require it) and multiple appenders (the record append
>> > operation). You can read about HDFS architecture at
>> > http://hadoop.apache.org/core/docs/current/hdfs_design.html. The
>> MapReduce
>> > part of Hadoop interacts with HDFS in the same way that Google's
>> MapReduce
>> > interacts with GFS (shipping computation to the data), although Hadoop
>> > MapReduce also supports running over other distributed filesystems.
>> >
>> > Matei
>> >
>> > On Sun, Feb 15, 2009 at 11:57 AM, Amandeep Khurana <ama...@gmail.com>
>> > wrote:
>> >
>> > > Hi
>> > >
>> > > Is the HDFS architecture completely based on the Google Filesystem? If
>> it
>> > > isnt, what are the differences between the two?
>> > >
>> > > Secondly, is the coupling between Hadoop and HDFS same as how it is
>> > between
>> > > the Google's version of Map Reduce and GFS?
>> > >
>> > > Amandeep
>> > >
>> > >
>> > > Amandeep Khurana
>> > > Computer Science Graduate Student
>> > > University of California, Santa Cruz
>> > >
>> >
>>
>
>

Reply via email to