With this configuration, any user having that group name will be able to write to any location.. (I've tried this in local network, though)
2009/2/14 Rasit OZDAS <rasitoz...@gmail.com>: > I agree with Amar and James, > > if you require permissions for your project, > then > 1. create a group in linux for your user. > 2. give group write access to all files in HDFS. (hadoop dfs -chmod -R > g+w / - or sth, I'm not totally sure.) > 3. change group ownership of all files in HDFS. (hadoop dfs -chgrp -R > <your_group_name> / - I'm not totally sure again..) > > cheers, > Rasit > > > 2009/2/12 james warren <ja...@rockyou.com>: >> Like Amar said. Try adding >> >> <property> >> <name>dfs.permissions</name> >> <value>false</value> >> </property> >> >> >> to your conf/hadoop-site.xml file (or flip the value in hadoop-default.xml), >> restart your daemons and give it a whirl. >> >> cheers, >> -jw >> >> On Wed, Feb 11, 2009 at 8:44 PM, Amar Kamat <ama...@yahoo-inc.com> wrote: >> >>> bjday wrote: >>> >>>> Good morning everyone, >>>> >>>> I have a question about correct setup for hadoop. I have 14 Dell >>>> computers in a lab. Each connected to the internet and each independent >>>> of >>>> each other. All run CentOS. Logins are handled by NIS. If userA logs >>>> into >>>> the master and starts the daemons and UserB logs into the master and wants >>>> to run a job while the daemons from UserA are still running the following >>>> error occurs: >>>> >>>> copyFromLocal: org.apache.hadoop.security.AccessControlException: >>>> Permission denied: user=UserB, access=WRITE, >>>> inode="user":UserA:supergroup:rwxr-xr-x >>>> >>> Looks like one of your files (input or output) is of different user. Seems >>> like your DFS has permissions enabled. If you dont require permissions then >>> disable it else make sure that the input/output paths are under your >>> permission (/user/userB is the hone directory for userB). >>> Amar >>> >>> >>>> what needs to be changed to allow UserB-UserZ to run their jobs? Does >>>> there need to be a local user the everyone logs into as and run from there? >>>> Should Hadoop be ran in an actual cluster instead of independent >>>> computers? >>>> Any ideas what is the correct configuration settings that allow it? >>>> >>>> I followed Ravi Phulari suggestions and followed: >>>> >>>> http://hadoop.apache.org/core/docs/current/quickstart.html >>>> >>>> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_(Multi-Node_Cluster)< >>>> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Multi-Node_Cluster%29> >>>> >>>> >>>> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_(Single-Node_Cluster)< >>>> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29> >>>> >>>> >>>> These allowed me to get Hadoop running on the 14 computers when I login >>>> and everything works fine, thank you Ravi. The problem occurs when >>>> additional people attempt to run jobs simultaneously. >>>> >>>> Thank you, >>>> >>>> Brian >>>> >>>> >>> >> > > > > -- > M. Raşit ÖZDAŞ > -- M. Raşit ÖZDAŞ