Re: failed to build trunk, what's wrong?

2012-07-16 Thread scorpiovn
Thank you for this guide.

--
View this message in context: 
http://hadoop-common.472056.n3.nabble.com/failed-to-build-trunk-what-s-wrong-tp3660097p3991340.html
Sent from the Users mailing list archive at Nabble.com.


Re: failed to build trunk, what's wrong?

2012-07-16 Thread scorpiovn
thank you for this guide.

--
View this message in context: 
http://hadoop-common.472056.n3.nabble.com/failed-to-build-trunk-what-s-wrong-tp3660097p3991341.html
Sent from the Users mailing list archive at Nabble.com.


Re: hadoop dfs -ls

2012-07-16 Thread Nitin Pawar
I managed to solve this by moving contents on hdfs-site.xml to
core-site.xml


Thanks

On Fri, Jul 13, 2012 at 9:54 PM, Leo Leung  wrote:

> Hi Nitin,
>
>
>
> Normally your conf should reside in /etc/hadoop/conf (if you don't have
> one. Copy it from the namenode - and keep it sync)
>
>
>
> hadoop (script) by default depends on hadoop-setup.sh which depends on
> hadoop-env.sh in /etc/hadoop/conf
>
>
>
> Or during runtime specify the config dir
>
> i.e:
>
>
>
> [hdfs]$  hadoop [--config ] 
>
>
>
>
>
> P.S. Some useful links:
>
> http://wiki.apache.org/hadoop/FAQ
>
> http://wiki.apache.org/hadoop/FrontPage
>
> http://wiki.apache.org/hadoop/
>
> http://hadoop.apache.org/common/docs/r1.0.3/
>
>
>
>
>
>
>
>
>
> -Original Message-
> From: d...@paraliatech.com [mailto:d...@paraliatech.com] On Behalf Of
> Dave Beech
> Sent: Friday, July 13, 2012 6:18 AM
> To: common-user@hadoop.apache.org
> Subject: Re: hadoop dfs -ls
>
>
>
> Hi Nitin
>
>
>
> It's likely that your hadoop command isn't finding the right configuration.
>
> In particular it doesn't know where your namenode is
> (fs.default.namesetting in core-site.xml)
>
>
>
> Maybe you need to set the HADOOP_CONF_DIR environment variable to point to
> your conf directory.
>
>
>
> Dave
>
>
>
> On 13 July 2012 14:11, Nitin Pawar  nitinpawar...@gmail.com>> wrote:
>
>
>
> > Hi,
>
> >
>
> > I have done setup numerous times but this time i did after some break.
>
> >
>
> > I managed to get the cluster up and running fine but when I do  hadoop
>
> > dfs -ls /
>
> >
>
> > it actually shows me contents of linux file system
>
> >
>
> > I am using hadoop-1.0.3 on rhel5.6
>
> >
>
> > Can anyone suggest what I must have done wrong?
>
> >
>
> > --
>
> > Nitin Pawar
>
> >
>



-- 
Nitin Pawar


Re: stuck in safe mode after restarting dfs after found dead node

2012-07-16 Thread Juan Pino
Thanks for your help. I ended up reformatting everything after backing up
the data.
After reinstalling the cluster, I still have problems with hadoop commands
hanging forever before returning the command line prompt.
Also, if I ctrl-Z and bg a hadoop command, then it finishes very quickly,
if I leave running in the foreground, then it hangs before returning the
prompt.
Same thing happens for a pseudo-distributed cluster installation on one
machine.

Juan

On Sat, Jul 14, 2012 at 4:23 PM, Edward Capriolo wrote:

> If the files are gone forever you should run:
>
> hadoop fsck -delete /
>
> To acknowledge they have moved on from existence. Otherwise things
> that attempt to read this files will, to put it in a technical way,
> BARF.
>
> On Fri, Jul 13, 2012 at 12:22 PM, Juan Pino 
> wrote:
> > Thank you for your reply. I ran that command before and it works fine but
> > hadoop fs -ls diplays the list of files in the user's directory but then
> > hangs for quite a while (~ 10 minutes) before
> > handing the command line prompt back, then if I rerun the same command
> > there is no problem. That is why I would like to be able to leave safe
> mode
> > automatically (at least I think it's related).
> > Also, in the hdfs web page, clicking on the Live Nodes or Dead Nodes
> links
> > hangs forever but I am able to browse the file
> > system without any problem with the browser.
> > There is no error in the logs.
> > Please let me know what sort of details I can provide to help resolve
> this
> > issue.
> >
> > Best,
> >
> > Juan
> >
> > On Fri, Jul 13, 2012 at 4:10 PM, Edward Capriolo  >wrote:
> >
> >> If the datanode is not coming back you have to explicitly tell hadoop
> >> to leave safemode.
> >>
> >>
> http://hadoop.apache.org/common/docs/r0.17.2/hdfs_user_guide.html#Safemode
> >>
> >> hadoop dfsadmin -safemode leave
> >>
> >>
> >> On Fri, Jul 13, 2012 at 9:35 AM, Juan Pino  >
> >> wrote:
> >> > Hi,
> >> >
> >> > I can't get HDFS to leave safe mode automatically. Here is what I did:
> >> >
> >> > -- there was a dead node
> >> > -- I stopped dfs
> >> > -- I restarted dfs
> >> > -- Safe mode wouldn't leave automatically
> >> >
> >> > I am using hadoop-1.0.2
> >> >
> >> > Here are the logs:
> >> >
> >> > end of hadoop-hadoop-namenode.log (attached):
> >> >
> >> > 2012-07-13 13:22:29,372 INFO org.apache.hadoop.hdfs.StateChange:
> STATE*
> >> Safe
> >> > mode ON.
> >> > The ratio of reported blocks 0.9795 has not reached the threshold
> 0.9990.
> >> > Safe mode will be turned off automatically.
> >> > 2012-07-13 13:22:29,375 INFO org.apache.hadoop.hdfs.StateChange:
> STATE*
> >> Safe
> >> > mode extension entered.
> >> > The ratio of reported blocks 0.9990 has reached the threshold 0.9990.
> >> Safe
> >> > mode will be turned off automatically in 29 seconds.
> >> > 2012-07-13 13:22:29,375 INFO org.apache.hadoop.hdfs.StateChange:
> *BLOCK*
> >> > NameSystem.processReport: from , blocks: 3128, processing time: 4
> msecs
> >> > 2012-07-13 13:31:29,201 INFO org.apache.hadoop.hdfs.StateChange:
> BLOCK*
> >> > NameSystem.processReport: discarded non-initial block report from
> because
> >> > namenode still in startup phase
> >> >
> >> > Any help would be greatly appreciated.
> >> >
> >> > Best,
> >> >
> >> > Juan
> >> >
> >>
>


Data Nodes not seeing NameNode / Task Trackers not seeing JobTracker

2012-07-16 Thread Ronan Lehane
Hi All,

I was wondering if anyone could help me figure out what's going wrong in my
five node Hadoop cluster, please?

It consists of:
1. NameNode
hduser@namenode:/usr/local/hadoop$ jps
13049 DataNode
13387 Jps
12740 NameNode
13316 SecondaryNameNode

2. JobTracker
hduser@jobtracker:/usr/local/hadoop$ jps
21817 TaskTracker
21448 DataNode
21542 JobTracker
21862 Jps

3. Slave1
hduser@slave1:/usr/local/hadoop$ jps
21226 DataNode
21514 Jps
21463 TaskTracker

4. Slave2
hduser@slave2:/usr/local/hadoop$ jps
20938 Jps
20650 DataNode
20887 TaskTracker

5. Slave3
hduser@slave3:/usr/local/hadoop$ jps
22145 Jps
21854 DataNode
22091 TaskTracker

All DataNodes have been kicked off by running start-dfs.sh on the NameNode
All TaskTrackers have been kicked off by running start-mapred.sh on the
JobTracker

When I try to execute a simple wordcount job from the NameNode I receive
the following error:
12/07/16 19:25:22 ERROR security.UserGroupInformation:
PriviledgedActionException as:hduser cause:java.net.ConnectException: Call
to jobtracker/10.21.68.218:54311 failed on connection exception:
java.net.ConnectException: Connection refused

If I check the jobtracker:
1. I can ping in both directions by both IP and Hostname
2. I can see that the jobtracker is listening on port 54311
tcp0  0 127.0.0.1:54311 0.0.0.0:*
LISTEN  1001   425093  21542/java
3. Telnet to this port from the NameNode fails with "Connection Refused"
telnet: Unable to connect to remote host: Connection refused

This issue can be worked around by moving the JobTracker functionality to
the NameNode, but when this is done the job is executed on the NameNode
rather than distributed across the cluster.
Checking the log files on the slaves nodes, I see Server Not Available
messages referenced at the below wiki.
http://wiki.apache.org/hadoop/ServerNotAvailable
The Data Nodes not seeing the NameNode and the Task Trackers not seeing
JobTracker.
Checking the JobTracker web interface, it always states there is only 1
node available.

I've checked the 5 troubleshooting steps provided but it all looks to be ok
in my environment.

Would anyone have any idea's of what could be causing this?
Any help would be appreciated.

Cheers,
Ronan


Re: Data Nodes not seeing NameNode / Task Trackers not seeing JobTracker

2012-07-16 Thread Harsh J
Ronan,

A couple of simple things to ensure first:

1. Make sure the firewall isn't the one at fault here. Best to disable
firewall if you do not need it, or carefully configure the rules to
allow in/out traffic over chosen ports.
2. Ensure that the hostnames fs.default.name and mapred.job.tracker
bind to, are external IP-resolving hostnames and not localhost
(loopback interface bound) addresses.

On Tue, Jul 17, 2012 at 12:05 AM, Ronan Lehane  wrote:
> Hi All,
>
> I was wondering if anyone could help me figure out what's going wrong in my
> five node Hadoop cluster, please?
>
> It consists of:
> 1. NameNode
> hduser@namenode:/usr/local/hadoop$ jps
> 13049 DataNode
> 13387 Jps
> 12740 NameNode
> 13316 SecondaryNameNode
>
> 2. JobTracker
> hduser@jobtracker:/usr/local/hadoop$ jps
> 21817 TaskTracker
> 21448 DataNode
> 21542 JobTracker
> 21862 Jps
>
> 3. Slave1
> hduser@slave1:/usr/local/hadoop$ jps
> 21226 DataNode
> 21514 Jps
> 21463 TaskTracker
>
> 4. Slave2
> hduser@slave2:/usr/local/hadoop$ jps
> 20938 Jps
> 20650 DataNode
> 20887 TaskTracker
>
> 5. Slave3
> hduser@slave3:/usr/local/hadoop$ jps
> 22145 Jps
> 21854 DataNode
> 22091 TaskTracker
>
> All DataNodes have been kicked off by running start-dfs.sh on the NameNode
> All TaskTrackers have been kicked off by running start-mapred.sh on the
> JobTracker
>
> When I try to execute a simple wordcount job from the NameNode I receive
> the following error:
> 12/07/16 19:25:22 ERROR security.UserGroupInformation:
> PriviledgedActionException as:hduser cause:java.net.ConnectException: Call
> to jobtracker/10.21.68.218:54311 failed on connection exception:
> java.net.ConnectException: Connection refused
>
> If I check the jobtracker:
> 1. I can ping in both directions by both IP and Hostname
> 2. I can see that the jobtracker is listening on port 54311
> tcp0  0 127.0.0.1:54311 0.0.0.0:*
> LISTEN  1001   425093  21542/java
> 3. Telnet to this port from the NameNode fails with "Connection Refused"
> telnet: Unable to connect to remote host: Connection refused
>
> This issue can be worked around by moving the JobTracker functionality to
> the NameNode, but when this is done the job is executed on the NameNode
> rather than distributed across the cluster.
> Checking the log files on the slaves nodes, I see Server Not Available
> messages referenced at the below wiki.
> http://wiki.apache.org/hadoop/ServerNotAvailable
> The Data Nodes not seeing the NameNode and the Task Trackers not seeing
> JobTracker.
> Checking the JobTracker web interface, it always states there is only 1
> node available.
>
> I've checked the 5 troubleshooting steps provided but it all looks to be ok
> in my environment.
>
> Would anyone have any idea's of what could be causing this?
> Any help would be appreciated.
>
> Cheers,
> Ronan



-- 
Harsh J


Group mismatches?

2012-07-16 Thread Clay B.

Hi all,

I have a Hadoop cluster which uses Samba to map an Active Directory domain 
to my CentOS 5.7 Hadoop cluster. However, I notice a strange mismatch with 
groups. Does anyone have any debugging advice, or how to refresh the DFS 
groups mapping? If not, should I file a bug at 
https://issues.apache.org/jira/browse/HADOOP?


# I see the following error:
[clayb@hamster ~]$ hadoop fs -ls /projects/foobarcommander
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".
ls: could not get get listing for 'hdfs://hamster:9000/projects/foobarcommander' : 
org.apache.hadoop.security.AccessControlException: Permission denied: user=clayb, 
access=READ_EXECUTE, 
inode="/projects/foobarcommander":hadmin:foobarcommander:drwxrwx---

# I verify group membership -- look a mismatch!
[clayb@hamster ~]$ which groups
/usr/bin/groups
[clayb@hamster ~]$ groups
foobarcommander xxx_rec_eng domain users all all_north america batchlogon xxx-s 
xxx03-s xxx1-admins xxx-emr-users xxx-emr-admins xxx1-users BUILTIN\users
[clayb@hamster ~]$ hadoop dfsgroups
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".
clayb : domain users xxx_rec_eng xxx-emr-users all xxx-emr-admins batchlogon 
all_north america xxx1-users xxx-s xxx03-s xxx1-admins BUILTIN\users

Notice, in particular the foobarcommander group is only shown for my 
/usr/bin/groups output. It looks like the following from the HDFS 
Permissions Guide[1] is not correct, in my case:

"The group list is the equivalent of `bash -c groups`."

# I have tried the following to no useful effect:
[admin@hamster ~]$ hadoop dfsadmin -refreshUserToGroupsMappings
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".

# I do, however, see other users with the foobarcommander group, so the group should be 
"visible" to Hadoop:
[clayb@hamster ~]$ hadoop dfsgroups pat
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".
pat : domain users all_north america all_san diego all foobarcommander 
BUILTIN\users
# And 'hadoop mrgroups' (like dfsgroups) returns the same bad data, for me:
[clayb@hamster ~]$ hadoop mrgroups
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".
clayb : domain users xxx_rec_eng xxx-emr-users all xxx-emr-admins batchlogon 
all_north america xxx1-users xxx-s xxx03-s xxx1-admins BUILTIN\users

# And I see that the system is getting the right data via getent(1):
[clayb@hamster ~]$ getent group foobarcommander
foobarcommander:*:16777316:pat,user1,user2,user3,clayb,user4,user5,user6,user7,user8,user9,user10,user12,user13,user14,user15

# I am using Cloudera's CDH3u4 Hadoop:
[clayb@hamster ~]$ hadoop version
Hadoop 0.20.2-cdh3u4
Subversion file:///data/1/tmp/topdir/BUILD/hadoop-0.20.2-cdh3u4 -r 
214dd731e3bdb687cb55988d3f47dd9e248c5690
Compiled by root on Mon May  7 14:03:02 PDT 2012

From source with checksum a60c9795e41a3248b212344fb131c12c


I also do not see any obviously useful errors in my namenode logs.

-Clay

[1]: 
http://hadoop.apache.org/common/docs/r0.20.2/hdfs_permissions_guide.html#User+Identity




Re: Group mismatches?

2012-07-16 Thread Edward Capriolo
In all places I have found it only to be the primary group, not all
the users supplemental groups.

On Mon, Jul 16, 2012 at 3:05 PM, Clay B.  wrote:
> Hi all,
>
> I have a Hadoop cluster which uses Samba to map an Active Directory domain
> to my CentOS 5.7 Hadoop cluster. However, I notice a strange mismatch with
> groups. Does anyone have any debugging advice, or how to refresh the DFS
> groups mapping? If not, should I file a bug at
> https://issues.apache.org/jira/browse/HADOOP?
>
> # I see the following error:
> [clayb@hamster ~]$ hadoop fs -ls /projects/foobarcommander
> log4j:ERROR Could not find value for key log4j.appender.NullAppender
> log4j:ERROR Could not instantiate appender named "NullAppender".
> ls: could not get get listing for
> 'hdfs://hamster:9000/projects/foobarcommander' :
> org.apache.hadoop.security.AccessControlException: Permission denied:
> user=clayb, access=READ_EXECUTE,
> inode="/projects/foobarcommander":hadmin:foobarcommander:drwxrwx---
>
> # I verify group membership -- look a mismatch!
> [clayb@hamster ~]$ which groups
> /usr/bin/groups
> [clayb@hamster ~]$ groups
> foobarcommander xxx_rec_eng domain users all all_north america batchlogon
> xxx-s xxx03-s xxx1-admins xxx-emr-users xxx-emr-admins xxx1-users
> BUILTIN\users
> [clayb@hamster ~]$ hadoop dfsgroups
> log4j:ERROR Could not find value for key log4j.appender.NullAppender
> log4j:ERROR Could not instantiate appender named "NullAppender".
> clayb : domain users xxx_rec_eng xxx-emr-users all xxx-emr-admins batchlogon
> all_north america xxx1-users xxx-s xxx03-s xxx1-admins BUILTIN\users
>
> Notice, in particular the foobarcommander group is only shown for my
> /usr/bin/groups output. It looks like the following from the HDFS
> Permissions Guide[1] is not correct, in my case:
> "The group list is the equivalent of `bash -c groups`."
>
> # I have tried the following to no useful effect:
> [admin@hamster ~]$ hadoop dfsadmin -refreshUserToGroupsMappings
> log4j:ERROR Could not find value for key log4j.appender.NullAppender
> log4j:ERROR Could not instantiate appender named "NullAppender".
>
> # I do, however, see other users with the foobarcommander group, so the
> group should be "visible" to Hadoop:
> [clayb@hamster ~]$ hadoop dfsgroups pat
> log4j:ERROR Could not find value for key log4j.appender.NullAppender
> log4j:ERROR Could not instantiate appender named "NullAppender".
> pat : domain users all_north america all_san diego all foobarcommander
> BUILTIN\users
> # And 'hadoop mrgroups' (like dfsgroups) returns the same bad data, for me:
> [clayb@hamster ~]$ hadoop mrgroups
> log4j:ERROR Could not find value for key log4j.appender.NullAppender
> log4j:ERROR Could not instantiate appender named "NullAppender".
> clayb : domain users xxx_rec_eng xxx-emr-users all xxx-emr-admins batchlogon
> all_north america xxx1-users xxx-s xxx03-s xxx1-admins BUILTIN\users
>
> # And I see that the system is getting the right data via getent(1):
> [clayb@hamster ~]$ getent group foobarcommander
> foobarcommander:*:16777316:pat,user1,user2,user3,clayb,user4,user5,user6,user7,user8,user9,user10,user12,user13,user14,user15
>
> # I am using Cloudera's CDH3u4 Hadoop:
> [clayb@hamster ~]$ hadoop version
> Hadoop 0.20.2-cdh3u4
> Subversion file:///data/1/tmp/topdir/BUILD/hadoop-0.20.2-cdh3u4 -r
> 214dd731e3bdb687cb55988d3f47dd9e248c5690
> Compiled by root on Mon May  7 14:03:02 PDT 2012
> From source with checksum a60c9795e41a3248b212344fb131c12c
>
> I also do not see any obviously useful errors in my namenode logs.
>
> -Clay
>
> [1]:
> http://hadoop.apache.org/common/docs/r0.20.2/hdfs_permissions_guide.html#User+Identity
>


Re: Data Nodes not seeing NameNode / Task Trackers not seeing JobTracker

2012-07-16 Thread Ronan Lehane
Thanks for the quick reply Harsh.
I think you may be onto something with the second suggestion.

I found an earlier thread saying that some of the troubleshooting steps
outlined below resolved a similar issue for that person:
http://wiki.apache.org/hadoop/Hbase/Troubleshooting

Like you suggested, the /etc/hosts file definitely looks to be involved as
I hit different issues depending on what hostnames are set against the
loopback addresses.
I'll try reset them to see if it resolves the issue.

Thanks,
Ronan


On Mon, Jul 16, 2012 at 7:44 PM, Harsh J  wrote:

> Ronan,
>
> A couple of simple things to ensure first:
>
> 1. Make sure the firewall isn't the one at fault here. Best to disable
> firewall if you do not need it, or carefully configure the rules to
> allow in/out traffic over chosen ports.
> 2. Ensure that the hostnames fs.default.name and mapred.job.tracker
> bind to, are external IP-resolving hostnames and not localhost
> (loopback interface bound) addresses.
>
> On Tue, Jul 17, 2012 at 12:05 AM, Ronan Lehane 
> wrote:
> > Hi All,
> >
> > I was wondering if anyone could help me figure out what's going wrong in
> my
> > five node Hadoop cluster, please?
> >
> > It consists of:
> > 1. NameNode
> > hduser@namenode:/usr/local/hadoop$ jps
> > 13049 DataNode
> > 13387 Jps
> > 12740 NameNode
> > 13316 SecondaryNameNode
> >
> > 2. JobTracker
> > hduser@jobtracker:/usr/local/hadoop$ jps
> > 21817 TaskTracker
> > 21448 DataNode
> > 21542 JobTracker
> > 21862 Jps
> >
> > 3. Slave1
> > hduser@slave1:/usr/local/hadoop$ jps
> > 21226 DataNode
> > 21514 Jps
> > 21463 TaskTracker
> >
> > 4. Slave2
> > hduser@slave2:/usr/local/hadoop$ jps
> > 20938 Jps
> > 20650 DataNode
> > 20887 TaskTracker
> >
> > 5. Slave3
> > hduser@slave3:/usr/local/hadoop$ jps
> > 22145 Jps
> > 21854 DataNode
> > 22091 TaskTracker
> >
> > All DataNodes have been kicked off by running start-dfs.sh on the
> NameNode
> > All TaskTrackers have been kicked off by running start-mapred.sh on the
> > JobTracker
> >
> > When I try to execute a simple wordcount job from the NameNode I receive
> > the following error:
> > 12/07/16 19:25:22 ERROR security.UserGroupInformation:
> > PriviledgedActionException as:hduser cause:java.net.ConnectException:
> Call
> > to jobtracker/10.21.68.218:54311 failed on connection exception:
> > java.net.ConnectException: Connection refused
> >
> > If I check the jobtracker:
> > 1. I can ping in both directions by both IP and Hostname
> > 2. I can see that the jobtracker is listening on port 54311
> > tcp0  0 127.0.0.1:54311 0.0.0.0:*
> > LISTEN  1001   425093  21542/java
> > 3. Telnet to this port from the NameNode fails with "Connection Refused"
> > telnet: Unable to connect to remote host: Connection refused
> >
> > This issue can be worked around by moving the JobTracker functionality to
> > the NameNode, but when this is done the job is executed on the NameNode
> > rather than distributed across the cluster.
> > Checking the log files on the slaves nodes, I see Server Not Available
> > messages referenced at the below wiki.
> > http://wiki.apache.org/hadoop/ServerNotAvailable
> > The Data Nodes not seeing the NameNode and the Task Trackers not seeing
> > JobTracker.
> > Checking the JobTracker web interface, it always states there is only 1
> > node available.
> >
> > I've checked the 5 troubleshooting steps provided but it all looks to be
> ok
> > in my environment.
> >
> > Would anyone have any idea's of what could be causing this?
> > Any help would be appreciated.
> >
> > Cheers,
> > Ronan
>
>
>
> --
> Harsh J
>


Re: Group mismatches?

2012-07-16 Thread Clay B.

Hi Edward,

I'm not sure I understand.

We all share the primary group domain_users (since these come from Active 
Directory over Samba). Still, we are able to be in multiple projects by 
using secondary groups, and it seems dfsgroups/mrgroups tries to report 
secondary groups.


Regardless, if I use the newgrp(1) command to switch my primary group, 
Hadoop still seems blind to the foobarcommander group, however:

[clayb@hamster ~]$ newgrp foobarcommander
[clayb@hamster ~]$ id
uid=16777217(clayb) gid=16777316(foobarcommander) 
groups=5(xxx_rec_eng),16777216(domain 
users),16777218(all),16777221(all_north 
america),16777223(batchlogon),16777226(xxx-s),16777229(xxx03-s),16777235(xxx1-admins),16777236(xxx-emr-users),16777237(xxx-emr-admins),16777265(xxx1-users),16777285(BUILTIN\users),16777316(foobarcommander)
[clayb@hamster ~]$ hadoop dfsgroups
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".
clayb : domain users xxx_rec_eng xxx-emr-users all xxx-emr-admins batchlogon 
all_north america xxx1-users xxx-s xxx03-s xxx1-admins BUILTIN\users

-Clay

On Mon, 16 Jul 2012, Edward Capriolo wrote:


In all places I have found it only to be the primary group, not all
the users supplemental groups.

On Mon, Jul 16, 2012 at 3:05 PM, Clay B.  wrote:

Hi all,

I have a Hadoop cluster which uses Samba to map an Active Directory domain
to my CentOS 5.7 Hadoop cluster. However, I notice a strange mismatch with
groups. Does anyone have any debugging advice, or how to refresh the DFS
groups mapping? If not, should I file a bug at
https://issues.apache.org/jira/browse/HADOOP?

# I see the following error:
[clayb@hamster ~]$ hadoop fs -ls /projects/foobarcommander
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".
ls: could not get get listing for
'hdfs://hamster:9000/projects/foobarcommander' :
org.apache.hadoop.security.AccessControlException: Permission denied:
user=clayb, access=READ_EXECUTE,
inode="/projects/foobarcommander":hadmin:foobarcommander:drwxrwx---

# I verify group membership -- look a mismatch!
[clayb@hamster ~]$ which groups
/usr/bin/groups
[clayb@hamster ~]$ groups
foobarcommander xxx_rec_eng domain users all all_north america batchlogon
xxx-s xxx03-s xxx1-admins xxx-emr-users xxx-emr-admins xxx1-users
BUILTIN\users
[clayb@hamster ~]$ hadoop dfsgroups
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".
clayb : domain users xxx_rec_eng xxx-emr-users all xxx-emr-admins batchlogon
all_north america xxx1-users xxx-s xxx03-s xxx1-admins BUILTIN\users

Notice, in particular the foobarcommander group is only shown for my
/usr/bin/groups output. It looks like the following from the HDFS
Permissions Guide[1] is not correct, in my case:
"The group list is the equivalent of `bash -c groups`."

# I have tried the following to no useful effect:
[admin@hamster ~]$ hadoop dfsadmin -refreshUserToGroupsMappings
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".

# I do, however, see other users with the foobarcommander group, so the
group should be "visible" to Hadoop:
[clayb@hamster ~]$ hadoop dfsgroups pat
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".
pat : domain users all_north america all_san diego all foobarcommander
BUILTIN\users
# And 'hadoop mrgroups' (like dfsgroups) returns the same bad data, for me:
[clayb@hamster ~]$ hadoop mrgroups
log4j:ERROR Could not find value for key log4j.appender.NullAppender
log4j:ERROR Could not instantiate appender named "NullAppender".
clayb : domain users xxx_rec_eng xxx-emr-users all xxx-emr-admins batchlogon
all_north america xxx1-users xxx-s xxx03-s xxx1-admins BUILTIN\users

# And I see that the system is getting the right data via getent(1):
[clayb@hamster ~]$ getent group foobarcommander
foobarcommander:*:16777316:pat,user1,user2,user3,clayb,user4,user5,user6,user7,user8,user9,user10,user12,user13,user14,user15

# I am using Cloudera's CDH3u4 Hadoop:
[clayb@hamster ~]$ hadoop version
Hadoop 0.20.2-cdh3u4
Subversion file:///data/1/tmp/topdir/BUILD/hadoop-0.20.2-cdh3u4 -r
214dd731e3bdb687cb55988d3f47dd9e248c5690
Compiled by root on Mon May  7 14:03:02 PDT 2012
From source with checksum a60c9795e41a3248b212344fb131c12c

I also do not see any obviously useful errors in my namenode logs.

-Clay

[1]:
http://hadoop.apache.org/common/docs/r0.20.2/hdfs_permissions_guide.html#User+Identity





hadoop fs -tail

2012-07-16 Thread Sukhendu Chakraborty
Hi,

Is there a way to get around the 1KB limitation of the hadoop fs -tail
command (http://hadoop.apache.org/common/docs/r0.20.0/hdfs_shell.html#tail)?
In my application some of the records can have length greater than 1KB
before the newline and I would like to get the full records as part of
'tail' (not truncated versions).

Thanks,
-Sukhendu


committing job fails : The stream is closed

2012-07-16 Thread han zheng
hi,allthis happens time by time, and disappears when I made the thread sleep for some msecs before commit the job.Can anybody help?thanks in advance!StackTraces:2012-07-11 16:55:03,069 [Thread-25] ERROR exec.Task (SessionState.java:printError(400)) - Job Commit failed with exception 'org.apache.hadoop.hive.ql.metadata.HiveException(java.io.IOException: Call to master/192.168.2.20:8020 failed on local exception: java.io.IOException: The stream is closed)'org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException: Call to master/192.168.2.20:8020 failed on local exception: java.io.IOException: The stream is closed	at org.apache.hadoop.hive.ql.exec.FileSinkOperator.jobClose(FileSinkOperator.java:752)	at org.apache.hadoop.hive.ql.exec.Operator.jobClose(Operator.java:599)	at org.apache.hadoop.hive.ql.exec.Operator.jobClose(Operator.java:599)	at org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:486)	at org.apache.hadoop.hive.ql.exec.MapRedTask.execute(MapRedTask.java:137)	at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:134)	at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)	at org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:48)Caused by: java.io.IOException: Call to master/192.168.2.20:8020 failed on local exception: java.io.IOException: The stream is closed	at org.apache.hadoop.ipc.Client.wrapException(Client.java:1130)	at org.apache.hadoop.ipc.Client.call(Client.java:1098)	at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:235)	at $Proxy9.getFileInfo(Unknown Source)	at sun.reflect.GeneratedMethodAccessor80.invoke(Unknown Source)	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)	at java.lang.reflect.Method.invoke(Method.java:597)	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)	at $Proxy9.getFileInfo(Unknown Source)	at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:875)	at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:513)	at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:768)	at org.apache.hadoop.hive.ql.exec.Utilities.mvFileToFinalPath(Utilities.java:1263)	at org.apache.hadoop.hive.ql.exec.FileSinkOperator.jobClose(FileSinkOperator.java:749)	at org.apache.hadoop.hive.ql.exec.Operator.jobClose(Operator.java:599)	at org.apache.hadoop.hive.ql.exec.Operator.jobClose(Operator.java:599)	at org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:486)	at org.apache.hadoop.hive.ql.exec.MapRedTask.execute(MapRedTask.java:137)	at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:134)	at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)	at org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:47)Caused by: java.io.IOException: The stream is closed	at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:108)	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:66)	at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:124)	at java.io.DataOutputStream.flush(DataOutputStream.java:107)	at org.apache.hadoop.ipc.Client$Connection.sendParam(Client.java:790)	at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:235)	at $Proxy9.getProtocolVersion(Unknown Source)	at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:405)	at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:388)	at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119)	at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:238)	at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:204)	at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)	at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386)	at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:251)	at org.apache.hadoop.fs.Path.getFileSystem(Path.java:187)	at org.apache.hadoop.hive.ql.exec.Utilities.clearMapRedWork(Utilities.java:181)	at org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:459)	at org.apache.hadoop.hive.ql.exec.MapRedTask.execute(MapRedTask.java:137)	at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:134)	at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)	at org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:48)
韩铮格物穷理 知行合一---杭州数云信息技术有限公司公司地址:上海市长宁区江苏路470号A楼803室手机:18610264627邮箱:zheng@huaat.net--- 



Running hadoop job remotely

2012-07-16 Thread Chandra Mohan, Ananda Vel Murugan
Hi,

I am trying to run HBase importTSV hadoop job to load data into HBase from a 
TSV file.

I am using the following code.

Configuration config = new Configuration();
Iterator iter = config.iterator();
while(iter.hasNext())
{
Object obj = iter.next();
System.out.println(obj);
}

Job job = new Job(config);
job.setJarByClass(ImportTsv.class);
job.setJobName("ImportTsv");
job.getConfiguration().set("user", "hadoop");
job.waitForCompletion(true);

I am getting this error
ERROR security.UserGroupInformation: PriviledgedActionException as:E317376 
cause:org.apache.hadoop.security.AccessControlException: 
org.apache.hadoop.security.AccessControlException: Permission denied: 
user=E317376, access=WRITE, inode="staging":hadoop:supergroup:rwxr-xr-x
I dont know how user name E317376 is being set. This is my windows machine user 
name from where I am trying to run this job in a remote cluster. My haddop user 
account in linux machine is "hadoop"
when i run this in linux machine which is part of Hadoop cluster under hadoop 
user account, everything works well. But I want to programatically run this job 
in a java web application. Am I doing anything wrong. Please help...