Re: QueueMetrics.AppsKilled/Failed metrics and failure reasons

2015-02-03 Thread Suma Shivaprasad
Using hadoop 2.4.0. #of Applications running on average is small ~ 40 -60.
The metrics in Ganglia shows around around 10-30 apps killed every 5 mins
which is very high wrt to the apps running at any given time(40-60). The RM
logs though show 0 failed apps in audit logs during that hour.
The RM UI also doesnt show any apps in Applications-Failed tab . The logs
are getting rolled over at a slower rate ..every 1-2 hours. Am searching
for Application Finished - Failed to find the apps failed. Please let me
know if I am missing something here.

Thanks
Suma




On Wed, Feb 4, 2015 at 10:03 AM, Rohith Sharma K S 
rohithsharm...@huawei.com wrote:

  Hi



 Could you give more information, which version of hadoop are you using?



  QueueMetrics.AppsKilled/Failed metrics shows much higher nos i.e ~100.
 However RMAuditLogger shows 1 or 2 Apps as Killed/Failed in the logs.

 May be I suspect that Logs might be rolled out. Does more applications are
 running?



 All the applications history will be displayed  on RM web UI (provided RM
 is not restarted or RM recovery enabled). May be you can check these
 applications lists.



 For finding reasons for application killed/failed, one way is you can
 check in NodeManager logs also. Here  you need to check using container_id
 for corresponding application.



 Thanks  Regards

 Rohith Sharma K S



 *From:* Suma Shivaprasad [mailto:sumasai.shivapra...@gmail.com]
 *Sent:* 03 February 2015 21:35
 *To:* user@hadoop.apache.org; yarn-...@hadoop.apache.org
 *Subject:* QueueMetrics.AppsKilled/Failed metrics and failure reasons



 Hello,


 Was trying to debug reasons for Killed/Failed apps and was checking for
 the applications that were killed/failed in RM logs - from RMAuditLogger.

  QueueMetrics.AppsKilled/Failed metrics shows much higher nos i.e ~100.
 However RMAuditLogger shows 1 or 2 Apps as Killed/Failed in the logs. Is it
 possible that some logs are missed by AuditLogger or is it the other way
 round and metrics are being reported higher ?

 Thanks

 Suma



Can I configure multiple M/Rs and normal processes to one workflow?

2015-02-03 Thread 임정택
Hello all.

We're periodically scan HBase tables to aggregate statistic information,
and store it to MySQL.

We have 3 kinds of CP (kind of data source), each has one Channel and one
Article table.
(Channel : Article is 1:N relation.)

All CPs table schema are different a bit, so in order to aggregate we
should apply different logics, with joining Channel and Article.

I've thought about workflow like this, but I wonder it can make sense.

1. run single process which initializes MySQL by creating table, deleting
row, etc.
2. run 3 M/Rs simultaneously to aggregate statistic information for each
CP, and insert rows  per Channel to MySQL.
3. run single process which finalizes whole aggregation - runs aggregation
query from MySQL to insert new row to MySQL, rolling table, etc.

Definitely 1,2,3 should be run in a row.

Any helps are really appreciated!
Thanks.

Regards.
Jungtaek Lim (HeartSaVioR)


RE: QueueMetrics.AppsKilled/Failed metrics and failure reasons

2015-02-03 Thread Rohith Sharma K S
Hi

Could you give more information, which version of hadoop are you using?


 QueueMetrics.AppsKilled/Failed metrics shows much higher nos i.e ~100. 
 However RMAuditLogger shows 1 or 2 Apps as Killed/Failed in the logs.
May be I suspect that Logs might be rolled out. Does more applications are 
running?

All the applications history will be displayed  on RM web UI (provided RM is 
not restarted or RM recovery enabled). May be you can check these applications 
lists.

For finding reasons for application killed/failed, one way is you can check in 
NodeManager logs also. Here  you need to check using container_id for 
corresponding application.

Thanks  Regards
Rohith Sharma K S

From: Suma Shivaprasad [mailto:sumasai.shivapra...@gmail.com]
Sent: 03 February 2015 21:35
To: user@hadoop.apache.org; yarn-...@hadoop.apache.org
Subject: QueueMetrics.AppsKilled/Failed metrics and failure reasons

Hello,

Was trying to debug reasons for Killed/Failed apps and was checking for the 
applications that were killed/failed in RM logs - from RMAuditLogger.
QueueMetrics.AppsKilled/Failed metrics shows much higher nos i.e ~100. However 
RMAuditLogger shows 1 or 2 Apps as Killed/Failed in the logs. Is it possible 
that some logs are missed by AuditLogger or is it the other way round and 
metrics are being reported higher ?
Thanks
Suma


RE: QueueMetrics.AppsKilled/Failed metrics and failure reasons

2015-02-03 Thread Rohith Sharma K S
There are several ways to confirm from YARN that total number of Killed/Failed 
applications in cluster
1. Get from RM web UI lists OR
2. From admin try using this to get numbers of failed and killed applications: 
./yarn application -list -appStates FAILED,KILLED
3. Using client API's

Since metrics values are displayed in ganglia is incorrect, I get doubt that 
1. does ganglia is pointing out to correct RM cluster? Or 
2. what is the method ganglia uses to retrieve QueueMetrics? 
3. Any client program calculates you have written retrieve apps and calculate 
it?


Thanks  Regards
Rohith Sharma K S

-Original Message-
From: Suma Shivaprasad [mailto:sumasai.shivapra...@gmail.com] 
Sent: 04 February 2015 11:03
To: user@hadoop.apache.org
Cc: yarn-...@hadoop.apache.org
Subject: Re: QueueMetrics.AppsKilled/Failed metrics and failure reasons

Using hadoop 2.4.0. #of Applications running on average is small ~ 40 -60.
The metrics in Ganglia shows around around 10-30 apps killed every 5 mins which 
is very high wrt to the apps running at any given time(40-60). The RM logs 
though show 0 failed apps in audit logs during that hour.
The RM UI also doesnt show any apps in Applications-Failed tab . The logs are 
getting rolled over at a slower rate ..every 1-2 hours. Am searching for 
Application Finished - Failed to find the apps failed. Please let me know if 
I am missing something here.

Thanks
Suma




On Wed, Feb 4, 2015 at 10:03 AM, Rohith Sharma K S  rohithsharm...@huawei.com 
wrote:

  Hi



 Could you give more information, which version of hadoop are you using?



  QueueMetrics.AppsKilled/Failed metrics shows much higher nos i.e ~100.
 However RMAuditLogger shows 1 or 2 Apps as Killed/Failed in the logs.

 May be I suspect that Logs might be rolled out. Does more applications 
 are running?



 All the applications history will be displayed  on RM web UI (provided 
 RM is not restarted or RM recovery enabled). May be you can check 
 these applications lists.



 For finding reasons for application killed/failed, one way is you can 
 check in NodeManager logs also. Here  you need to check using 
 container_id for corresponding application.



 Thanks  Regards

 Rohith Sharma K S



 *From:* Suma Shivaprasad [mailto:sumasai.shivapra...@gmail.com]
 *Sent:* 03 February 2015 21:35
 *To:* user@hadoop.apache.org; yarn-...@hadoop.apache.org
 *Subject:* QueueMetrics.AppsKilled/Failed metrics and failure reasons



 Hello,


 Was trying to debug reasons for Killed/Failed apps and was checking 
 for the applications that were killed/failed in RM logs - from RMAuditLogger.

  QueueMetrics.AppsKilled/Failed metrics shows much higher nos i.e ~100.
 However RMAuditLogger shows 1 or 2 Apps as Killed/Failed in the logs. 
 Is it possible that some logs are missed by AuditLogger or is it the 
 other way round and metrics are being reported higher ?

 Thanks

 Suma



Pass lib jars when invoking an hadoop program

2015-02-03 Thread xeonmailinglist

Hi,

I am trying to run |distcp| using a java class, but I get the error of 
class not found |DistCpOptions|. I have used the argument |-libjars 
./share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar| to pass the jar file, 
but it seems that is not right. How I pass the lib properly?


Output:

|vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop jar wordcount.jar  -libjars 
./share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar /input1 /outputmp /output1
Configuration: core-default.xml, core-site.xml, mapred-default.xml, 
mapred-site.xml, yarn-default.xml, yarn-site.xml, hdfs-default.xml, 
hdfs-site.xml
-libjars
./share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar
/input1
/outputmp
/output1
Exception in thread main java.lang.NoClassDefFoundError: 
org/apache/hadoop/tools/DistCpOptions
at org.apache.hadoop.mapred.examples.WordCount.main(WordCount.java:101)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Caused by: java.lang.ClassNotFoundException: 
org.apache.hadoop.tools.DistCpOptions
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
|

My class:

|public static void main(String[] args) throws Exception {
Configuration conf = new Configuration();

String[] otherArgs = new GenericOptionsParser(conf, 
args).getRemainingArgs();
if (otherArgs.length  2) {
System.err.println(Usage: wordcount in [in...] out);
System.exit(2);
}

System.out.println(conf.toString());
for (int i = 0; i  args.length; i++) {
System.out.println(args[i]);
}

// distcp
String proto = webhdfs://;
String src = hadoop-coc-1/input1;
String dest = hadoop-coc-2/input1;
ListPath lsrc = new ArrayListPath();
lsrc.add(new Path(src));
DistCpOptions options = new DistCpOptions(lsrc, new Path(dest));
DistCp distcp = new DistCp(new Configuration(), options);
distcp.execute();
}
|

​


Re: tools.DistCp: Invalid arguments

2015-02-03 Thread xeonmailinglist

I have found the problem. I started to use `webhdfs` and everything is ok.


On 03-02-2015 10:40, xeonmailinglist wrote:


What do you mean by no path is given? Even if I launch this command, I 
get the same error…. What path should I put here?


|$ hadoop distcp hdfs://hadoop-coc-1:50070/input1 
hdfs://hadoop-coc-2:50070/input1|


Thanks,

On 02-02-2015 19:59, Alexander Alten-Lorenz wrote:


Have a closer look:


|hdfs://hadoop-coc-2:50070/|


No Path is given.


On 02 Feb 2015, at 20:52, xeonmailinglist xeonmailingl...@gmail.com 
mailto:xeonmailingl...@gmail.com wrote:


Hi,

I am trying to copy data using |distcp| but I get this error. Both 
hadoop runtime are working properly. Why is this happening?


|
vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop 
distcphdfs://hadoop-coc-1:50070/input1  hdfs://hadoop-coc-2:50070/
15/02/02 19:46:37 ERROR tools.DistCp: Invalid arguments:
java.io.IOException: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: 
Protocol message end-group tag did not match expected tag.; Host Details : local host is: 
hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
 at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
 at org.apache.hadoop.ipc.Client.call(Client.java:1472)
 at org.apache.hadoop.ipc.Client.call(Client.java:1399)
 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
 at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
 at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
 at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
 at 
org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
 at 
org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
 at 
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
 at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
 at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400)
 at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:188)
 at org.apache.hadoop.tools.DistCp.run(DistCp.java:111)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.hadoop.tools.DistCp.main(DistCp.java:401)
Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol message 
end-group tag did not match expected tag.
 at 
com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94)
 at 
com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124)
 at 
com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202)
 at 
com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241)
 at 
com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253)
 at 
com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259)
 at 
com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49)
 at 
org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:3167)
 at 
org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1072)
 at org.apache.hadoop.ipc.Client$Connection.run(Client.java:966)
Invalid arguments: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: 
Protocol message end-group tag did not match expected tag.; Host Details : local host is: 
hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
usage: distcp OPTIONS [source_path...] target_path
|

Thanks,

​



​




Re: tools.DistCp: Invalid arguments

2015-02-03 Thread Alexander Alten-Lorenz
Hi,

Can you please try webhdfs instead hdfs?

- Alexander 

 On 03 Feb 2015, at 12:05, xeonmailinglist xeonmailingl...@gmail.com wrote:
 
 Maybe this has to do with this error… I can’t do ls to my own machine using 
 the command below. Can this be related to the other problem? Shouldn't I list 
 the files with this command?
 vagrant@hadoop-coc-1:~$ hdfs dfs -ls hdfs://192.168.56.100/
 ls: Call From hadoop-coc-1/192.168.56.100 to hadoop-coc-1:8020 failed on 
 connection exception: java.net.ConnectException: Connection refused; For more 
 details see:  http://wiki.apache.org/hadoop/ConnectionRefused 
 http://wiki.apache.org/hadoop/ConnectionRefused
 On 02-02-2015 19:59, Alexander Alten-Lorenz wrote:
 
 
 
 Have a closer look:
 
 hdfs://hadoop-coc-2:50070/ hdfs://hadoop-coc-2:50070/
 No Path is given.
 
 
 On 02 Feb 2015, at 20:52, xeonmailinglist xeonmailingl...@gmail.com 
 mailto:xeonmailingl...@gmail.com wrote:
 
 Hi,
 
 I am trying to copy data using distcp but I get this error. Both hadoop 
 runtime are working properly. Why is this happening?
 
 
 vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop distcp 
 hdfs://hadoop-coc-1:50070/input1 hdfs://hadoop-coc-1:50070/input1 
 hdfs://hadoop-coc-2:50070/ hdfs://hadoop-coc-2:50070/
 15/02/02 19:46:37 ERROR tools.DistCp: Invalid arguments: 
 java.io.IOException: Failed on local exception: 
 com.google.protobuf.InvalidProtocolBufferException: Protocol message 
 end-group tag did not match expected tag.; Host Details : local host is: 
 hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070; 
 at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
 at org.apache.hadoop.ipc.Client.call(Client.java:1472)
 at org.apache.hadoop.ipc.Client.call(Client.java:1399)
 at 
 org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
 at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
 at 
 org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
 at 
 org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
 at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
 at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
 at 
 org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
 at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400)
 at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:188)
 at org.apache.hadoop.tools.DistCp.run(DistCp.java:111)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.hadoop.tools.DistCp.main(DistCp.java:401)
 Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol 
 message end-group tag did not match expected tag.
 at 
 com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94)
 at 
 com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124)
 at 
 com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202)
 at 
 com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49)
 at 
 org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:3167)
 at 
 org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1072)
 at org.apache.hadoop.ipc.Client$Connection.run(Client.java:966)
 Invalid arguments: Failed on local exception: 
 com.google.protobuf.InvalidProtocolBufferException: Protocol message 
 end-group tag did not match expected tag.; Host Details : local host is: 
 hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070; 
 usage: distcp OPTIONS [source_path...] target_path
 Thanks,
 
 
 
 



Re: tools.DistCp: Invalid arguments

2015-02-03 Thread Alexander Alten-Lorenz
Ah, good. Cross-posting :)

BR,
 Alex

 On 03 Feb 2015, at 12:41, xeonmailinglist xeonmailingl...@gmail.com wrote:
 
 I have found the problem. I started to use `webhdfs` and everything is ok.
 
 
 On 03-02-2015 10:40, xeonmailinglist wrote:
 What do you mean by no path is given? Even if I launch this command, I get 
 the same error…. What path should I put here?
 
 $ hadoop distcp
 hdfs://hadoop-coc-1:50070/input1
 hdfs://hadoop-coc-2:50070/input1
 
 Thanks,
 On 02-02-2015 19:59, Alexander Alten-Lorenz wrote:
 
 Have a closer look:
 
 hdfs://hadoop-coc-2:50070/ hdfs://hadoop-coc-2:50070/
 No Path is given.
 
 
 On 02 Feb 2015, at 20:52, xeonmailinglist xeonmailingl...@gmail.com 
 mailto:xeonmailingl...@gmail.com wrote:
 
 Hi,
 
 I am trying to copy data using distcp but I get this error. Both hadoop 
 runtime are working properly. Why is this happening?
 
 
 vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop distcp 
 hdfs://hadoop-coc-1:50070/input1 hdfs://hadoop-coc-1:50070/input1 
 hdfs://hadoop-coc-2:50070/ hdfs://hadoop-coc-2:50070/
 15/02/02 19:46:37 ERROR tools.DistCp: Invalid arguments: 
 java.io.IOException: Failed on local exception: 
 com.google.protobuf.InvalidProtocolBufferException: Protocol message 
 end-group tag did not match expected tag.; Host Details : local host is: 
 hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070; 
 at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
 at org.apache.hadoop.ipc.Client.call(Client.java:1472)
 at org.apache.hadoop.ipc.Client.call(Client.java:1399)
 at 
 org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
 at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
 at 
 org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
 at 
 org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
 at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
 at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
 at 
 org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
 at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400)
 at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:188)
 at org.apache.hadoop.tools.DistCp.run(DistCp.java:111)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.hadoop.tools.DistCp.main(DistCp.java:401)
 Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol 
 message end-group tag did not match expected tag.
 at 
 com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94)
 at 
 com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124)
 at 
 com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202)
 at 
 com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49)
 at 
 org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:3167)
 at 
 org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1072)
 at org.apache.hadoop.ipc.Client$Connection.run(Client.java:966)
 Invalid arguments: Failed on local exception: 
 com.google.protobuf.InvalidProtocolBufferException: Protocol message 
 end-group tag did not match expected tag.; Host Details : local host is: 
 hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070; 
 usage: distcp OPTIONS [source_path...] target_path
 Thanks,
 
 
 
 



Re: tools.DistCp: Invalid arguments

2015-02-03 Thread Artem Ervits
Another good option is hftp.

Artem Ervits
On Feb 3, 2015 6:42 AM, xeonmailinglist xeonmailingl...@gmail.com wrote:

  I have found the problem. I started to use `webhdfs` and everything is ok.


 On 03-02-2015 10:40, xeonmailinglist wrote:

  What do you mean by no path is given? Even if I launch this command, I
 get the same error…. What path should I put here?

 $ hadoop distcp hdfs://hadoop-coc-1:50070/input1
 hdfs://hadoop-coc-2:50070/input1

 Thanks,

 On 02-02-2015 19:59, Alexander Alten-Lorenz wrote:

 Have a closer look:

hdfs://hadoop-coc-2:50070/


  No Path is given.


  On 02 Feb 2015, at 20:52, xeonmailinglist xeonmailingl...@gmail.com
 wrote:

   Hi,

 I am trying to copy data using distcp but I get this error. Both hadoop
 runtime are working properly. Why is this happening?


 vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop distcp 
 hdfs://hadoop-coc-1:50070/input1 hdfs://hadoop-coc-2:50070/
 15/02/02 19:46:37 ERROR tools.DistCp: Invalid arguments:
 java.io.IOException: Failed on local exception: 
 com.google.protobuf.InvalidProtocolBufferException: Protocol message 
 end-group tag did not match expected tag.; Host Details : local host is: 
 hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
 at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
 at org.apache.hadoop.ipc.Client.call(Client.java:1472)
 at org.apache.hadoop.ipc.Client.call(Client.java:1399)
 at 
 org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
 at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
 at 
 org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
 at 
 org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
 at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
 at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
 at 
 org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
 at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400)
 at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:188)
 at org.apache.hadoop.tools.DistCp.run(DistCp.java:111)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.hadoop.tools.DistCp.main(DistCp.java:401)
 Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol 
 message end-group tag did not match expected tag.
 at 
 com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94)
 at 
 com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124)
 at 
 com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202)
 at 
 com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49)
 at 
 org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:3167)
 at 
 org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1072)
 at org.apache.hadoop.ipc.Client$Connection.run(Client.java:966)
 Invalid arguments: Failed on local exception: 
 com.google.protobuf.InvalidProtocolBufferException: Protocol message 
 end-group tag did not match expected tag.; Host Details : local host is: 
 hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
 usage: distcp OPTIONS [source_path...] target_path

 Thanks,
 ​


  ​





Can not start HA namenode with security enabled

2015-02-03 Thread 郝东
I am converting a secure non-HA cluster into a secure HA cluster. After the 
configuration and started all the journalnodes, I executed the following 
commands on the original NameNode:
1. hdfs name -initializeSharedEdits #this step succeeded
2. hadoop-daemon.sh start namenode # this step failed.

The namenode did not start successfully. I verified that my principals are 
right. And I checked the DNS is configured correctly so that I could use the 
nslookup command to lookup and reverse-lookup the Namenode and JournalNodes.

I also checked the logs. The JournalNodes did not report any ERROR. The 
Namenode Log report some ERRORs, but I still could not find the reason 
according to these ERRORS.

In the following I listed the main part of my hdfs-site.xml and the error log 
from my Namenode.  Could anyone help me to figure it out?


Many Thanks!

**The main part of my hdfs-site.xml*

property
namedfs.nameservices/name
valuebgdt-dev-hrb/value
/property

property
namedfs.ha.namenodes.bgdt-dev-hrb/name
valuenn1,nn2/value
/property

property
namedfs.namenode.rpc-address.bgdt-dev-hrb.nn1/name
valuebgdt01.dev.hrb:9000/value
/property

property
namedfs.namenode.rpc-address.bgdt-dev-hrb.nn2/name
valuebgdt02.dev.hrb:9000/value
/property

property
namedfs.namenode.shared.edits.dir/name
valueqjournal://bgdt01.dev.hrb:8485;bgdt03.dev.hrb:8485;bgdt04.dev.hrb:8485/bgdt-dev-hrb/value
/property

property
namedfs.client.failover.proxy.provider.bgdt-dev-hrb/name
valueorg.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider/value
/property

property
namedfs.ha.fencing.methods/name
valuesshfence
shell(/bin/true)
/value
/property

property
namedfs.ha.fencing.ssh.private-key-files/name
value/home/hadoop/.ssh/id_rsa/value
/property

property
namedfs.journalnode.edits.dir/name
value/bgdt/hadoop/hdfs/jn/value
/property

property
namedfs.permissions.enabled/name
valuetrue/value
/property
property
namedfs.namenode.name.dir/name
valuefile:///bgdt/hadoop/hdfs/nn/value
finaltrue/final
/property
property
namedfs.datanode.name.dir/name
valuefile:///bgdt/hadoop/hdfs/dn/value
/property

property
namedfs.namenode.http-address.bgdt-dev-hrb.nn1/name
valuebgdt01.dev.hrb:50070/value
/property

property
namedfs.namenode.http-address.bgdt-dev-hrb.nn2/name
valuebgdt02.dev.hrb:50070/value
/property

property
namedfs.permissions.superusergroup/name
valuebgdtgrp/value
/property

property
namedfs.block.access.token.enable/name
valuetrue/value
/property

property
namedfs.http.policy/name
valueHTTP_ONLY/value
/property

property
namedfs.namenode.https-address.bgdt-dev-hrb.nn1/name
valuebgdt01.dev.hrb:50470/value
/property

property
namedfs.namenode.https-address.bgdt-dev-hrb.nn2/name
valuebgdt02.dev.hrb:50470/value
/property

property
namedfs.namenode.keytab.file/name
value/etc/hadoop/keytab/hadoop.service.keytab/value
/property
property
namedfs.namenode.kerberos.principal/name
valuehdfs/_h...@bgdt.dev.hrb/value
/property
property
namedfs.namenode.kerberos.https.principal/name
valuehost/_h...@bgdt.dev.hrb/value
/property

property
namedfs.webhdfs.enabled/name
valuetrue/value
/property

property
namedfs.web.authentication.kerberos.principal/name
valuehttp/_h...@bgdt.dev.hrb/value
/property

property
namedfs.web.authentication.kerberos.keytab/name
value/etc/hadoop/keytab/hadoop.service.keytab/value
/property

property
namedfs.journalnode.kerberos.principal/name
valuehdfs/_h...@bgdt.dev.hrb/value
/property

property
namedfs.journalnode.kerberos.https.principal/name
valuehost/_h...@bgdt.dev.hrb/value
/property

property
namedfs.journalnode.kerberos.internal.spnego.principal/name
valuehttp/_h...@bgdt.dev.hrb/value
/property

property
namedfs.journalnode.keytab.file/name
value/etc/hadoop/keytab/hadoop.service.keytab/value
/property

*The Error Log from the 
Namenode**


2015-02-03 17:42:06,020 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: 
Start loading edits file 
http://bgdt04.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3,
 
http://bgdt01.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3
2015-02-03 17:42:06,024 INFO 
org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding 
stream 
'http://bgdt04.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3,
 
http://bgdt01.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3'
 to transaction ID 68994
2015-02-03 17:42:06,024 INFO 
org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding 
stream 
'http://bgdt04.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3'
 to transaction 

QueueMetrics.AppsKilled/Failed metrics and failure reasons

2015-02-03 Thread Suma Shivaprasad
Hello,

Was trying to debug reasons for Killed/Failed apps and was checking for the
applications that were killed/failed in RM logs - from RMAuditLogger.
QueueMetrics.AppsKilled/Failed metrics shows much higher nos i.e ~100.
However RMAuditLogger shows 1 or 2 Apps as Killed/Failed in the logs. Is it
possible that some logs are missed by AuditLogger or is it the other way
round and metrics are being reported higher ?

Thanks
Suma


IBM JAVA and KerberosTestUtils

2015-02-03 Thread Sangamesh Mallayya
Hi All,

This is with respect to the JIRA defect: HADOOP-10774 related to kerberose 
authentication using IBM JAVA.
Looks like there are lot of changes haven been done to properly handle the 
kerbserose authentication using the JIRA defect: HADOOP-9446 for IBM JAVA.
But their are still some failures can been seen in Apache Hadoop Common 
tests in case of IBM JAVA.
Available patch for HADOOP-10774 will solve the authentication issues plus 
the path issues.

Please let me know how to proceed on this issue or let me know the right 
contact or mailing list.

Thanks  Regards,
-Sangamesh
AIX BOS Development
Bangalore, EGL D Block, 6th floor

Phone: +91 (80) 417-76965

Re: Can not start HA namenode with security enabled

2015-02-03 Thread Manoj Samel
Have you added all host specific principals in kerberos database ?

Thanks,

On Tue, Feb 3, 2015 at 7:59 AM, 郝东 donhof...@163.com wrote:

 I am converting a secure non-HA cluster into a secure HA cluster. After
 the configuration and started all the journalnodes, I executed the
 following commands on the original NameNode:
 1. hdfs name -initializeSharedEdits #this step succeeded
 2. hadoop-daemon.sh start namenode # this step failed.

 The namenode did not start successfully. I verified that my principals are
 right. And I checked the DNS is configured correctly so that I could use
 the nslookup command to lookup and reverse-lookup the Namenode and
 JournalNodes.

 I also checked the logs. The JournalNodes did not report any ERROR. The
 Namenode Log report some ERRORs, but I still could not find the reason
 according to these ERRORS.

 In the following I listed the main part of my hdfs-site.xml and the error
 log from my Namenode.  Could anyone help me to figure it out?

 Many Thanks!

 **The main part of my hdfs-site.xml*

 property
 namedfs.nameservices/name
 valuebgdt-dev-hrb/value
 /property

 property
 namedfs.ha.namenodes.bgdt-dev-hrb/name
 valuenn1,nn2/value
 /property

 property
 namedfs.namenode.rpc-address.bgdt-dev-hrb.nn1/name
 valuebgdt01.dev.hrb:9000/value
 /property

 property
 namedfs.namenode.rpc-address.bgdt-dev-hrb.nn2/name
 valuebgdt02.dev.hrb:9000/value
 /property

 property
 namedfs.namenode.shared.edits.dir/name

 valueqjournal://bgdt01.dev.hrb:8485;bgdt03.dev.hrb:8485;bgdt04.dev.hrb:8485/bgdt-dev-hrb/value
 /property

 property
 namedfs.client.failover.proxy.provider.bgdt-dev-hrb/name

 valueorg.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider/value
 /property

 property
 namedfs.ha.fencing.methods/name
 valuesshfence
 shell(/bin/true)
 /value
 /property

 property
 namedfs.ha.fencing.ssh.private-key-files/name
 value/home/hadoop/.ssh/id_rsa/value
 /property

 property
 namedfs.journalnode.edits.dir/name
 value/bgdt/hadoop/hdfs/jn/value
 /property

 property
 namedfs.permissions.enabled/name
 valuetrue/value
 /property
 property
 namedfs.namenode.name.dir/name
 valuefile:///bgdt/hadoop/hdfs/nn/value
 finaltrue/final
 /property
 property
 namedfs.datanode.name.dir/name
 valuefile:///bgdt/hadoop/hdfs/dn/value
 /property

 property
 namedfs.namenode.http-address.bgdt-dev-hrb.nn1/name
 valuebgdt01.dev.hrb:50070/value
 /property

 property
 namedfs.namenode.http-address.bgdt-dev-hrb.nn2/name
 valuebgdt02.dev.hrb:50070/value
 /property

 property
 namedfs.permissions.superusergroup/name
 valuebgdtgrp/value
 /property

 property
 namedfs.block.access.token.enable/name
 valuetrue/value
 /property

 property
 namedfs.http.policy/name
 valueHTTP_ONLY/value
 /property

 property
 namedfs.namenode.https-address.bgdt-dev-hrb.nn1/name
 valuebgdt01.dev.hrb:50470/value
 /property

 property
 namedfs.namenode.https-address.bgdt-dev-hrb.nn2/name
 valuebgdt02.dev.hrb:50470/value
 /property

 property
 namedfs.namenode.keytab.file/name
 value/etc/hadoop/keytab/hadoop.service.keytab/value
 /property
 property
 namedfs.namenode.kerberos.principal/name
 valuehdfs/_h...@bgdt.dev.hrb/value
 /property
 property
 namedfs.namenode.kerberos.https.principal/name
 valuehost/_h...@bgdt.dev.hrb/value
 /property

 property
 namedfs.webhdfs.enabled/name
 valuetrue/value
 /property

 property
 namedfs.web.authentication.kerberos.principal/name
 valuehttp/_h...@bgdt.dev.hrb/value
 /property

 property
 namedfs.web.authentication.kerberos.keytab/name
 value/etc/hadoop/keytab/hadoop.service.keytab/value
 /property

 property
 namedfs.journalnode.kerberos.principal/name
 valuehdfs/_h...@bgdt.dev.hrb/value
 /property

 property
 namedfs.journalnode.kerberos.https.principal/name
 valuehost/_h...@bgdt.dev.hrb/value
 /property

 property
 namedfs.journalnode.kerberos.internal.spnego.principal/name
 valuehttp/_h...@bgdt.dev.hrb/value
 /property

 property
 namedfs.journalnode.keytab.file/name
 value/etc/hadoop/keytab/hadoop.service.keytab/value
 /property

 *The Error Log from the
 Namenode**

 2015-02-03 17:42:06,020 INFO
 org.apache.hadoop.hdfs.server.namenode.FSImage: Start loading edits file
 http://bgdt04.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3,

 http://bgdt01.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3
 2015-02-03 17:42:06,024 INFO
 org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding
 stream '
 http://bgdt04.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3,

 http://bgdt01.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3'
 to transaction ID 68994
 

Re: Pass lib jars when invoking an hadoop program

2015-02-03 Thread xeonmailinglist

Got it. Here's the solution:

```
vagrant@hadoop-coc-1:~/Programs/hadoop$ export 
HADOOP_CLASSPATH=share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar; hadoop 
jar wordcount.jar -libjars 
$HADOOP_HOME/share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar /input1 
/outputmp /output1

```

On 03-02-2015 14:58, xeonmailinglist wrote:


Hi,

I am trying to run |distcp| using a java class, but I get the error of 
class not found |DistCpOptions|. I have used the argument |-libjars 
./share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar| to pass the jar 
file, but it seems that is not right. How I pass the lib properly?


Output:

|vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop jar wordcount.jar  -libjars 
./share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar /input1 /outputmp /output1
Configuration: core-default.xml, core-site.xml, mapred-default.xml, 
mapred-site.xml, yarn-default.xml, yarn-site.xml, hdfs-default.xml, 
hdfs-site.xml
-libjars
./share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar
/input1
/outputmp
/output1
Exception in thread main java.lang.NoClassDefFoundError: 
org/apache/hadoop/tools/DistCpOptions
 at org.apache.hadoop.mapred.examples.WordCount.main(WordCount.java:101)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Caused by: java.lang.ClassNotFoundException: 
org.apache.hadoop.tools.DistCpOptions
 at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
 at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
 at java.security.AccessController.doPrivileged(Native Method)
 at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
|

My class:

|public static void main(String[] args) throws Exception {
 Configuration conf = new Configuration();

 String[] otherArgs = new GenericOptionsParser(conf, 
args).getRemainingArgs();
 if (otherArgs.length  2) {
 System.err.println(Usage: wordcount in [in...] out);
 System.exit(2);
 }

 System.out.println(conf.toString());
 for (int i = 0; i  args.length; i++) {
 System.out.println(args[i]);
 }

 // distcp
 String proto = webhdfs://;
 String src = hadoop-coc-1/input1;
 String dest = hadoop-coc-2/input1;
 ListPath lsrc = new ArrayListPath();
 lsrc.add(new Path(src));
 DistCpOptions options = new DistCpOptions(lsrc, new Path(dest));
 DistCp distcp = new DistCp(new Configuration(), options);
 distcp.execute();
 }
|
​




Re: Pass lib jars when invoking an hadoop program

2015-02-03 Thread xeonmailinglist

Got it. Here’s the solution:

|vagrant@hadoop-coc-1:~/Programs/hadoop$ export 
HADOOP_CLASSPATH=share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar; hadoop jar 
wordcount.jar -libjars 
$HADOOP_HOME/share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar /input1 /outputmp 
/output1
|

On 03-02-2015 14:58, xeonmailinglist wrote:


Hi,

I am trying to run |distcp| using a java class, but I get the error of 
class not found |DistCpOptions|. I have used the argument |-libjars 
./share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar| to pass the jar 
file, but it seems that is not right. How I pass the lib properly?


Output:

|vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop jar wordcount.jar  -libjars 
./share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar /input1 /outputmp /output1
Configuration: core-default.xml, core-site.xml, mapred-default.xml, 
mapred-site.xml, yarn-default.xml, yarn-site.xml, hdfs-default.xml, 
hdfs-site.xml
-libjars
./share/hadoop/tools/lib/hadoop-distcp-2.6.0.jar
/input1
/outputmp
/output1
Exception in thread main java.lang.NoClassDefFoundError: 
org/apache/hadoop/tools/DistCpOptions
 at org.apache.hadoop.mapred.examples.WordCount.main(WordCount.java:101)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Caused by: java.lang.ClassNotFoundException: 
org.apache.hadoop.tools.DistCpOptions
 at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
 at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
 at java.security.AccessController.doPrivileged(Native Method)
 at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
|

My class:

|public static void main(String[] args) throws Exception {
 Configuration conf = new Configuration();

 String[] otherArgs = new GenericOptionsParser(conf, 
args).getRemainingArgs();
 if (otherArgs.length  2) {
 System.err.println(Usage: wordcount in [in...] out);
 System.exit(2);
 }

 System.out.println(conf.toString());
 for (int i = 0; i  args.length; i++) {
 System.out.println(args[i]);
 }

 // distcp
 String proto = webhdfs://;
 String src = hadoop-coc-1/input1;
 String dest = hadoop-coc-2/input1;
 ListPath lsrc = new ArrayListPath();
 lsrc.add(new Path(src));
 DistCpOptions options = new DistCpOptions(lsrc, new Path(dest));
 DistCp distcp = new DistCp(new Configuration(), options);
 distcp.execute();
 }
|
​


​


create Job with java code.

2015-02-03 Thread xeonmailinglist

Hi,

I want this because I want to create depency between 2 jobs. The first 
job execute the wordcount example, and the second job copy the output of 
the wordcount to another HDFS.
Therefore, I want to create a job (job 2) that includes the code to copy 
data to another HDFS. The code is below. Is there a way to do it with 
the API?


| // Put inside a job()
String proto = webhdfs://;
String src = hadoop-coc-1/input1;
String dest = hadoop-coc-2/input1;
ListPath lsrc = new ArrayListPath();
lsrc.add(new Path(proto+src));
DistCpOptions options = new DistCpOptions(lsrc, new Path(proto+dest));
DistCp distcp = new DistCp(conf, options);
distcp.execute();
|

​


Re: tools.DistCp: Invalid arguments

2015-02-03 Thread xeonmailinglist
Maybe this has to do with this error… I can’t do |ls| to my own machine 
using the command below. Can this be related to the other problem? 
Shouldn't I list the files with this command?


|vagrant@hadoop-coc-1:~$ hdfs dfs -ls hdfs://192.168.56.100/
ls: Call From hadoop-coc-1/192.168.56.100 to hadoop-coc-1:8020 failed on 
connection exception: java.net.ConnectException: Connection refused; For more 
details see:  http://wiki.apache.org/hadoop/ConnectionRefused
|

On 02-02-2015 19:59, Alexander Alten-Lorenz wrote:


Have a closer look:


|hdfs://hadoop-coc-2:50070/|


No Path is given.


On 02 Feb 2015, at 20:52, xeonmailinglist xeonmailingl...@gmail.com 
mailto:xeonmailingl...@gmail.com wrote:


Hi,

I am trying to copy data using |distcp| but I get this error. Both 
hadoop runtime are working properly. Why is this happening?


|
vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop 
distcphdfs://hadoop-coc-1:50070/input1  hdfs://hadoop-coc-2:50070/
15/02/02 19:46:37 ERROR tools.DistCp: Invalid arguments:
java.io.IOException: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: 
Protocol message end-group tag did not match expected tag.; Host Details : local host is: 
hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
 at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
 at org.apache.hadoop.ipc.Client.call(Client.java:1472)
 at org.apache.hadoop.ipc.Client.call(Client.java:1399)
 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
 at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
 at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
 at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
 at 
org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
 at 
org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
 at 
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
 at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
 at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400)
 at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:188)
 at org.apache.hadoop.tools.DistCp.run(DistCp.java:111)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.hadoop.tools.DistCp.main(DistCp.java:401)
Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol message 
end-group tag did not match expected tag.
 at 
com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94)
 at 
com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124)
 at 
com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202)
 at 
com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241)
 at 
com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253)
 at 
com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259)
 at 
com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49)
 at 
org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:3167)
 at 
org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1072)
 at org.apache.hadoop.ipc.Client$Connection.run(Client.java:966)
Invalid arguments: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: 
Protocol message end-group tag did not match expected tag.; Host Details : local host is: 
hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
usage: distcp OPTIONS [source_path...] target_path
|

Thanks,

​



​


Re: tools.DistCp: Invalid arguments

2015-02-03 Thread xeonmailinglist
What do you mean by no path is given? Even if I launch this command, I 
get the same error…. What path should I put here?


|$ hadoop distcp hdfs://hadoop-coc-1:50070/input1 
hdfs://hadoop-coc-2:50070/input1|


Thanks,

On 02-02-2015 19:59, Alexander Alten-Lorenz wrote:


Have a closer look:


|hdfs://hadoop-coc-2:50070/|


No Path is given.


On 02 Feb 2015, at 20:52, xeonmailinglist xeonmailingl...@gmail.com 
mailto:xeonmailingl...@gmail.com wrote:


Hi,

I am trying to copy data using |distcp| but I get this error. Both 
hadoop runtime are working properly. Why is this happening?


|
vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop 
distcphdfs://hadoop-coc-1:50070/input1  hdfs://hadoop-coc-2:50070/
15/02/02 19:46:37 ERROR tools.DistCp: Invalid arguments:
java.io.IOException: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: 
Protocol message end-group tag did not match expected tag.; Host Details : local host is: 
hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
 at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
 at org.apache.hadoop.ipc.Client.call(Client.java:1472)
 at org.apache.hadoop.ipc.Client.call(Client.java:1399)
 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
 at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
 at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
 at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
 at 
org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
 at 
org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
 at 
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
 at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
 at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400)
 at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:188)
 at org.apache.hadoop.tools.DistCp.run(DistCp.java:111)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.hadoop.tools.DistCp.main(DistCp.java:401)
Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol message 
end-group tag did not match expected tag.
 at 
com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94)
 at 
com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124)
 at 
com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202)
 at 
com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241)
 at 
com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253)
 at 
com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259)
 at 
com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49)
 at 
org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:3167)
 at 
org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1072)
 at org.apache.hadoop.ipc.Client$Connection.run(Client.java:966)
Invalid arguments: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: 
Protocol message end-group tag did not match expected tag.; Host Details : local host is: 
hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
usage: distcp OPTIONS [source_path...] target_path
|

Thanks,

​



​


Re: unsubscribe

2015-02-03 Thread Ram Kumar
Check http://hadoop.apache.org/mailing_lists.html#User

Regards,
Ramkumar Bashyam

On Wed, Jan 7, 2015 at 7:01 PM, Kiran Prasad Gorigay 
kiranprasa...@imimobile.com wrote:

unsubscribe






How to rolling upgrade??

2015-02-03 Thread Mr.J
my cluster A, and cluster B. To upgrade to version 2.6
In what order should I upgrade? 
Journalnode 1 gt;gt; Journalnode 2 gt; Journalnode 3 gt;gt; Namenode Std 
gt;gt; Namenode Act gt;gt; Datanode   ??
Do I also need to upgrade the zookeeper? 
 
hadoop-2.4.1 :  JournalNode, Namenode, Datanode
zookeeper-3.4.6 : zookeeper
 
http://hadoop.apache.org/docs/r2.6.0/hadoop-project-dist/hadoop-hdfs/HdfsRollingUpgrade.html
 
 
 
 
 
 

 


unsubscribe

2015-02-03 Thread Kiran Prasad Gorigay
unsubscribe



Re:Re: Can not start HA namenode with security enabled

2015-02-03 Thread 郝东
Hi,
I have checked my kerberos database. All the principals are there. By the way, 
if I did not enable HA, just enable the secure-mode, the Namenode can be 
started correctly.







At 2015-02-04 01:24:21, Manoj Samel manojsamelt...@gmail.com wrote:

Have you added all host specific principals in kerberos database ?


Thanks,


On Tue, Feb 3, 2015 at 7:59 AM, 郝东 donhof...@163.com wrote:


I am converting a secure non-HA cluster into a secure HA cluster. After the 
configuration and started all the journalnodes, I executed the following 
commands on the original NameNode:
1. hdfs name -initializeSharedEdits #this step succeeded
2. hadoop-daemon.sh start namenode # this step failed.

The namenode did not start successfully. I verified that my principals are 
right. And I checked the DNS is configured correctly so that I could use the 
nslookup command to lookup and reverse-lookup the Namenode and JournalNodes.

I also checked the logs. The JournalNodes did not report any ERROR. The 
Namenode Log report some ERRORs, but I still could not find the reason 
according to these ERRORS.

In the following I listed the main part of my hdfs-site.xml and the error log 
from my Namenode.  Could anyone help me to figure it out?


Many Thanks!

**The main part of my hdfs-site.xml*

property
namedfs.nameservices/name
valuebgdt-dev-hrb/value
/property

property
namedfs.ha.namenodes.bgdt-dev-hrb/name
valuenn1,nn2/value
/property

property
namedfs.namenode.rpc-address.bgdt-dev-hrb.nn1/name
valuebgdt01.dev.hrb:9000/value
/property

property
namedfs.namenode.rpc-address.bgdt-dev-hrb.nn2/name
valuebgdt02.dev.hrb:9000/value
/property

property
namedfs.namenode.shared.edits.dir/name
valueqjournal://bgdt01.dev.hrb:8485;bgdt03.dev.hrb:8485;bgdt04.dev.hrb:8485/bgdt-dev-hrb/value
/property

property
namedfs.client.failover.proxy.provider.bgdt-dev-hrb/name
valueorg.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider/value
/property

property
namedfs.ha.fencing.methods/name
valuesshfence
shell(/bin/true)
/value
/property

property
namedfs.ha.fencing.ssh.private-key-files/name
value/home/hadoop/.ssh/id_rsa/value
/property

property
namedfs.journalnode.edits.dir/name
value/bgdt/hadoop/hdfs/jn/value
/property

property
namedfs.permissions.enabled/name
valuetrue/value
/property
property
namedfs.namenode.name.dir/name
valuefile:///bgdt/hadoop/hdfs/nn/value
finaltrue/final
/property
property
namedfs.datanode.name.dir/name
valuefile:///bgdt/hadoop/hdfs/dn/value
/property

property
namedfs.namenode.http-address.bgdt-dev-hrb.nn1/name
valuebgdt01.dev.hrb:50070/value
/property

property
namedfs.namenode.http-address.bgdt-dev-hrb.nn2/name
valuebgdt02.dev.hrb:50070/value
/property

property
namedfs.permissions.superusergroup/name
valuebgdtgrp/value
/property

property
namedfs.block.access.token.enable/name
valuetrue/value
/property

property
namedfs.http.policy/name
valueHTTP_ONLY/value
/property

property
namedfs.namenode.https-address.bgdt-dev-hrb.nn1/name
valuebgdt01.dev.hrb:50470/value
/property

property
namedfs.namenode.https-address.bgdt-dev-hrb.nn2/name
valuebgdt02.dev.hrb:50470/value
/property

property
namedfs.namenode.keytab.file/name
value/etc/hadoop/keytab/hadoop.service.keytab/value
/property
property
namedfs.namenode.kerberos.principal/name
valuehdfs/_h...@bgdt.dev.hrb/value
/property
property
namedfs.namenode.kerberos.https.principal/name
valuehost/_h...@bgdt.dev.hrb/value
/property

property
namedfs.webhdfs.enabled/name
valuetrue/value
/property

property
namedfs.web.authentication.kerberos.principal/name
valuehttp/_h...@bgdt.dev.hrb/value
/property

property
namedfs.web.authentication.kerberos.keytab/name
value/etc/hadoop/keytab/hadoop.service.keytab/value
/property

property
namedfs.journalnode.kerberos.principal/name
valuehdfs/_h...@bgdt.dev.hrb/value
/property

property
namedfs.journalnode.kerberos.https.principal/name
valuehost/_h...@bgdt.dev.hrb/value
/property

property
namedfs.journalnode.kerberos.internal.spnego.principal/name
valuehttp/_h...@bgdt.dev.hrb/value
/property

property
namedfs.journalnode.keytab.file/name
value/etc/hadoop/keytab/hadoop.service.keytab/value
/property

*The Error Log from the 
Namenode**


2015-02-03 17:42:06,020 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: 
Start loading edits file 
http://bgdt04.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3,
 
http://bgdt01.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3
2015-02-03 17:42:06,024 INFO 
org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding 
stream 
'http://bgdt04.dev.hrb:8480/getJournal?jid=bgdt-dev-hrbsegmentTxId=68994storageInfo=-57%3A876630880%3A0%3ACID-ea4c77aa-882d-4adf-a347-42f1344421f3,