Re: Large number of application attempts

2015-07-08 Thread Akmal Abbasov
 But I’ve noticed that there are ~4000 znodes in 
 /hadoop-ha/testhbase1/rmstore/ZKRMStateRoot/RMAppRoot in Zookeeper.
Can I manually remove this znodes, or do I need to do this through yarn command 
line?

Thank you.

 On 07 Jul 2015, at 10:23, Akmal Abbasov akmal.abba...@icloud.com wrote:
 
 Hi all,
 I have HBase cluster. 
 I run a cron job each hour to create a snapshot of tables in HBase and export 
 these snapshots to another HBase cluster for backup. 
 But I’ve noticed that there are ~4000 znodes in 
 /hadoop-ha/testhbase1/rmstore/ZKRMStateRoot/RMAppRoot in Zookeeper.
 What are these znodes? 
 Are they failed yarn applications?
 How I need to handle them?
 
 Thank you.
 Akmal



Re: Large number of application attempts

2015-07-08 Thread Akmal Abbasov
Hi,
Are you running a lots of apps ?
I have 5 tables in HBase, and I run a cron job each hour which will create a 
snapshot of each table, 
and tries to export it to a remote HBase cluster. 

  Or could you, maybe, share some of file names of those znodes, please ?
they are all have the name format as follows ‘application_x_’

Thank you.

 On 07 Jul 2015, at 19:00, Xuan Gong xg...@hortonworks.com wrote:
 
 Hey, Akmal:
Are you running a lots of apps ? Or could you, maybe, share some of
 file names of those znodes, please ?
 
 
 Thanks
 
 Xuan Gong
 On 7/7/15, 1:23 AM, Akmal Abbasov akmal.abba...@icloud.com wrote:
 
 Hi all,
 I have HBase cluster.
 I run a cron job each hour to create a snapshot of tables in HBase and
 export these snapshots to another HBase cluster for backup.
 But I¹ve noticed that there are ~4000 znodes in
 /hadoop-ha/testhbase1/rmstore/ZKRMStateRoot/RMAppRoot in Zookeeper.
 What are these znodes?
 Are they failed yarn applications?
 How I need to handle them?
 
 Thank you.
 Akmal
 



Re: Large number of application attempts

2015-07-08 Thread Xuan Gong
bq. Can I manually remove this znodes, or do I need to do this through
yarn command line?

  Which version of hadoop you are using ? If the version is 2.6 or after
2.6, you could do yarn resourcemanager -format-state-store which will
format all the rmstore (basically delete path
ha/testhbase1/rmstore/ZKRMStateRoot and all the subdir/files)

Thanks

Xuan Gong


On 7/8/15, 5:45 AM, Akmal Abbasov akmal.abba...@icloud.com wrote:

 But I¹ve noticed that there are ~4000 znodes in
/hadoop-ha/testhbase1/rmstore/ZKRMStateRoot/RMAppRoot in Zookeeper.
Can I manually remove this znodes, or do I need to do this through yarn
command line?

Thank you.

 On 07 Jul 2015, at 10:23, Akmal Abbasov akmal.abba...@icloud.com
wrote:
 
 Hi all,
 I have HBase cluster.
 I run a cron job each hour to create a snapshot of tables in HBase and
export these snapshots to another HBase cluster for backup.
 But I¹ve noticed that there are ~4000 znodes in
/hadoop-ha/testhbase1/rmstore/ZKRMStateRoot/RMAppRoot in Zookeeper.
 What are these znodes?
 Are they failed yarn applications?
 How I need to handle them?
 
 Thank you.
 Akmal




Different outputformats in avro map reduce job

2015-07-08 Thread Nishanth S
Hi All,

I  have a map reduce job which  reads a binary file and  needs to output
multiple avro files and a textformat file.I was able to output multiplle
avro files using (Avromultipleouts).How would I modify the job to output
textformat as well along with these avro files.Is it possible.

Thanks,
Nishanth


Re: tools.DistCp: Invalid arguments

2015-07-08 Thread Giri P
will distcp do checksum after it copied the data to target ?

On Tue, Feb 3, 2015 at 4:15 AM, Artem Ervits artemerv...@gmail.com wrote:

 Another good option is hftp.

 Artem Ervits
 On Feb 3, 2015 6:42 AM, xeonmailinglist xeonmailingl...@gmail.com
 wrote:

  I have found the problem. I started to use `webhdfs` and everything is
 ok.


 On 03-02-2015 10:40, xeonmailinglist wrote:

  What do you mean by no path is given? Even if I launch this command, I
 get the same error…. What path should I put here?

 $ hadoop distcp hdfs://hadoop-coc-1:50070/input1
 hdfs://hadoop-coc-2:50070/input1

 Thanks,

 On 02-02-2015 19:59, Alexander Alten-Lorenz wrote:

 Have a closer look:

hdfs://hadoop-coc-2:50070/


  No Path is given.


  On 02 Feb 2015, at 20:52, xeonmailinglist xeonmailingl...@gmail.com
 wrote:

   Hi,

 I am trying to copy data using distcp but I get this error. Both hadoop
 runtime are working properly. Why is this happening?


 vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop distcp 
 hdfs://hadoop-coc-1:50070/input1 hdfs://hadoop-coc-2:50070/
 15/02/02 19:46:37 ERROR tools.DistCp: Invalid arguments:
 java.io.IOException: Failed on local exception: 
 com.google.protobuf.InvalidProtocolBufferException: Protocol message 
 end-group tag did not match expected tag.; Host Details : local host is: 
 hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
 at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
 at org.apache.hadoop.ipc.Client.call(Client.java:1472)
 at org.apache.hadoop.ipc.Client.call(Client.java:1399)
 at 
 org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
 at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
 at 
 org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
 at 
 org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
 at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
 at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
 at 
 org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
 at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400)
 at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:188)
 at org.apache.hadoop.tools.DistCp.run(DistCp.java:111)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.hadoop.tools.DistCp.main(DistCp.java:401)
 Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol 
 message end-group tag did not match expected tag.
 at 
 com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94)
 at 
 com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124)
 at 
 com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202)
 at 
 com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259)
 at 
 com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49)
 at 
 org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:3167)
 at 
 org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1072)
 at org.apache.hadoop.ipc.Client$Connection.run(Client.java:966)
 Invalid arguments: Failed on local exception: 
 com.google.protobuf.InvalidProtocolBufferException: Protocol message 
 end-group tag did not match expected tag.; Host Details : local host is: 
 hadoop-coc-1/127.0.1.1; destination host is: hadoop-coc-2:50070;
 usage: distcp OPTIONS [source_path...] target_path

 Thanks,
 ​


  ​





Re: Different outputformats in avro map reduce job

2015-07-08 Thread Harsh J
You can write out the text file by using a direct HDFS file writer.
Your only concern in doing this approach would be the use of proper
target directories, which is documented at
http://wiki.apache.org/hadoop/FAQ#Can_I_write_create.2Fwrite-to_hdfs_files_directly_from_map.2Freduce_tasks.3F.

On Thu, Jul 9, 2015 at 2:40 AM, Nishanth S chinchu2...@gmail.com wrote:
 Hi All,

 I  have a map reduce job which  reads a binary file and  needs to output
 multiple avro files and a textformat file.I was able to output multiplle
 avro files using (Avromultipleouts).How would I modify the job to output
 textformat as well along with these avro files.Is it possible.

 Thanks,
 Nishanth



-- 
Harsh J


Re: share the same namespace in 2 YARN instances?

2015-07-08 Thread Anoop Dawar
Xeon

I believe there is a YARN federation project going on and another project
Apache Myriad that you may want to look at.

YARN Federation https://issues.apache.org/jira/browse/YARN-2915

Apache Myriad http://incubator.apache.org/projects/myriad.html

 --
 Thanks,

 Anoop Dawar
 Senior Director, Product Management
 m: 408 464 2273
 skypeid: anooprdawar

www.mapr.com

On Mon, Mar 2, 2015 at 7:12 AM, xeonmailinglist xeonmailingl...@gmail.com
wrote:

 Hi,

 I was reading about Federation of HDFS, which is possible in YARN (
 http://www.devx.com/opensource/enhance-existing-hdfs-architecture-with-hadoop-federation.html),
 and  I started to wonder if is it possible to have 2 YARN runtimes that
 share the same HDFS namespace?

 Thanks,