so ,there is no need to explicitly do checksum on source and target after we migrate.
I was thinking of using comparechecksum in distcpUtils to compare after migration On Fri, Jul 10, 2015 at 4:20 PM, Harsh J <ha...@cloudera.com> wrote: > Yes, if the length matches and if you haven't specifically asked it to > ignore checksums. > > > On Thursday, July 9, 2015, Giri P <gpatc...@gmail.com> wrote: > >> will distcp do checksum after it copied the data to target ? >> >> On Tue, Feb 3, 2015 at 4:15 AM, Artem Ervits <artemerv...@gmail.com> >> wrote: >> >>> Another good option is hftp. >>> >>> Artem Ervits >>> On Feb 3, 2015 6:42 AM, "xeonmailinglist" <xeonmailingl...@gmail.com> >>> wrote: >>> >>>> I have found the problem. I started to use `webhdfs` and everything is >>>> ok. >>>> >>>> >>>> On 03-02-2015 10:40, xeonmailinglist wrote: >>>> >>>> What do you mean by no path is given? Even if I launch this command, >>>> I get the same error…. What path should I put here? >>>> >>>> $ hadoop distcp hdfs://hadoop-coc-1:50070/input1 >>>> hdfs://hadoop-coc-2:50070/input1 >>>> >>>> Thanks, >>>> >>>> On 02-02-2015 19:59, Alexander Alten-Lorenz wrote: >>>> >>>> Have a closer look: >>>> >>>> hdfs://hadoop-coc-2:50070/ >>>> >>>> >>>> No Path is given. >>>> >>>> >>>> On 02 Feb 2015, at 20:52, xeonmailinglist <xeonmailingl...@gmail.com> >>>> wrote: >>>> >>>> Hi, >>>> >>>> I am trying to copy data using distcp but I get this error. Both >>>> hadoop runtime are working properly. Why is this happening? >>>> >>>> >>>> vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop distcp >>>> hdfs://hadoop-coc-1:50070/input1 hdfs://hadoop-coc-2:50070/ >>>> 15/02/02 19:46:37 ERROR tools.DistCp: Invalid arguments: >>>> java.io.IOException: Failed on local exception: >>>> com.google.protobuf.InvalidProtocolBufferException: Protocol message >>>> end-group tag did not match expected tag.; Host Details : local host is: >>>> "hadoop-coc-1/127.0.1.1"; destination host is: "hadoop-coc-2":50070; >>>> at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772) >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1472) >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1399) >>>> at >>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232) >>>> at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source) >>>> at >>>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752) >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>> at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>> at >>>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187) >>>> at >>>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102) >>>> at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source) >>>> at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988) >>>> at >>>> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118) >>>> at >>>> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114) >>>> at >>>> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) >>>> at >>>> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114) >>>> at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400) >>>> at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:188) >>>> at org.apache.hadoop.tools.DistCp.run(DistCp.java:111) >>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>> at org.apache.hadoop.tools.DistCp.main(DistCp.java:401) >>>> Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol >>>> message end-group tag did not match expected tag. >>>> at >>>> com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94) >>>> at >>>> com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124) >>>> at >>>> com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202) >>>> at >>>> com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241) >>>> at >>>> com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253) >>>> at >>>> com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259) >>>> at >>>> com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49) >>>> at >>>> org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:3167) >>>> at >>>> org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1072) >>>> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:966) >>>> Invalid arguments: Failed on local exception: >>>> com.google.protobuf.InvalidProtocolBufferException: Protocol message >>>> end-group tag did not match expected tag.; Host Details : local host is: >>>> "hadoop-coc-1/127.0.1.1"; destination host is: "hadoop-coc-2":50070; >>>> usage: distcp OPTIONS [source_path...] <target_path> >>>> >>>> Thanks, >>>> >>>> >>>> >>>> >>>> >>>> >>>> >> > > -- > Harsh J >