[ 
https://issues.apache.org/jira/browse/HIVE-26933?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-26933:
----------------------------------
    Labels: pull-request-available  (was: )

> Cleanup dump directory for eventId which was failed in previous dump cycle
> --------------------------------------------------------------------------
>
>                 Key: HIVE-26933
>                 URL: https://issues.apache.org/jira/browse/HIVE-26933
>             Project: Hive
>          Issue Type: Improvement
>            Reporter: Harshal Patel
>            Assignee: Harshal Patel
>            Priority: Major
>              Labels: pull-request-available
>          Time Spent: 10m
>  Remaining Estimate: 0h
>
> # If Incremental Dump operation failes while dumping any event id  in the 
> staging directory. Then dump directory for this event id along with file 
> _dumpmetadata  still exists in the dump location. which is getting stored in 
> _events_dump file
>  # When user triggers dump operation for this policy again, It again resumes 
> dumping from failed event id, and tries to dump it again but as that event id 
> directory already created in previous cycle, it fails with the exception
> {noformat}
> [Scheduled Query Executor(schedule:repl_policytest7, execution_id:7181)]: 
> FAILED: Execution Error, return code 40000 from 
> org.apache.hadoop.hive.ql.exec.repl.ReplDumpTask. 
> org.apache.hadoop.fs.FileAlreadyExistsException: 
> /warehouse/tablespace/staging/policytest7/dGVzdDc=/14bcf976-662b-4237-b5bb-e7d63a1d089f/hive/137961/_dumpmetadata
>  for client 172.27.182.5 already exists
>     at 
> org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.startFile(FSDirWriteFileOp.java:388)
>     at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2576)
>     at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2473)
>     at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:773)
>     at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:490)
>     at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>     at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:533)
>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070)
>     at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:989)
>     at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:917)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:422)
>     at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1898)
>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2894){noformat}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to