[ https://issues.apache.org/jira/browse/HIVE-26933?focusedWorklogId=842162&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-842162 ]
ASF GitHub Bot logged work on HIVE-26933: ----------------------------------------- Author: ASF GitHub Bot Created on: 30/Jan/23 01:46 Start Date: 30/Jan/23 01:46 Worklog Time Spent: 10m Work Description: pudidic merged PR #3984: URL: https://github.com/apache/hive/pull/3984 Issue Time Tracking ------------------- Worklog Id: (was: 842162) Time Spent: 3h 20m (was: 3h 10m) > Cleanup dump directory for eventId which was failed in previous dump cycle > -------------------------------------------------------------------------- > > Key: HIVE-26933 > URL: https://issues.apache.org/jira/browse/HIVE-26933 > Project: Hive > Issue Type: Improvement > Reporter: Harshal Patel > Assignee: Harshal Patel > Priority: Major > Labels: pull-request-available > Time Spent: 3h 20m > Remaining Estimate: 0h > > # If Incremental Dump operation failes while dumping any event id in the > staging directory. Then dump directory for this event id along with file > _dumpmetadata still exists in the dump location. which is getting stored in > _events_dump file > # When user triggers dump operation for this policy again, It again resumes > dumping from failed event id, and tries to dump it again but as that event id > directory already created in previous cycle, it fails with the exception > {noformat} > [Scheduled Query Executor(schedule:repl_policytest7, execution_id:7181)]: > FAILED: Execution Error, return code 40000 from > org.apache.hadoop.hive.ql.exec.repl.ReplDumpTask. > org.apache.hadoop.fs.FileAlreadyExistsException: > /warehouse/tablespace/staging/policytest7/dGVzdDc=/14bcf976-662b-4237-b5bb-e7d63a1d089f/hive/137961/_dumpmetadata > for client 172.27.182.5 already exists > at > org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.startFile(FSDirWriteFileOp.java:388) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2576) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2473) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:773) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:490) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:533) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:989) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:917) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1898) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2894){noformat} > -- This message was sent by Atlassian Jira (v8.20.10#820010)