gianm commented on issue #6124: KafkaIndexTask can delete published segments on 
restart
URL: 
https://github.com/apache/incubator-druid/issues/6124#issuecomment-411254406
 
 
   > 5. Since the segments which are published twice are already stored in 
metastore, the publish fails.
   
   This doesn't seem right: there is code specifically to handle the case where 
a task tries to publish a segment that some task already published. It happens 
all the time with replicas, and they just ignore that segment and move on to 
the next one.
   
   I wonder if the real reason for publish failure is that the startMetadata 
doesn't match up. I bet it wouldn't match up: it sounds like the task is trying 
to publish from the point it _originally_ started from rather than from the 
point it last published.
   
   It sounds like there are two separate problems here:
   
   - In (7) the task should not have removed the published segments (this is 
the biggest bug).
   - In (3) the task should have done something smarter instead of restoring a 
setup that couldn't possibly work out.

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@druid.apache.org
For additional commands, e-mail: commits-h...@druid.apache.org

Reply via email to