GWphua commented on PR #19016:
URL: https://github.com/apache/druid/pull/19016#issuecomment-3956270711

   Thanks for taking a look @cecemei 
   
   This method depends on the SEGMENT lock introduced in #7547 (See #7491 also 
for more info). The use of segment lock assumes that the number of segments in 
the same datasource and same interval will not exceed `Short.MAX_VALUE` (which 
im not a fan of).  
   
   Let's say there are 100 segments in the same interval. we specify segments 
0~4 in the SpecificSegmentsSpec to be compacted:
   
   A new segment `segmentId_32768` will be created from segments 0~4, having 
shardSpec: 
   ```
   {
     "type": "numbered_overwrite",
     "partitionId": 32768,
     "startRootPartitionId": 0,
     "endRootPartitionId": 5,
     "minorVersion": 1,
     "atomicUpdateGroupSize": 1
   }
   ```
   
   The rest of the segments (5~100) will still stay queryable + available. The 
segments 0~4 will then be scheduled to be killed. 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to