This reminded me of https://issues.apache.org/jira/browse/ACCUMULO-3710

I don't feel like 3000 ranges is too many, but this isn't quantitative.

IIRC, the BatchScanner will take each Range you provide, bin each Range to the TabletServer(s) currently hosting the corresponding data, clip (truncate) each Range to match the Tablet boundaries, and then does an RPC to each TabletServer with just the Ranges hosted there.

Inside the TabletServer, it will then have many Ranges, binned by Tablet (KeyExtent, to be precise). This will spawn a org.apache.accumulo.tserver.scan.LookupTask will will start collecting results to send back to the client.

The caveat here is that those ranges are processed serially on a TabletServer. Maybe, you're swamping one TabletServer with lots of Ranges that it could be processing in parallel.

Could you experiment with using multiple BatchScanners and something like Guava's Iterables.concat to make it appear like one Iterator?

I'm curious if we should put an optimization into the BatchScanner itself to limit the number of ranges we send in one RPC to a TabletServer (e.g. one BatchScanner might open multiple MultiScanSessions to a TabletServer).

Sven Hodapp wrote:
Hi there,

currently we're experimenting with a two node Accumulo cluster (two tablet 
servers) setup for document storage.
This documents are decomposed up to the sentence level.

Now I'm using a BatchScanner to assemble the full document like this:

     val bscan = instance.createBatchScanner(ARTIFACTS, auths, 10) // ARTIFACTS 
table currently hosts ~30GB data, ~200M entries on ~45 tablets
     bscan.setRanges(ranges)  // there are like 3000 Range.exact's in the 
ranges-list
       for (entry<- bscan.asScala) yield {
         val key = entry.getKey()
         val value = entry.getValue()
         // etc.
       }

For larger full documents (e.g. 3000 exact ranges), this operation will take 
about 12 seconds.
But shorter documents are assembled blazing fast...

Is that to much for a BatchScanner / I'm misusing the BatchScaner?
Is that a normal time for such a (seek) operation?
Can I do something to get a better seek performance?

Note: I have already enabled bloom filtering on that table.

Thank you for any advice!

Regards,
Sven

Reply via email to