Thanks

> 在 2017年2月10日,12:39,Josh Elser <els...@apache.org> 写道:
> 
> Just to be clear, Lu, for now stick to using a Scanner with the RowIterator :)
> 
> It sounds like we might have to re-think how the RowIterator works with the 
> BatchScanner...
> 
> Christopher wrote:
>> I suspected that was the case. BatchScanner does not guarantee ordering
>> of entries, which is needed for the behavior you're expecting with
>> RowIterator. This means that the RowIterator could see the same row
>> multiple times with different subsets of the row's columns. This is
>> probably affecting your count.
>> 
>> On Thu, Feb 9, 2017 at 10:29 PM Lu Q <luq.j...@gmail.com
>> <mailto:luq.j...@gmail.com>> wrote:
>> 
>>    I use BatchScanner
>> 
>>>    在 2017年2月10日,11:24,Christopher <ctubb...@apache.org
>>>    <mailto:ctubb...@apache.org>> 写道:
>>> 
>>>    Does it matter if your scanner is a BatchScanner or a Scanner?
>>>    I wonder if this is due to the way BatchScanner could split rows up.
>>> 
>>>    On Thu, Feb 9, 2017 at 9:50 PM Lu Q <luq.j...@gmail.com
>>>    <mailto:luq.j...@gmail.com>> wrote:
>>> 
>>> 
>>>        I use accumulo 1.8.0,and I develop a ORM framework for
>>>        conversion the scan result to a object.
>>> 
>>>        Before,I use Rowiterator because it faster than direct to use scan
>>> 
>>>        RowIterator rows = new RowIterator(scan);
>>>        rows.forEachRemaining(rowIterator -> {
>>>        while (rowIterator.hasNext()) {
>>>        Map.Entry<Key, Value> entry = rowIterator.next();
>>>        ...
>>>        }
>>>        }
>>> 
>>>        it works ok until I query 1000+ once .I found that when the
>>>        range size bigger then 1000,some data miss.
>>>        I think maybe I conversion it error ,so I change it to a map
>>>        struct ,the row_id as the map key ,and other as the map value
>>>        ,the problem still exists.
>>> 
>>>        Then I not use RowIterator,it works ok.
>>>        for (Map.Entry<Key, Value> entry : scan) {
>>>        ...
>>>        }
>>> 
>>> 
>>>        Is the bug or my program error ?
>>>        Thanks.
>>> 
>>>    --
>>>    Christopher
>> 
>> --
>> Christopher

Reply via email to