Hi Anil,

Partition processing time can be high due to non-uniform data distribution
when data collocation or bad affinity function used.
It is ok for collocate data, and it is shouldn't be an  issue if partition
size variation is low enough and you iterate over partitions in parallel
manner.

Please, check that partition sizes.


You approach looks correct.

On Fri, Feb 24, 2017 at 4:13 PM, Anil <anilk...@gmail.com> wrote:

> Hi Andrey,
>
> if you notice in the log, time taken to process the partition is high ( >
> 15 sec). Not sure what is causing that high query time.
>
> In my case, both caches are collocated, and eqId column is indexed and
> setLocal is true for the query.
>
> I wonder if my approach is correct. please correct it, in case you see it
> is suspicious.
>
> Thanks.
>
>
>
> On 24 February 2017 at 18:37, Anil <anilk...@gmail.com> wrote:
>
>> Hi Andrey,
>>
>> I have attached the log. thanks.
>>
>> Thanks.
>>
>>
>>
>>
>>
>> On 24 February 2017 at 18:16, Andrey Mashenkov <
>> andrey.mashen...@gmail.com> wrote:
>>
>>> Hi Anil,
>>>
>>> Would you please provide ignite logs as well?
>>>
>>>
>>> On Fri, Feb 24, 2017 at 3:33 PM, Andrey Gura <ag...@apache.org> wrote:
>>>
>>>> Hi, Anil
>>>>
>>>> Could you please provide crash dump? In your case it is
>>>> /opt/ignite-manager/api/hs_err_pid18543.log file.
>>>>
>>>> On Fri, Feb 24, 2017 at 9:05 AM, Anil <anilk...@gmail.com> wrote:
>>>> > Hi ,
>>>> >
>>>> > I see the node is down with following error while running compute task
>>>> >
>>>> >
>>>> > # A fatal error has been detected by the Java Runtime Environment:
>>>> > #
>>>> > #  SIGSEGV (0xb) at pc=0x00007facd5cae561, pid=18543,
>>>> tid=0x00007fab8a9ea700
>>>> > #
>>>> > # JRE version: OpenJDK Runtime Environment (8.0_111-b14) (build
>>>> > 1.8.0_111-8u111-b14-3~14.04.1-b14)
>>>> > # Java VM: OpenJDK 64-Bit Server VM (25.111-b14 mixed mode linux-amd64
>>>> > compressed oops)
>>>> > # Problematic frame:
>>>> > # J 8676 C2
>>>> > org.apache.ignite.internal.processors.query.h2.opt.GridH2Key
>>>> ValueRowOffheap.getOffheapValue(I)Lorg/h2/value/Value;
>>>> > (290 bytes) @ 0x00007facd5cae561 [0x00007facd5cae180+0x3e1]
>>>> > #
>>>> > # Failed to write core dump. Core dumps have been disabled. To enable
>>>> core
>>>> > dumping, try "ulimit -c unlimited" before starting Java again
>>>> > #
>>>> > # An error report file with more information is saved as:
>>>> > # /opt/ignite-manager/api/hs_err_pid18543.log
>>>> > #
>>>> > # If you would like to submit a bug report, please visit:
>>>> > #   http://bugreport.java.com/bugreport/crash.jsp
>>>> > #
>>>> >
>>>> >
>>>> > I have two 2 caches on 4 node cluster each cache is configured with
>>>> 10 gb
>>>> > off heap.
>>>> >
>>>> > ComputeTask performs the following execution and it is broad casted
>>>> to all
>>>> > nodes.
>>>> >
>>>> >                for (Integer part : parts) {
>>>> > ScanQuery<String, Person> scanQuery = new ScanQuery<String, Person>();
>>>> > scanQuery.setLocal(true);
>>>> > scanQuery.setPartition(part);
>>>> >
>>>> > Iterator<Cache.Entry<String, Person>> iterator =
>>>> > cache.query(scanQuery).iterator();
>>>> >
>>>> > while (iterator.hasNext()) {
>>>> > Cache.Entry<String, Person> row = iterator.next();
>>>> > String eqId =   row.getValue().getEqId();
>>>> > try {
>>>> > QueryCursor<Entry<AffinityKey<String>, Contract>> pdCursor =
>>>> > detailsCache.query(new SqlQuery<AffinityKey<String>,
>>>> > PersonDetail>(PersonDetail.class,
>>>> > "select * from DETAIL_CACHE.PersonDetail where eqId = ? order by
>>>> enddate
>>>> > desc").setLocal(true).setArgs(eqId));
>>>> > Long prev = null;
>>>> > for (Entry<AffinityKey<String>, PersonDetail> d : pdCursor) {
>>>> > // populate person info into person detail
>>>> > dataStreamer.addData(new AffinityKey<String>(detaildId, eqId),
>>>> > d);
>>>> > }
>>>> > pdCursor.close();
>>>> > }catch (Exception ex){
>>>> > }
>>>> > }
>>>> >
>>>> > }
>>>> >
>>>> >
>>>> > Please let me know if you see any issues with approach or any
>>>> > configurations.
>>>> >
>>>> > Thanks.
>>>> >
>>>>
>>>
>>>
>>>
>>> --
>>> Best regards,
>>> Andrey V. Mashenkov
>>>
>>
>>
>


-- 
Best regards,
Andrey V. Mashenkov

Reply via email to