GridCachePartitionExchangeManager Null pointer exception

2019-01-28 Thread mahesh76private
We got this null pointer exception and the nodes rebooted. 

Attached are the logs. 
GridCachePartitionExchanger.GridCachePartitionExchanger

  





--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: The cluster with persistence enable is stuck for 60 seconds when restarting a node. Is this normal?

2019-01-28 Thread Justin Ji
Sergey - 

Thank for your reply.

Yes, I can find them in the logs, I attached the logs of three nodes, the
node that was restarted is node3.

BTW, I use the default Rebalance Mode(ASYNC), so the rebalancing may not
impact the whole cluster.

Am I getting it wrong?
ignite-003.log
  



--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Is there a way to restart ignite without data but only table structures.

2019-01-28 Thread yangjiajun
Hello.

I want to clear data after restart ignite.But I also want to keep the table
structure which I dynamically created.Is there a way to do this?



--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: Ignite and dynamic linking

2019-01-28 Thread Igor Sapego
Hi,

Currently, Ignite on start creates JVM instance internally, but
it never stops it. Also, it currently can not work with already started
JVM.

So when you start Ignite the first time, it loads JVM, when you stop
and unload it, the JVM remains loaded in process memory. When
you start Ignite again, it discovers that JVM was already loaded, and
as it can not work with pre-loaded JVM, it just returns you the error.

To solve the issue, the following ticket should be implemented [1], but
currently, it is not. As a workaround you may try to call
JNI_DestroyJavaVM() after you have unloaded Ignite, I'm not sure
of the result though. This is simply is not a use case we have tested.

[1] - https://issues.apache.org/jira/browse/IGNITE-4618

Best Regards,
Igor


On Mon, Jan 28, 2019 at 3:49 PM F.D.  wrote:

> Hi Igniters,
> I'm trying to use Ignite in a dll (using c++) that is dinamically loaded.
> I wrapped the method start/end/... bihind a "c" pure interface that I
> export.
>
> It works quite well. I can call the LoadLibrary and start a Ignite node. I
> can stop it and restart it again smoothly.
>
> I've the problem when I LoadLibrary and then I call FreeLibrary (and until
> here it works), but when I try to LoadLibrary again and to start the node,
> I get the error: Failed to initialize JVM* [errCls=, errMsg=JVM already
> created.]*
>
> Do you any ideas why I got this error?
>
> Thanks,
>F.D.
>


Problem setting baseline programmatically

2019-01-28 Thread kellan
I'm trying to set the baseline using this code:

val expectedReplicas = context.igniteReplicas
while (ignite.cluster().forServers().nodes().size() !=
context.igniteReplicas) {
Thread.sleep(5000)
}


ignite.cluster().setBaselineTopology(ignite.cluster().forServers().nodes())

However, when I reach my target number of nodes and attempt to set the
baseline topology not all of the nodes are always added. Is there another
condition that I need to check for before setting?



--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: Blocked system-critical thread has been detected. This can lead to cluster-wide undefined behaviour

2019-01-28 Thread Humphrey
Hi Ilya,

1) Which thread pool is used by compute? (is that the ignite public thread
pool [1])?

I'm now using the following from when I listen to events:

CompletableFuture.runAsync(() -> { 
  ignite.compute().run(new MyRunnable(event.getValue())) 
}, Executors.newFixedThreadPool(10));

This seems to work now but I'm not sure if this is the correct way to handle
the long running events. 
2) I think this will will queue all those jobs until a thread (one of the
10) finishes it's job right?

I've also tried with a compute.runAsync and then listen on the future, after
doing the put in the callback method.
3) Which of these is the best approach?

Humphrey


[1] https://apacheignite.readme.io/docs/thread-pools




--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: failure due to IGNITE_BPLUS_TREE_LOCK_RETRIES

2019-01-28 Thread mahesh76private
We actually think while building index, the tree is getting corrupted. 
Increasing LOCK_RETRIES didnt fix it too...

Please review the below link as well. Infact, we saw creating index was
getting into a infinite loop and throwing out errors at other clients. 


http://apache-ignite-users.70518.x6.nabble.com/processDynamicIndexChange-exception-at-client-node-while-creating-index-td26770.html

My suspicion is this issue happens only when table is partitioned, no
affinity keys, and definitely spread across multiple nodes. 

Please review the logs





--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: The cluster with persistence enable is stuck for 60 seconds when restarting a node. Is this normal?

2019-01-28 Thread Sergey Antonov
Justin, may be problem in rebalancing?
https://apacheignite.readme.io/docs/rebalancing
Could you grep log messages by GridCachePartitionsExchangeManager with text
"Skipping rebalancing" or "Rebalancing scheduled" ?

пн, 28 янв. 2019 г. в 15:04, Justin Ji :

> Can someone give me some suggestions?
>
>
>
> --
> Sent from: http://apache-ignite-users.70518.x6.nabble.com/
>


-- 
BR, Sergey Antonov


Ignite and dynamic linking

2019-01-28 Thread F.D.
Hi Igniters,
I'm trying to use Ignite in a dll (using c++) that is dinamically loaded. I
wrapped the method start/end/... bihind a "c" pure interface that I export.

It works quite well. I can call the LoadLibrary and start a Ignite node. I
can stop it and restart it again smoothly.

I've the problem when I LoadLibrary and then I call FreeLibrary (and until
here it works), but when I try to LoadLibrary again and to start the node,
I get the error: Failed to initialize JVM* [errCls=, errMsg=JVM already
created.]*

Do you any ideas why I got this error?

Thanks,
   F.D.


Re:

2019-01-28 Thread Som Som
.Net part:

using (var ignite =Ignition.StartClient(Ignition
.ClientConfigurationSectionName))

{

var queryEntity = new QueryEntity

{

KeyTypeName = typeof(string).FullName,

KeyType = typeof(string),

ValueTypeName = typeof(int).FullName,

ValueType = typeof(int),

Fields = new[] { newQueryField("ClassCode", typeof(
string)) { IsKeyField =true }, new QueryField("Priority", typeof(int)) {
IsKeyField = false } }

};



queryEntity.TableName =IgniteCacheName
.QUIK_CLASSCODE_PRIORITY;



var cfg = new CacheClientConfiguration(IgniteCacheName
.QUIK_CLASSCODE_PRIORITY, new[] { queryEntity })

{

DataRegionName = "persistent",

Backups = 1,

SqlSchema = "PUBLIC"

};



var c = ignite.GetOrCreateCache(cfg);



c.Put("a", 1);

}



Sql query part:



This query works ok – SELECT _Key, _Val FROM "QUIK.CLASSCODEPRIORITY"

But this one throws an error mentioned above SELECT SecCode, Prioruty FROM
"QUIK.CLASSCODEPRIORITY"

пн, 28 янв. 2019 г., 14:23 Ilya Kasnacheev ilya.kasnach...@gmail.com:

> Hello!
>
> Can you please show your cache configuration and the exact SQL statement
> used?
>
> What happens here is that Ignite expects some composite value type as
> opposed to bare integer. Not so clear why yet.
>
> Regards,
> --
> Ilya Kasnacheev
>
>
> пн, 28 янв. 2019 г. в 14:12, Som Som <2av10...@gmail.com>:
>
>>
>> hi, i'v got a roblem reading the cache throgh the odbc.
>>
>> cahche was created in .net, key type is string, value type is int.
>>
>> error text:
>>
>> SQL Error [5]: javax.cache.CacheException: Failed to execute map
>> query on remote node [nodeId=c3ef8d97-09d0-432d-a0a2-7fd73e8413bc,
>> errMsg=Failed to execute SQL query. General error: "class
>> org.apache.ignite.IgniteCheckedException: Unexpected binary object class
>> [type=class java.lang.Integer]"; SQL statement:
>>
>> SELECT
>>
>> __Z0.CLASSCODE __C0_0,
>>
>> __Z0.PRIORITY __C0_1
>>
>> FROM PUBLIC."QUIK.CLASSCODEPRIORITY" __Z0 [5-197]]
>>
>>   javax.cache.CacheException: Failed to execute map query on remote node
>> [nodeId=c3ef8d97-09d0-432d-a0a2-7fd73e8413bc, errMsg=Failed to execute SQL
>> query. General error: "class org.apache.ignite.IgniteCheckedException:
>> Unexpected binary object class [type=class java.lang.Integer]"; SQL
>> statement:
>>
>> SELECT
>>
>> __Z0.CLASSCODE __C0_0,
>>
>> __Z0.PRIORITY __C0_1
>>
>> FROM PUBLIC."QUIK.CLASSCODEPRIORITY" __Z0 [5-197]]
>>
>>   javax.cache.CacheException: Failed to execute map query on remote node
>> [nodeId=c3ef8d97-09d0-432d-a0a2-7fd73e8413bc, errMsg=Failed to execute SQL
>> query. General error: "class org.apache.ignite.IgniteCheckedException:
>> Unexpected binary object class [type=class java.lang.Integer]"; SQL
>> statement:
>>
>> SELECT
>>
>> __Z0.CLASSCODE __C0_0,
>>
>> __Z0.PRIORITY __C0_1
>>
>> FROM PUBLIC."QUIK.CLASSCODEPRIORITY" __Z0 [5-197]]
>>
>


Re: The cluster with persistence enable is stuck for 60 seconds when restarting a node. Is this normal?

2019-01-28 Thread Justin Ji
Can someone give me some suggestions?



--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: QueryCursor checkpoint

2019-01-28 Thread Ilya Kasnacheev
Hello!

You can indeed scan on per partition basis. Data will not move between
partitions unless key is changed or affinity function is modified. Note
that partition is normally stored on a single node but it might be moved in
case of rebalance.

You can probably read more here:
https://cwiki.apache.org/confluence/display/IGNITE/%28Partition+Map%29+Exchange+-+under+the+hood

Regards,
-- 
Ilya Kasnacheev


чт, 24 янв. 2019 г. в 22:03, msuh :

> Hello,
>
> Our end production cluster would be working with many billions of entities
> in many caches, and have use cases where we need to run ScanQuery over an
> entire cache to update certain fields.
>
> We expect that there could definitely be failures in the middle of a single
> ScanQuery due to the sheer size of the caches. Since we wouldn't want to
> rerun ScanQuery from the start, we're wondering if we could keep some
> checkpoint of up to which point we've processed in the QueryCursor. The
> QueryCursor API doesn't seem to show any methods that allow that, but I may
> not be looking at the right place? Would there be any other efficient ways
> to keep track of vaguely up to which point we've processed? If QueryCursor
> doesn't provide anything externally, would partition number be the best
> option?
>
> But from what I've seen, it seemed like entities in partitions shift around
> (from rebalancing or something?), so not sure if that's even possible.
>
>
>
> --
> Sent from: http://apache-ignite-users.70518.x6.nabble.com/
>


SQL Query | Cache | Partition - strange exception after 3 days in production

2019-01-28 Thread Aat
Hello, 

After 3 days in production : now when we try to execute this query we get an
exception : 

var query = 
new SqlFieldsQuery("select Perimeter, sum(delta) from farVe" + 
" where Perimeter in('A','B')" + 
" and arDate='2019-01-25'" + 
" and UndlName='FTSE' GROUP BY Perimeter"); 

this query worked well until this morning. 

now in the app logs i have : 
 javax.cache.CacheException: Failed to execute map query on remote node
[nodeId=673edfe7-aec7-4d1f-b476-3d4e0ef3ee98, errMsg=Failed to execute SQL
query. General error: "class org.apache.ignite.binary.BinaryObjectException:
Not enough data to read the value [position=0, requiredBytes=1,
remainingBytes=0]"; SQL statement: 
SELECT 
__Z0.PERIMETER AS __C0_0, 
SUM(__Z0.DELTA) AS __C0_1 
FROM "farVe".FARVE __Z0 
WHERE (__Z0.UNDLNAME = 'FTS') AND ((__Z0.PERIMETER IN('A', 'B')) AND
(__Z0.ARDATE = DATE '2019-01-25')) 
GROUP BY __Z0.PERIMETER [5-197]] 

And from the random node i get this message: 

[12:34:45,961][SEVERE][query-#24551][GridMapQueryExecutor] Failed to execute
local query. 
  85601 class org.apache.ignite.cache.query.QueryCancelledException: The
query was cancelled while executing. 
  85602 at
org.apache.ignite.internal.processors.query.h2.IgniteH2Indexing.executeSqlQuery(IgniteH2Indexing.java:1426)
 
  85603 at
org.apache.ignite.internal.processors.query.h2.IgniteH2Indexing.executeSqlQueryWithTimer(IgniteH2Indexing.java:1489)
 
  85604 at
org.apache.ignite.internal.processors.query.h2.twostep.GridMapQueryExecutor.onQueryRequest0(GridMapQueryExecutor.java:930)
 
  85605 at
org.apache.ignite.internal.processors.query.h2.twostep.GridMapQueryExecutor.onQueryRequest(GridMapQueryExecutor.java:705)
 
  85606 at
org.apache.ignite.internal.processors.query.h2.twostep.GridMapQueryExecutor.onMessage(GridMapQueryExecutor.java:240)
 
  85607 at
org.apache.ignite.internal.processors.query.h2.twostep.GridMapQueryExecutor$2.onMessage(GridMapQueryExecutor.java:200)
 
  85608 at
org.apache.ignite.internal.managers.communication.GridIoManager$ArrayListener.onMessage(GridIoManager.java:2349)
 
  85609 at
org.apache.ignite.internal.managers.communication.GridIoManager.invokeListener(GridIoManager.java:1569)
 
  85610 at
org.apache.ignite.internal.managers.communication.GridIoManager.processRegularMessage0(GridIoManager.java:1197)
 
  85611 at
org.apache.ignite.internal.managers.communication.GridIoManager.access$4200(GridIoManager.java:127)
 
  85612 at
org.apache.ignite.internal.managers.communication.GridIoManager$9.run(GridIoManager.java:1093)
 
  85613 at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
  85614 at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
  85615 at java.lang.Thread.run(Thread.java:748) 

Infra: 
--  5 nodes 
-- version : 2.7 


cache definition: 
createCache("farVe", new CacheConfig().init().setIndexedTypes(Long.class, FarVE.class)); 

 // Specify cache mode and/or any other Ignite-specific configuration
properties. 
setCacheMode(CacheMode.PARTITIONED); 

setStoreByValue(false) 
.setWriteThrough(false) 
.setReadThrough(false) 

.setBackups(1) 
.setWriteSynchronizationMode(FULL_SYNC) 

.setStatisticsEnabled(true) 
.setManagementEnabled(true); 


java class: 

@Data 
public class FarVE implements Serializable { 
@QuerySqlField(index = true) 
private LocalDate arDate; 

@QuerySqlField 
private Double delta; 

@QuerySqlField(index = true) 
private String perimeter; 

} 

__ sorry if this error has  already been evoked but i search and i did not
found  answer.

Aat,



--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: Eviction policy is not working for default and new data region

2019-01-28 Thread Ilya Kasnacheev
Hello!

I have looked at your reproducer, and got as far as:
Exception in thread "main" java.io.IOException: Это каталог
at sun.nio.ch.FileDispatcherImpl.read0(Native Method)
at sun.nio.ch.FileDispatcherImpl.read(FileDispatcherImpl.java:46)
at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
at sun.nio.ch.IOUtil.read(IOUtil.java:197)
at sun.nio.ch.FileChannelImpl.read(FileChannelImpl.java:159)
at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:65)
at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:109)
at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:103)
at sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:284)
at sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:326)
at sun.nio.cs.StreamDecoder.read(StreamDecoder.java:178)
at java.io.InputStreamReader.read(InputStreamReader.java:184)
at java.io.BufferedReader.fill(BufferedReader.java:161)
at java.io.BufferedReader.readLine(BufferedReader.java:324)
at java.io.BufferedReader.readLine(BufferedReader.java:389)
at java.nio.file.Files.readAllLines(Files.java:3205)
at java.nio.file.Files.readAllLines(Files.java:3242)
at org.dcache.nfs.ExportFile.parse(ExportFile.java:130)
at org.dcache.nfs.ExportFile.rescan(ExportFile.java:322)
at org.dcache.nfs.ExportFile.(ExportFile.java:98)
at org.dcache.nfs.ExportFile.(ExportFile.java:68)
at org.dcache.vfs4j.NfsMain.main(NfsMain.java:102)

Can you please modify it so it doesn't have extensive dependencies or data
requirements? That I can simply run it and reproduce this behavior?

Regards,
-- 
Ilya Kasnacheev


чт, 24 янв. 2019 г. в 17:34, vyhignite1 :

> Can anyone help?
>
>
>
> --
> Sent from: http://apache-ignite-users.70518.x6.nabble.com/
>


Re:

2019-01-28 Thread Ilya Kasnacheev
Hello!

Can you please show your cache configuration and the exact SQL statement
used?

What happens here is that Ignite expects some composite value type as
opposed to bare integer. Not so clear why yet.

Regards,
-- 
Ilya Kasnacheev


пн, 28 янв. 2019 г. в 14:12, Som Som <2av10...@gmail.com>:

>
> hi, i'v got a roblem reading the cache throgh the odbc.
>
> cahche was created in .net, key type is string, value type is int.
>
> error text:
>
> SQL Error [5]: javax.cache.CacheException: Failed to execute map query
> on remote node [nodeId=c3ef8d97-09d0-432d-a0a2-7fd73e8413bc, errMsg=Failed
> to execute SQL query. General error: "class
> org.apache.ignite.IgniteCheckedException: Unexpected binary object class
> [type=class java.lang.Integer]"; SQL statement:
>
> SELECT
>
> __Z0.CLASSCODE __C0_0,
>
> __Z0.PRIORITY __C0_1
>
> FROM PUBLIC."QUIK.CLASSCODEPRIORITY" __Z0 [5-197]]
>
>   javax.cache.CacheException: Failed to execute map query on remote node
> [nodeId=c3ef8d97-09d0-432d-a0a2-7fd73e8413bc, errMsg=Failed to execute SQL
> query. General error: "class org.apache.ignite.IgniteCheckedException:
> Unexpected binary object class [type=class java.lang.Integer]"; SQL
> statement:
>
> SELECT
>
> __Z0.CLASSCODE __C0_0,
>
> __Z0.PRIORITY __C0_1
>
> FROM PUBLIC."QUIK.CLASSCODEPRIORITY" __Z0 [5-197]]
>
>   javax.cache.CacheException: Failed to execute map query on remote node
> [nodeId=c3ef8d97-09d0-432d-a0a2-7fd73e8413bc, errMsg=Failed to execute SQL
> query. General error: "class org.apache.ignite.IgniteCheckedException:
> Unexpected binary object class [type=class java.lang.Integer]"; SQL
> statement:
>
> SELECT
>
> __Z0.CLASSCODE __C0_0,
>
> __Z0.PRIORITY __C0_1
>
> FROM PUBLIC."QUIK.CLASSCODEPRIORITY" __Z0 [5-197]]
>


Re: Some questions about store and distributed processing

2019-01-28 Thread Ilya Kasnacheev
Hello!

You can also use Continuous Queries for that. They will execute where the
data is modified, when it is modified.
https://apacheignite.readme.io/docs/continuous-queries

Regards,
-- 
Ilya Kasnacheev


сб, 26 янв. 2019 г. в 13:41, yann Blazart :

> Hello all !
>
> I will have to use Ignite because I think it's the best solution to all my
> concerns, but I have a few question.
>
> I have to process very big json files (200GB), with lot of objects of
> different type generated from it.
> These objects, I will have to do multiple controls on it (with a dsl), and
> check  unicity, and in the end do some complexe join request between them.
>
> So for complexe request, ok, I store all in partitionned tables, and do
> request, easy.
>
> But for all the dsl rule ton apply on each object, It could be very nice
> if it can be applied when it's stored on the node, instead of doing it when
> I read the file, I mean :
>
> cache.putAll(mymap);
>
> then something on node to say :  new Entry listener -> execute dsl rules.
>
> I think I can gain lot of processing time like that. But is it possible ?
>
> I checked the doc, but I only see ways to 1st store all then run dsl rule
> on all node.
>
> Thanks in advance ;)
>
> Regards
>


Re: ScanQuery for List<...>

2019-01-28 Thread AndrewV
Thanks a lot.
I tried to create CustomIgniteBiPredicate, but inside the method "apply"
List list is always empty.

*Example with an empty list in filter predicate:*
public boolean apply(Integer key, List list) {
   // Here the list is always empty
for (Foo item: list) {
if (item.id == 2)
return true;
}
return false;
}

*But actually the cache contains data:*
iCache.iterator().forEachRemaining(data -> {
List list = data.getValue(); // Here I have fetched list
}
});





--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: failure due to IGNITE_BPLUS_TREE_LOCK_RETRIES

2019-01-28 Thread Ilya Kasnacheev
Hello!

This is highly unusual, I guess the default value for this property should
be far larger that anything attainable in practice.

Dmitriy, Alexey, can you please clarify?

Regards.
-- 
Ilya Kasnacheev


вс, 27 янв. 2019 г. в 21:16, mahesh76private :

> Increasing the size of IGNITE_BPLUS_TREE_LOCK_RETRIES makes the issue go
> away.
> But the explanation that is occurs due to contention is somewhat not
> clear.
>
> In all cases, in occurred to be when I was creating an index. Feels like it
> is somehow related to the height of the tree... and this variable seems to
> limit the height.
>
>
>
>
> --
> Sent from: http://apache-ignite-users.70518.x6.nabble.com/
>


[no subject]

2019-01-28 Thread Som Som
hi, i'v got a roblem reading the cache throgh the odbc.

cahche was created in .net, key type is string, value type is int.

error text:

SQL Error [5]: javax.cache.CacheException: Failed to execute map query
on remote node [nodeId=c3ef8d97-09d0-432d-a0a2-7fd73e8413bc, errMsg=Failed
to execute SQL query. General error: "class
org.apache.ignite.IgniteCheckedException: Unexpected binary object class
[type=class java.lang.Integer]"; SQL statement:

SELECT

__Z0.CLASSCODE __C0_0,

__Z0.PRIORITY __C0_1

FROM PUBLIC."QUIK.CLASSCODEPRIORITY" __Z0 [5-197]]

  javax.cache.CacheException: Failed to execute map query on remote node
[nodeId=c3ef8d97-09d0-432d-a0a2-7fd73e8413bc, errMsg=Failed to execute SQL
query. General error: "class org.apache.ignite.IgniteCheckedException:
Unexpected binary object class [type=class java.lang.Integer]"; SQL
statement:

SELECT

__Z0.CLASSCODE __C0_0,

__Z0.PRIORITY __C0_1

FROM PUBLIC."QUIK.CLASSCODEPRIORITY" __Z0 [5-197]]

  javax.cache.CacheException: Failed to execute map query on remote node
[nodeId=c3ef8d97-09d0-432d-a0a2-7fd73e8413bc, errMsg=Failed to execute SQL
query. General error: "class org.apache.ignite.IgniteCheckedException:
Unexpected binary object class [type=class java.lang.Integer]"; SQL
statement:

SELECT

__Z0.CLASSCODE __C0_0,

__Z0.PRIORITY __C0_1

FROM PUBLIC."QUIK.CLASSCODEPRIORITY" __Z0 [5-197]]


Re: The cluster with persistence enable is stuck for 60 seconds when restarting a node. Is this normal?

2019-01-28 Thread Justin Ji
Sergey - 

Thank for your reply.

Yes, I defined consistentId for every node. Is it relates to partition map
exchange?



--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: The cluster with persistence enable is stuck for 60 seconds when restarting a node. Is this normal?

2019-01-28 Thread Justin Ji
Sergey - 

Thank for your reply.

Yes, I defined consistentId for every node. Does it relates to partition map
exchange?



--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/


Re: The cluster with persistence enable is stuck for 60 seconds when restarting a node. Is this normal?

2019-01-28 Thread Sergey Antonov
Hi!

Did you define consistentId for nodes?
https://ignite.apache.org/releases/latest/javadoc/org/apache/ignite/configuration/IgniteConfiguration.html#setConsistentId-java.io.Serializable-

пн, 28 янв. 2019 г. в 11:56, Justin Ji :

> I have a cluster(2.7.0) with persistence enabled, the cluster has three
> server nodes(8c16g) and every node has 8 Gb data(1 primary and 1 backup).
>
> When I restart one of them, the whole cluster will stop about 60 seconds
> and
> do not process any request from the client.
>
> From the logs, I saw that the node is processing partition map exchange,
> almost 60 seconds was spent on it.
>
> My disk is SSD which has a high read and write speed.
> The network is well too, the nodes can connect with each other, and the
> latency is very low(less than 0.2 ms) when restarting.
> I also have investigated the GC logs of three nodes, most of GC is finished
> in 50ms.
> My cache AtomicityMode is CacheAtomicityMode.ATOMIC, so there are no long
> transactions.
>
> So I want to ask is it normal that the cluster is stuck for 60 seconds?
>
> If it is not normal, then where should the problem be most, how to
> optimize?
>
> Looking forward to your replies.
>
>
>
>
>
>
> --
> Sent from: http://apache-ignite-users.70518.x6.nabble.com/
>


-- 
BR, Sergey Antonov


The cluster with persistence enable is stuck for 60 seconds when restarting a node. Is this normal?

2019-01-28 Thread Justin Ji
I have a cluster(2.7.0) with persistence enabled, the cluster has three
server nodes(8c16g) and every node has 8 Gb data(1 primary and 1 backup).

When I restart one of them, the whole cluster will stop about 60 seconds and
do not process any request from the client.

>From the logs, I saw that the node is processing partition map exchange,
almost 60 seconds was spent on it.

My disk is SSD which has a high read and write speed.
The network is well too, the nodes can connect with each other, and the
latency is very low(less than 0.2 ms) when restarting.
I also have investigated the GC logs of three nodes, most of GC is finished
in 50ms.
My cache AtomicityMode is CacheAtomicityMode.ATOMIC, so there are no long
transactions.

So I want to ask is it normal that the cluster is stuck for 60 seconds?

If it is not normal, then where should the problem be most, how to optimize?

Looking forward to your replies.






--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/