It started fine and it also recreated the partitions for that cache.

On Fri, Dec 16, 2022 at 10:48 AM John Smith <java.dev....@gmail.com> wrote:

> Weird because after restart and just deleting the
> work/db/node-xxxxxx/cache-my-cache folder on the node that shutdown. It
> started up fine and I have the same amount of caches...
> And sudo lsof -p -a PID returns only 3600 files
>
> On Fri, Dec 16, 2022 at 12:10 AM Gianluca Bonetti <
> gianluca.bone...@gmail.com> wrote:
>
>> Hello
>>
>> I had the same problem, with far more caches in use, as total number (but
>> each cache was very small in size).
>>
>> 32768 files is definitely too low.
>> In my ase, I had to raise it to 262144 hard limit and 131072 soft limit.
>> Please update your /etc/security/limits.conf records for the user you run
>> your app with.
>>
>> I also raised fs.file-max to 2097152 which may be excessive, but I don't
>> see a problem with setting it that high.
>>
>> Cheers
>> Gianluca
>>
>> On Fri, 16 Dec 2022 at 01:39, John Smith <java.dev....@gmail.com> wrote:
>>
>>> Hi it seems the JVM was forcefully shutdown when I tried to create a new
>>> partitioned cache.
>>>
>>> The error seems to indicate that it was "too many files" can someone
>>> from Ignite confirm this?
>>>
>>> I have checked with lsof and Ignite only has about 3600 files open. It's
>>> the only service running on that server. So I don't see how this could
>>> happen? I have a total of 10 caches mixed between replicated and
>>> partitioned (1 backup) over 3 nodes.
>>>
>>> I have
>>>
>>> fs.file-max:300000
>>> and
>>> - soft    nofile          32768
>>> - hard    nofile          32768
>>> respectively on each node.
>>>
>>> What I did was delete the db/folder for that specific cache on that node
>>> and when I restarted it. It worked and recreated the folder for that cache.
>>>
>>> https://www.dropbox.com/s/zwf28akser9p4dt/ignite-XXXXXX.0.log?dl=0
>>>
>>

Reply via email to