[ 
https://issues.apache.org/jira/browse/ASTERIXDB-2276?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16355684#comment-16355684
 ] 

ASF subversion and git services commented on ASTERIXDB-2276:
------------------------------------------------------------

Commit 734e28c021e41e4ca04017ca3473eb0f6518ec55 in asterixdb-bad's branch 
refs/heads/master from [~mhubail]
[ https://git-wip-us.apache.org/repos/asf?p=asterixdb-bad.git;h=734e28c ]

[ASTERIXDB-2276][BAD] Coordinated Change

Change-Id: I5b82de3902550da24f0be54414cb167bec71834f


> Introduce max active writable datasets
> --------------------------------------
>
>                 Key: ASTERIXDB-2276
>                 URL: https://issues.apache.org/jira/browse/ASTERIXDB-2276
>             Project: Apache AsterixDB
>          Issue Type: Improvement
>            Reporter: Murtadha Hubail
>            Assignee: Murtadha Hubail
>            Priority: Major
>
> currently, it is possible for an operation that tries to allocates memory for 
> its dataset to fail without the user knowing about it. For example, if a feed 
> is connected to a number of datasets that exceeds the max global memory for 
> datasets, this issue will happen. The proposal is to add a configurable max 
> active writable datasets parameter and calculate the memory component size 
> based on it and the the memory component page size. After that, any operation 
> could check the currently modified datasets and fail early if that number 
> exceeds the configured value.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to