Possibly, though in this case (12 million slot dirs)
that would imply that the client has 12 million tasks,
which is unlikely because then the client would completely bog down
because of the N^2 scheduling logic.
It's more likely due to a failure to clean out slot dirs,
and I suspect it's specific to Volpex.

I'll make the ncpus*100 change.

-- David
On 24-Oct-2013 11:26 AM, McLeod, John wrote:
An unbounded number of slot dirs. Would be caused by High Priority picking up a 
task, running it for a few minutes, discovering that that particular task was 
no longer in high priority, and picking up one that was now in high priority.  
Repeat until all tasks have run for a couple of minutes in high priority.  
(This should have been fixed by the recent change to EDF for the project rather 
than selecting the tasks that were tagged as high priority).

The other way would be to have slots that cannot be cleaned out (but I thought 
that bug was fixed a while ago).

On a different note, could we change the calculated maximum number of slots to 
cope with machines with more than 1000 CPUs?  I know that machines like that 
are rare now.  How about n_cpus * 100 for the number of slot directories 
allowed?
_______________________________________________
boinc_dev mailing list
boinc_dev@ssl.berkeley.edu
http://lists.ssl.berkeley.edu/mailman/listinfo/boinc_dev
To unsubscribe, visit the above URL and
(near bottom of page) enter your email address.

_______________________________________________
boinc_dev mailing list
boinc_dev@ssl.berkeley.edu
http://lists.ssl.berkeley.edu/mailman/listinfo/boinc_dev
To unsubscribe, visit the above URL and
(near bottom of page) enter your email address.

Reply via email to