Kilian, Was this slowdown happening always or did it just started to be bad recently?
Thanks! Regards Hartmut --------------- http://boost-spirit.com http://stellar.cct.lsu.edu > -----Original Message----- > From: hpx-users-boun...@stellar.cct.lsu.edu [mailto:hpx-users- > boun...@stellar.cct.lsu.edu] On Behalf Of Kilian Werner > Sent: Tuesday, January 9, 2018 7:46 AM > To: hpx-users@stellar.cct.lsu.edu > Subject: [hpx-users] Performance Counter Data Interpretation > > Dear hpx user list, > > one of our projects shows unexpectedly bad speedups when > supplying additional OS-worker-threads to HPX. > The project is run locally and in parallel on a machine > with 8 cores, trying to pin down the parallelization > bottleneck we printed the built in HPX Performance > Counters as seen below. > The parallelization is achieved by scheduling tasks with > hpx::apply that themselves will schedule additional tasks > with hpx::apply. > The program terminates after a final task (that can > identify itself and will always finish last, independent > of task scheduling order) fires an event. > Synchronization is performed with some > hpx::lcos::local::mutex locks. > > The problem seems to be apparent when looking at the > harshly growing cumulative-overhead per worker-thread when > employing more OS threads. > However we are a bit clueless as to interpret the meaning > of this cumulative-overhead counter. > We were especially surprised to find, that the > per-worker-thread overhead at some point came close to and > even surpassed the total cumulative runtime (see > cumulative overhead of worker thread 0 when run with 8 os > threads vs. total cumulative runtime). > > What exactly does the performance counter > /threads/time/cumulative-overhead measure? How can the > overhead be larger than the total execution time? > How could we narrow down the causes for the growing > overhead? For example how could we measure how much time > is spend waiting at (specific) mutexes in total? > > Thanks in advance, > > Kilian Werner > > > > --hpx:threads 1: > > /threads{locality#0/total/total}/count/cumulative,1,2.015067,[s],127704 > /threads{locality#0/total/total}/time/average,1,2.015073,[s],14938,[ns] > /threads{locality#0/total/total}/time/cumulative,1,2.015074,[s],1.90769e+0 > 9,[ns] > /threads{locality#0/total/total}/time/cumulative- > overhead,1,2.015076,[s],1.03483e+08,[ns] > /threads{locality#0/pool#default/worker-thread#0}/time/cumulative- > overhead,1,2.015076,[s],1.03483e+08,[ns] > /threads{locality#0/total/total}/idle-rate,1,2.015078,[s],514,[0.01%] > > --hpx:threads 2: > > /threads{locality#0/total/total}/count/cumulative,1,1.814639,[s],112250 > /threads{locality#0/total/total}/time/average,1,1.814644,[s],17986,[ns] > /threads{locality#0/total/total}/time/cumulative,1,1.814654,[s],2.01907e+0 > 9,[ns] > /threads{locality#0/total/total}/time/cumulative- > overhead,1,1.814647,[s],1.60469e+09,[ns] > /threads{locality#0/pool#default/worker-thread#0}/time/cumulative- > overhead,1,1.814599,[s],1.12562e+09,[ns] > /threads{locality#0/pool#default/worker-thread#1}/time/cumulative- > overhead,1,1.814649,[s],4.79071e+08,[ns] > /threads{locality#0/total/total}/idle-rate,1,1.814603,[s],4428,[0.01%] > > --hpx:threads 8: > > /threads{locality#0/total/total}/count/cumulative,1,4.597361,[s],109476 > /threads{locality#0/total/total}/time/average,1,4.597373,[s],37988,[ns] > /threads{locality#0/total/total}/time/cumulative,1,4.597335,[s],4.1588e+09 > ,[ns] > /threads{locality#0/total/total}/time/cumulative- > overhead,1,4.597325,[s],3.25232e+10,[ns] > /threads{locality#0/pool#default/worker-thread#0}/time/cumulative- > overhead,1,4.597408,[s],4.20735e+09,[ns] > /threads{locality#0/pool#default/worker-thread#1}/time/cumulative- > overhead,1,4.597390,[s],4.08787e+09,[ns] > /threads{locality#0/pool#default/worker-thread#2}/time/cumulative- > overhead,1,4.597385,[s],3.62298e+09,[ns] > /threads{locality#0/pool#default/worker-thread#3}/time/cumulative- > overhead,1,4.597358,[s],4.12475e+09,[ns] > /threads{locality#0/pool#default/worker-thread#4}/time/cumulative- > overhead,1,4.597338,[s],4.10011e+09,[ns] > /threads{locality#0/pool#default/worker-thread#5}/time/cumulative- > overhead,1,4.597402,[s],4.14242e+09,[ns] > /threads{locality#0/pool#default/worker-thread#6}/time/cumulative- > overhead,1,4.597353,[s],4.13593e+09,[ns] > /threads{locality#0/pool#default/worker-thread#7}/time/cumulative- > overhead,1,4.597408,[s],4.13275e+09,[ns] > /threads{locality#0/total/total}/idle-rate,1,4.597350,[s],8867,[0.01%] > > > _______________________________________________ > hpx-users mailing list > hpx-users@stellar.cct.lsu.edu > https://mail.cct.lsu.edu/mailman/listinfo/hpx-users _______________________________________________ hpx-users mailing list hpx-users@stellar.cct.lsu.edu https://mail.cct.lsu.edu/mailman/listinfo/hpx-users