On Fri, Aug 5, 2016 at 11:31 AM, Maxim Uvarov <maxim.uva...@linaro.org> wrote:
> On 08/05/16 19:29, Mike Holmes wrote: > >> >> >> On 5 August 2016 at 12:27, Maxim Uvarov <maxim.uva...@linaro.org <mailto: >> maxim.uva...@linaro.org>> wrote: >> >> clear_sched_queues(); >> >> is missing just before destroying queues + fix form Bill with >> account all term errors. >> >> >> Should this go on the bug ? >> >> This this thread should be about this patch to enable process mode or it >> will get hijacked >> > > yes, that is test bug, fix has to be in monarch also. > > Bill, I think you will add this to your patch and send v2. > Sorry for the hijack, but I don't understand this analysis or suggestion. Are you saying the cause of the issue Brian identified is a missing clear_sched_queues() call somewhere in odp_scheduling.c? I don't see that as each of the scheduled routines -- test_schedule_single(), test_schedule_multi(), test_schedul_many() -- already contain that call. > > Maxim. > > > >> >> Maxim. >> >> >> >> On 08/05/16 19:01, Bill Fischofer wrote: >> >> On Fri, Aug 5, 2016 at 10:59 AM, Mike Holmes >> <mike.hol...@linaro.org <mailto:mike.hol...@linaro.org>> wrote: >> >> On 4 August 2016 at 11:26, Brian Brooks >> <brian.bro...@linaro.org <mailto:brian.bro...@linaro.org>> >> wrote: >> >> Reviewed-by: Brian Brooks <brian.bro...@linaro.org >> <mailto:brian.bro...@linaro.org>> >> >> >> Just wanted to follow up on this, this patch has also >> highlighted a bug in >> the original code and the thread below documents it, also >> a bug for it has >> been created https://bugs.linaro.org/show_bug.cgi?id=2457 >> <https://bugs.linaro.org/show_bug.cgi?id=2457> >> >> However this patch is orthogonal to that problem which >> existed before this >> patch was created so I think we need to take it so that >> TgrM testing can >> continue. >> >> >> Agreed. This one should be merged to permit parallel activity, >> but we'd >> still like to track down the other bug. >> >> >> On 08/04 09:18:14, Mike Holmes wrote: >> >> +ret=0 >> + >> +run() >> +{ >> + echo odp_scheduling_run_proc starts with $1 >> worker threads >> + echo >> ===================================================== >> + >> + $PERFORMANCE/odp_scheduling${EXEEXT} >> --odph_proc -c $1 || ret=1 >> +} >> + >> +run 1 >> +run 8 >> + >> +exit $ret >> >> Seeing this randomly in both multithread and >> multiprocess modes: >> >> ../../../odp/platform/linux-generic/odp_queue.c:328:odp_ >> >> queue_destroy():queue >> >> "sched_00_07" not empty >> ../../../odp/platform/linux-generic/odp_schedule.c:271: >> >> schedule_term_global():Queue >> >> not empty >> ../../../odp/platform/linux-generic/odp_schedule.c:294: >> >> schedule_term_global():Pool >> >> destroy fail. >> ../../../odp/platform/linux-generic/odp_init.c:188:_odp_ >> >> term_global():ODP >> >> schedule term failed. >> ../../../odp/platform/linux-generic/odp_queue.c:170:odp_ >> >> queue_term_global():Not >> >> destroyed queue: sched_00_07 >> ../../../odp/platform/linux-generic/odp_init.c:195:_odp_ >> >> term_global():ODP >> >> queue term failed. >> ../../../odp/platform/linux-generic/odp_pool.c:149:odp_ >> >> pool_term_global():Not >> >> destroyed pool: odp_sched_pool >> ../../../odp/platform/linux-generic/odp_pool.c:149:odp_ >> >> pool_term_global():Not >> >> destroyed pool: msg_pool >> ../../../odp/platform/linux-generic/odp_init.c:202:_odp_ >> >> term_global():ODP >> >> buffer pool term failed. >> ~/odp_incoming/odp_build/test/common_plat/performance$ >> echo $? >> 0 >> >> Potentially two items: one for correctly returning the >> failure code, and >> another related to teardown. Both beyond the scope of >> this patch which >> LGTM. >> >> >> >> -- >> Mike Holmes >> Technical Manager - Linaro Networking Group >> Linaro.org <http://www.linaro.org/> *│ *Open source >> software for ARM SoCs >> "Work should be fun and collaborative, the rest follows" >> >> >> >> >> >> -- >> Mike Holmes >> Technical Manager - Linaro Networking Group >> Linaro.org <http://www.linaro.org/>***│ *Open source software for ARM >> SoCs >> "Work should be fun and collaborative, the rest follows" >> >> >