I have deployed some models in RK 3288, which completed in for example 5ms. But
I found that the TVM threads seems consume CPU continuously after the model
inference completed. I tested by the following code:
tvm::runtime::Module mod =
(*tvm::runtime::Registry::Get("tvm.graph_runtime.create"))(json_data,
mod_syslib, device_type, device_id);
tvm::runtime::PackedFunc set_input = mod.GetFunction("set_input");
set_input("data", data);
tvm::runtime::PackedFunc load_params = mod.GetFunction("load_params");
load_params(params);
tvm::runtime::PackedFunc run = mod.GetFunction("run");
while (True)
{
run();
usleep(100000); // in microsecond
}
I used 4 tvm threads. The inference run() will be continued 5ms, then sleeped
100ms. But I used top cmd and saw that the cpu occupation is about 75%, use top
-H saw that there were 3 threads occupied cpu core continuesly (RK3288 has 4
arm core in total).

My question is how to sleep the threads in time without release the
tvm::runtime::Module mod.
---
[Visit
Topic](https://discuss.tvm.ai/t/do-tvm-runtime-threads-occupy-cpu-persistently-how-to-sleep-them-in-time/6178/1)
to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https://discuss.tvm.ai/email/unsubscribe/1279bf184ffe3743a05748ff5c4268fcebb572f598d65cec881ad0ef144bf046).