hcms1994 opened a new issue, #13574:
URL: https://github.com/apache/tvm/issues/13574
Hi all,
I try to use the autotuning module to search for the optimal library, but
an error was reported
pool is not optimized for arm cpu.
Extract tasks...
get network success...
Tuning...
[Task 1/52] Current/Best: 3.84/ 47.36 GFLOPS | Progress: (576/576) |
942.71 s Done.
[Task 2/52] Current/Best: 11.06/ 28.70 GFLOPS | Progress: (776/1000) |
1541.06 s Done.
[Task 3/52] Current/Best: 13.94/ 45.63 GFLOPS | Progress: (704/1000) |
1198.90 s Done.
[Task 4/52] Current/Best: 14.51/ 27.14 GFLOPS | Progress: (696/1000) |
1375.30 s Done.
[Task 5/52] Current/Best: 9.40/ 65.11 GFLOPS | Progress: (768/768) |
1282.82 s Done.
[Task 6/52] Current/Best: 27.17/ 35.97 GFLOPS | Progress: (1000/1000) |
2061.47 s Done.
[Task 7/52] Current/Best: 36.12/ 145.86 GFLOPS | Progress: (616/1000) |
1041.53 s Done.
[Task 8/52] Current/Best: 37.84/ 71.57 GFLOPS | Progress: (1000/1000) |
2304.04 s Done.
[Task 9/52] Current/Best: 30.33/ 91.75 GFLOPS | Progress: (1000/1000) |
1706.85 s Done.
[Task 10/52] Current/Best: 15.53/ 54.76 GFLOPS | Progress: (824/1000) |
1632.02 s Done.
[Task 11/52] Current/Best: 7.98/ 66.98 GFLOPS | Progress: (864/960) |
1513.51 s Done.
[Task 12/52] Current/Best: 26.76/ 67.60 GFLOPS | Progress: (1000/1000) |
2103.65 s Done.
[Task 13/52] Current/Best: 62.60/ 153.52 GFLOPS | Progress: (648/1000) |
1203.46 s Done.
[Task 14/52] Current/Best: 14.19/ 98.03 GFLOPS | Progress: (960/1000) |
3060.50 s Done.
[Task 15/52] Current/Best: 69.88/ 143.46 GFLOPS | Progress: (808/1000) |
1494.49 s Done.
[Task 16/52] Current/Best: 22.57/ 89.38 GFLOPS | Progress: (776/1000) |
1591.37 s Done.
[Task 17/52] Current/Best: 35.75/ 133.76 GFLOPS | Progress: (776/1000) |
1370.02 s Done.
[Task 18/52] Current/Best: 55.20/ 81.16 GFLOPS | Progress: (1000/1000) |
2558.72 s Done.
[Task 19/52] Current/Best: 62.47/ 130.66 GFLOPS | Progress: (608/1000) |
1026.17 s Done.
[Task 20/52] Current/Best: 31.73/ 69.75 GFLOPS | Progress: (864/1000) |
2060.20 s Done.
[Task 21/52] Current/Best: 18.59/ 110.98 GFLOPS | Progress: (816/1000) |
1324.28 s Done.
[Task 22/52] Current/Best: 11.29/ 51.83 GFLOPS | Progress: (1000/1000) |
2105.17 s Done.
[Task 23/52] Current/Best: 2.37/ 11.02 GFLOPS | Progress: (800/800) |
1250.26 s Done.
[Task 24/52] Current/Best: 4.44/ 6.52 GFLOPS | Progress: (912/1000) |
1622.29 s Done.
[Task 25/52] Current/Best: 21.41/ 106.14 GFLOPS | Progress: (616/1000) |
1040.57 s Done.
[Task 26/52] Current/Best: 16.15/ 46.35 GFLOPS | Progress: (616/1000) |
1420.74 s Done.
[Task 27/52] Current/Best: 23.23/ 64.34 GFLOPS | Progress: (776/1000) |
1216.78 s Done.
[Task 28/52] Current/Best: 16.34/ 27.72 GFLOPS | Progress: (1000/1000) |
2019.74 s Done.
[Task 29/52] Current/Best: 5.26/ 17.20 GFLOPS | Progress: (800/800) |
1294.08 s Done.
[Task 30/52] Current/Best: 4.21/ 11.27 GFLOPS | Progress: (768/1000) |
1394.74 s Done.
[Task 31/52] Current/Best: 3.37/ 16.32 GFLOPS | Progress: (752/768) |
1200.67 s Done.
[Task 32/52] Current/Best: 4.32/ 12.32 GFLOPS | Progress: (1000/1000) |
1838.50 s Done.
[Task 33/52] Current/Best: 16.47/ 70.21 GFLOPS | Progress: (624/1000) |
1024.04 s Done.
[Task 34/52] Current/Best: 20.03/ 36.60 GFLOPS | Progress: (1000/1000) |
2361.25 s Done.
[Task 35/52] Current/Best: 14.07/ 54.69 GFLOPS | Progress: (608/1000) |
951.49 s Done.
[Task 36/52] Current/Best: 14.23/ 24.36 GFLOPS | Progress: (1000/1000) |
2282.57 s Done.
[Task 37/52] Current/Best: 9.40/ 37.86 GFLOPS | Progress: (768/768) |
1238.55 s Done.
[Task 38/52] Current/Best: 10.34/ 20.86 GFLOPS | Progress: (976/1000) |
1790.84 s Done.
[Task 39/52] Current/Best: 7.46/ 29.37 GFLOPS | Progress: (576/576) |
915.50 s Done.
[Task 40/52] Current/Best: 5.07/ 19.89 GFLOPS | Progress: (1000/1000) |
1882.90 s Done.
[Task 41/52] Current/Best: 8.10/ 41.97 GFLOPS | Progress: (616/720) |
991.51 s Done.
[Task 42/52] Current/Best: 7.30/ 22.45 GFLOPS | Progress: (1000/1000) |
2165.44 s Done.
[Task 43/52] Current/Best: 18.56/ 45.00 GFLOPS | Progress: (576/576) |
941.53 s Done.
[Task 44/52] Current/Best: 9.23/ 28.91 GFLOPS | Progress: (888/1000) |
1649.56 s Done.
[Task 45/52] Current/Best: 16.35/ 76.66 GFLOPS | Progress: (648/1000) |
1101.68 s Done.
[Task 46/52] Current/Best: 19.67/ 42.18 GFLOPS | Progress: (944/1000) |
2283.94 s Done.
[Task 47/52] Current/Best: 9.82/ 58.84 GFLOPS | Progress: (632/960) |
1039.73 s Done.
[Task 48/52] Current/Best: 14.56/ 21.36 GFLOPS | Progress: (1000/1000) |
2293.03 s Done.
[Task 49/52] Current/Best: 5.48/ 72.21 GFLOPS | Progress: (1000/1000) |
1728.76 s Done.
[Task 50/52] Current/Best: 25.33/ 45.19 GFLOPS | Progress: (944/1000) |
1801.64
s/usr/local/lib/python3.6/dist-packages/setuptools-58.5.3-py3.6.egg/pkg_resources/__init__.py:119:
PkgResourcesDeprecationWarning: 0.18ubuntu0.18.04.1 is an invalid version and
will not be supported in a future release
PkgResourcesDeprecationWarning,
/usr/local/lib/python3.6/dist-packages/xgboost/training.py:17: UserWarning:
Old style callback is deprecated. See:
https://xgboost.readthedocs.io/en/latest/python/callbacks.html
warnings.warn(f'Old style callback is deprecated. See: {link}',
UserWarning)
Done.
Traceback (most recent call last):
File "autoTVM_tune_relay_cuda_agx_1000.py", line 282, in <module>
tune_and_evaluate(tuning_option)
File "autoTVM_tune_relay_cuda_agx_1000.py", line 250, in tune_and_evaluate
tune_tasks(tasks, **tuning_opt)
File "autoTVM_tune_relay_cuda_agx_1000.py", line 216, in tune_tasks
tuner_obj.load_history(autotvm.record.load_from_file(tmp_log_file))
File
"/home/caros/vis_work/code/apache-tvm-src-v0.10.0/python/tvm/autotvm/tuner/model_based_tuner.py",
line 314, in load_history
maximums = self.model_optimizer.find_maximums(base_model,
self.plan_size, self.visited)
File
"/home/caros/vis_work/code/apache-tvm-src-v0.10.0/python/tvm/autotvm/tuner/sa_model_optimizer.py",
line 89, in find_maximums
scores = model.predict(points)
File
"/home/caros/vis_work/code/apache-tvm-src-v0.10.0/python/tvm/autotvm/tuner/xgboost_cost_model.py",
line 311, in predict
return self.bst.predict(dtest, output_margin=output_margin)
File "/usr/local/lib/python3.6/dist-packages/xgboost/core.py", line 1920,
in predict
ctypes.byref(preds)
File "/usr/local/lib/python3.6/dist-packages/xgboost/core.py", line 218,
in _check_call
raise XGBoostError(py_str(_LIB.XGBGetLastError()))
xgboost.core.XGBoostError: [02:04:57] /workspace/src/learner.cc:1257: Check
failed: learner_model_param_.num_feature >= p_fmat->Info().num_col_ (835 vs.
919) : Number of columns does not match number of features in booster.
Stack trace:
[bt] (0)
/usr/local/lib/python3.6/dist-packages/xgboost/lib/libxgboost.so(+0x1658c0)
[0x7fa07368c0]
[bt] (1)
/usr/local/lib/python3.6/dist-packages/xgboost/lib/libxgboost.so(+0x165d90)
[0x7fa0736d90]
[bt] (2)
/usr/local/lib/python3.6/dist-packages/xgboost/lib/libxgboost.so(+0x16f6d0)
[0x7fa07406d0]
[bt] (3)
/usr/local/lib/python3.6/dist-packages/xgboost/lib/libxgboost.so(+0x16f840)
[0x7fa0740840]
[bt] (4)
/usr/local/lib/python3.6/dist-packages/xgboost/lib/libxgboost.so(XGBoosterPredictFromDMatrix+0x2dc)
[0x7fa062a18c]
[bt] (5) /usr/lib/aarch64-linux-gnu/libffi.so.6(ffi_call_SYSV+0x64)
[0x7facfccd28]
[bt] (6) /usr/lib/aarch64-linux-gnu/libffi.so.6(ffi_call+0xc8)
[0x7facfcd698]
[bt] (7)
/usr/lib/python3.6/lib-dynload/_ctypes.cpython-36m-aarch64-linux-gnu.so(_ctypes_callproc+0x420)
[0x7f99950198]
[bt] (8)
/usr/lib/python3.6/lib-dynload/_ctypes.cpython-36m-aarch64-linux-gnu.so(+0x104a8)
[0x7f999504a8]
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]