jwfromm commented on a change in pull request #7063: URL: https://github.com/apache/tvm/pull/7063#discussion_r539661492
########## File path: python/tvm/contrib/nvcc.py ########## @@ -269,15 +270,24 @@ def have_int8(compute_version): return False -def have_tensorcore(compute_version): +def have_tensorcore(compute_version=None): """Either TensorCore support is provided in the compute capability or not Parameters ---------- compute_version : str compute capability of a GPU (e.g. "7.0") """ + if compute_version is None: + if tvm.gpu(0).exist: + compute_version = tvm.gpu(0).compute_version + else: + compute_version = AutotvmGlobalScope.current.cuda_target_arch Review comment: I've changed `have_tensorcore` to instead extract the architecture from a target. Can you take another look and let me know what you think? One downside is that `have_tensorcore` is used in places that don't have access to the target object such as in topi functions, so we have to support both the old way of getting target from `tvm.gpu...` and the target object. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org