From 7063a09ef1b98849e98194e8a9e47455cd1b5fa3 Mon Sep 17 00:00:00 2001 From: Masahiro Masuda Date: Fri, 21 May 2021 17:57:03 +0900 Subject: [PATCH] minor fix --- python/tvm/relay/op/_transform.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/python/tvm/relay/op/_transform.py b/python/tvm/relay/op/_transform.py index c158075f53a8..76d417f0bdc7 100644 --- a/python/tvm/relay/op/_transform.py +++ b/python/tvm/relay/op/_transform.py @@ -1136,7 +1136,7 @@ def _gather_nd_shape(data_shape, indices_shape, batch_dims, num_indices_per_tupl # depend on a runtime shape dimension of indices tensor, even if the # dimension is always a known, fixed value. As a workaround, we assume that # the fixed gather dimension (the size of an indexing tuple) is recorded - # in `gather_nd` op attribute. + # in gather_nd op attributes. mdim = num_indices_per_tuple kdim = indices_shape.shape[0] - 1 out_shape = output_tensor((kdim + ndim - (mdim + batch_dims),), "int64") @@ -1150,7 +1150,7 @@ def _gather_nd_shape(data_shape, indices_shape, batch_dims, num_indices_per_tupl @_reg.register_shape_func("gather_nd", False) def gather_nd_shape_func(attrs, inputs, _): """ - Shape func for ghater_nd operator. + Shape func for gather_nd operator. """ batch_dims = get_const_int(attrs.batch_dims) num_indices_per_tuple = get_const_int(attrs.num_indices_per_tuple)