-
Notifications
You must be signed in to change notification settings - Fork 3.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Relay][Dyn] Dynamic TopK Op #6008
Merged
Merged
Changes from all commits
Commits
Show all changes
5 commits
Select commit
Hold shift + click to select a range
File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,71 @@ | ||
# Licensed to the Apache Software Foundation (ASF) under one | ||
# or more contributor license agreements. See the NOTICE file | ||
# distributed with this work for additional information | ||
# regarding copyright ownership. The ASF licenses this file | ||
# to you under the Apache License, Version 2.0 (the | ||
# "License"); you may not use this file except in compliance | ||
# with the License. You may obtain a copy of the License at | ||
# | ||
# http://www.apache.org/licenses/LICENSE-2.0 | ||
# | ||
# Unless required by applicable law or agreed to in writing, | ||
# software distributed under the License is distributed on an | ||
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY | ||
# KIND, either express or implied. See the License for the | ||
# specific language governing permissions and limitations | ||
# under the License. | ||
"Definition of classic algorithms" | ||
# pylint: disable=invalid-name,unused-argument | ||
from __future__ import absolute_import | ||
|
||
from tvm.te.hybrid import script | ||
from tvm.runtime import convert | ||
|
||
from .. import strategy | ||
from .. import op as _reg | ||
from ..op import OpPattern, register_pattern | ||
from ..op import register_strategy | ||
|
||
# topk | ||
register_strategy("dyn.topk", strategy.topk_strategy) | ||
register_pattern("dyn.topk", OpPattern.OPAQUE) | ||
|
||
@script | ||
def _topk_shape_func_input_data(data, k, axis): | ||
ndim = len(data.shape) | ||
val_out = output_tensor((ndim,), "int64") | ||
indices_out = output_tensor((ndim,), "int64") | ||
|
||
for i in const_range(ndim): | ||
if i != axis: | ||
val_out[i] = int64(data.shape[i]) | ||
indices_out[i] = int64(data.shape[i]) | ||
else: | ||
if k[0] < 1: | ||
val_out[i] = int64(data.shape[i]) | ||
indices_out[i] = int64(data.shape[i]) | ||
else: | ||
val_out[i] = int64(k[0]) | ||
indices_out[i] = int64(k[0]) | ||
return val_out, indices_out | ||
|
||
@_reg.register_shape_func("dyn.topk", True) | ||
def topk_shape_func(attrs, inputs, _): | ||
""" | ||
Shape func for topk. | ||
""" | ||
axis = attrs.axis | ||
if axis < 0: | ||
axis += len(inputs[0].shape) | ||
val_out, indices_out = \ | ||
_topk_shape_func_input_data(inputs[0], inputs[1], convert(axis)) | ||
|
||
ret_type = attrs.ret_type | ||
if ret_type == "both": | ||
ret = [val_out, indices_out] | ||
elif ret_type == "values": | ||
ret = [val_out] | ||
else: | ||
ret = [indices_out] | ||
|
||
return ret |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,107 @@ | ||
/* | ||
* Licensed to the Apache Software Foundation (ASF) under one | ||
* or more contributor license agreements. See the NOTICE file | ||
* distributed with this work for additional information | ||
* regarding copyright ownership. The ASF licenses this file | ||
* to you under the Apache License, Version 2.0 (the | ||
* "License"); you may not use this file except in compliance | ||
* with the License. You may obtain a copy of the License at | ||
* | ||
* http://www.apache.org/licenses/LICENSE-2.0 | ||
* | ||
* Unless required by applicable law or agreed to in writing, | ||
* software distributed under the License is distributed on an | ||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY | ||
* KIND, either express or implied. See the License for the | ||
* specific language governing permissions and limitations | ||
* under the License. | ||
*/ | ||
|
||
/*! | ||
* \file topk.cc | ||
* \brief TopK operators | ||
*/ | ||
#include <tvm/relay/attrs/algorithm.h> | ||
#include <tvm/relay/op.h> | ||
#include <tvm/tir/op.h> | ||
|
||
namespace tvm { | ||
namespace relay { | ||
namespace dyn { | ||
|
||
bool TopKRel(const Array<Type>& types, int num_inputs, const Attrs& attrs, | ||
const TypeReporter& reporter) { | ||
// `types` contains: [data, k, result] | ||
const TopKAttrs* param = attrs.as<TopKAttrs>(); | ||
CHECK_EQ(types.size(), 3); | ||
const auto* data = types[0].as<TensorTypeNode>(); | ||
const auto* k = types[1].as<TensorTypeNode>(); | ||
if (data == nullptr) { | ||
CHECK(types[0].as<IncompleteTypeNode>()) | ||
<< "tile: expect input type to be TensorType but get " << types[0]; | ||
return false; | ||
} | ||
if (k == nullptr) { | ||
CHECK(types[1].as<IncompleteTypeNode>()) | ||
<< "tile: expect input type to be TensorType but get " << types[1]; | ||
return false; | ||
} | ||
CHECK(k->shape.size() <= 1) << "Parameter k must be a Scalar or a Tensor of shape (1, )"; | ||
if (k->shape.size() == 1) { | ||
const IntImmNode* k_shape = k->shape[0].as<IntImmNode>(); | ||
CHECK(k_shape) << "Parameter k must have static shape"; | ||
CHECK_EQ(k_shape->value, 1) << "Parameter k must be a Scalar or a Tensor of shape (1, )"; | ||
} | ||
int ndim = data->shape.size(); | ||
int axis = param->axis; | ||
if (axis < 0) { | ||
axis += ndim; | ||
} | ||
CHECK(axis >= 0 && axis < ndim); | ||
Array<IndexExpr> out_shape; | ||
for (int i = 0; i < ndim; ++i) { | ||
if (i != axis) { | ||
out_shape.push_back(data->shape[i]); | ||
} else { | ||
out_shape.push_back(Any()); | ||
} | ||
} | ||
auto values_ty = TensorType(out_shape, data->dtype); | ||
auto indices_ty = TensorType(out_shape, param->dtype); | ||
if (param->ret_type == "both") { | ||
reporter->Assign(types[2], TupleType({values_ty, indices_ty})); | ||
} else if (param->ret_type == "values") { | ||
reporter->Assign(types[2], values_ty); | ||
} else if (param->ret_type == "indices") { | ||
reporter->Assign(types[2], indices_ty); | ||
} else { | ||
LOG(FATAL) << "Unsupported ret type: " << param->ret_type; | ||
} | ||
return true; | ||
} | ||
|
||
Expr MakeTopK(Expr data, Expr k, int axis, String ret_type, bool is_ascend, DataType dtype) { | ||
auto attrs = make_object<TopKAttrs>(); | ||
attrs->axis = axis; | ||
attrs->ret_type = ret_type; | ||
attrs->is_ascend = is_ascend; | ||
attrs->dtype = dtype; | ||
static const Op& op = Op::Get("dyn.topk"); | ||
return Call(op, {data, k}, Attrs(attrs), {}); | ||
} | ||
|
||
TVM_REGISTER_GLOBAL("relay.op.dyn._make.topk").set_body_typed(MakeTopK); | ||
|
||
RELAY_REGISTER_OP("dyn.topk") | ||
.describe(R"doc(Get the top k elements in an input tensor along the given axis. | ||
)doc" TVM_ADD_FILELINE) | ||
.set_num_inputs(2) | ||
.set_attrs_type<TopKAttrs>() | ||
.add_argument("data", "Tensor", "Input data.") | ||
.add_argument("k", "Tensor", "Number of top elements.") | ||
.set_support_level(6) | ||
.add_type_rel("DynTopK", TopKRel); | ||
|
||
} // namespace dyn | ||
} // namespace relay | ||
} // namespace tvm |
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm concered about this case: If data's shape is not static, it needs to use _dyn_make.topk() too.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I might be missing something. If it's a Relay Constant, we go static. If it's another Relay Expr, we go dynamic, if it's a python literal, we go static. Is there another possibility?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What I meant is
data
, notk
.data
can only be Relay Expr. Should we always use _dyn.topk and let DynmaicToStatic determine whether can convert it to static op?Because TopK's out_shape depends on shape of data, if shape of data is dynamic, out_shape becomes dynamic, we need to use shape function in that case
https://github.com/apache/incubator-tvm/blob/eafb2aa13d6cd223629f17d5f6aab5a8d4fce7f5/src/relay/op/algorithm/topk.cc#L50
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Current dyn namespace is for attrs to be Relay Expr which makes shape func data dependent. For data independent shape func we can still relay on "static" version op.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I understand that. But it's possbile that an op is input shape depandeant but input shape itself is dynamic. In this kind case, we still need to use shape function.
This is a simliar case which doesn't have topk, but has similar issue.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, I think this is a bit of complexity around how @icemelon9 and co implemented dynamic shapes in Relay. Basically, any op can take in
Any
shapes, at which point the graph becomes un-runable on the graph runtime and has to be executed through the VM.As @kevinthesun said, what we've been working on recently is allowing for non-constant attributes, the
dyn
namespace is mostly to make non-constant attributes explicit. If we need to separate all dynamic shapes, I think we'd need to implement avm
specific op namespace, and I'm not sure we want to go that far?There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@lixiaoquan Data independent shape func is still there and can handle dynamic input shape cases. @mbrookhart I think for now we can limit dyn namespace to Expr attrs and later consider how to have a more uniform interface.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for pointing that out, I just realize 'static' version op still have some dynamic shape handling ability. I think I misunderstood 'dyn' namespace before.