You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm on paper branch and trained yolor-p6 on custom dataset. Now I exported it to onnx model.
Shape of my onnx output layer is batch, 3, 96, 160, nc. That differs from shape of forward() function return when doing detect.py.
When I set model.model[-1].export = False in export.py, I will get the correct output layer in my onnx model, but that model does not work for any other inference engine.
Setting model.model[-1].exportTrue or False delivers two different onnx models, why is that the case? Any suggestion of what this problem could solve or how to get the correct output layer in onnx model?
Thanks.
The text was updated successfully, but these errors were encountered:
Hi,
I'm on paper branch and trained yolor-p6 on custom dataset. Now I exported it to onnx model.
Shape of my onnx output layer is
batch, 3, 96, 160, nc
. That differs from shape offorward()
function return when doingdetect.py
.When I set
model.model[-1].export = False
inexport.py
, I will get the correct output layer in my onnx model, but that model does not work for any other inference engine.Setting
model.model[-1].export
True
orFalse
delivers two different onnx models, why is that the case? Any suggestion of what this problem could solve or how to get the correct output layer in onnx model?Thanks.
The text was updated successfully, but these errors were encountered: