Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add custom max iou assigner to prevent CPU OOM in training phase #2228

Merged
merged 4 commits into from
Jun 8, 2023
Merged
Show file tree
Hide file tree
Changes from 3 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ All notable changes to this project will be documented in this file.
### New features

- Support encrypted dataset training (<https://github.com/openvinotoolkit/training_extensions/pull/2209>)
- Add custom max iou assigner to prevent CPU OOM when large annotations are used (<https://github.com/openvinotoolkit/training_extensions/pull/2228>)

### Enhancements

Expand Down
4 changes: 2 additions & 2 deletions otx/algorithms/detection/adapters/mmdet/models/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,6 @@
# SPDX-License-Identifier: Apache-2.0
#

from . import backbones, dense_heads, detectors, heads, losses, necks, roi_heads
from . import assigners, backbones, dense_heads, detectors, heads, losses, necks, roi_heads

__all__ = ["backbones", "dense_heads", "detectors", "heads", "losses", "necks", "roi_heads"]
__all__ = ["assigners", "backbones", "dense_heads", "detectors", "heads", "losses", "necks", "roi_heads"]
Original file line number Diff line number Diff line change
@@ -0,0 +1,8 @@
"""Initial file for mmdetection assigners."""
# Copyright (C) 2023 Intel Corporation
# SPDX-License-Identifier: Apache-2.0
#

from .custom_max_iou_assigner import CustomMaxIoUAssigner

__all__ = ["CustomMaxIoUAssigner"]
Original file line number Diff line number Diff line change
@@ -0,0 +1,107 @@
"""Custom assigner for mmdet MaxIouAssigner."""
# Copyright (C) 2023 Intel Corporation
# SPDX-License-Identifier: Apache-2.0
#

import torch
from mmdet.core.bbox.assigners import MaxIoUAssigner
from mmdet.core.bbox.builder import BBOX_ASSIGNERS


@BBOX_ASSIGNERS.register_module()
class CustomMaxIoUAssigner(MaxIoUAssigner):
"""Assign a corresponding gt bbox or background to each bbox.

Each proposals will be assigned with `-1`, or a semi-positive integer
indicating the ground truth index.

- -1: negative sample, no assigned gt
- semi-positive integer: positive sample, index (0-based) of assigned gt

This CustomMaxIoUAssigner patches assign funtion of mmdet's MaxIouAssigner
so that it can prevent CPU OOM for images whose gt is extremely large
"""

cpu_assign_thr = 1000

def assign(self, bboxes, gt_bboxes, gt_bboxes_ignore=None, gt_labels=None):
"""Assign gt to bboxes.

This method assign a gt bbox to every bbox (proposal/anchor), each bbox
will be assigned with -1, or a semi-positive number. -1 means negative
sample, semi-positive number is the index (0-based) of assigned gt.
The assignment is done in following steps, the order matters.

Especially CustomMaxIoUAssigner split gt_bboxes tensor into small tensors
when gt_bboxes is too large.

1. assign every bbox to the background
2. assign proposals whose iou with all gts < neg_iou_thr to 0
3. for each bbox, if the iou with its nearest gt >= pos_iou_thr,
assign it to that bbox
4. for each gt bbox, assign its nearest proposals (may be more than
one) to itself

Args:
bboxes (Tensor): Bounding boxes to be assigned, shape(n, 4).
gt_bboxes (Tensor): Groundtruth boxes, shape (k, 4).
gt_bboxes_ignore (Tensor, optional): Ground truth bboxes that are
labelled as `ignored`, e.g., crowd boxes in COCO.
gt_labels (Tensor, optional): Label of gt_bboxes, shape (k, ).

Returns:
:obj:`AssignResult`: The assign result.

Example:
>>> self = MaxIoUAssigner(0.5, 0.5)
>>> bboxes = torch.Tensor([[0, 0, 10, 10], [10, 10, 20, 20]])
>>> gt_bboxes = torch.Tensor([[0, 0, 10, 9]])
>>> assign_result = self.assign(bboxes, gt_bboxes)
>>> expected_gt_inds = torch.LongTensor([1, 0])
>>> assert torch.all(assign_result.gt_inds == expected_gt_inds)
"""
assign_on_cpu = True if (self.gpu_assign_thr > 0) and (gt_bboxes.shape[0] > self.gpu_assign_thr) else False
# compute overlap and assign gt on CPU when number of GT is large
if assign_on_cpu:
device = bboxes.device
bboxes = bboxes.cpu()
gt_bboxes = gt_bboxes.cpu()
if gt_bboxes_ignore is not None:
gt_bboxes_ignore = gt_bboxes_ignore.cpu()
if gt_labels is not None:
gt_labels = gt_labels.cpu()

if assign_on_cpu and gt_bboxes.shape[0] > self.cpu_assign_thr:
split_length = gt_bboxes.shape[0] // self.cpu_assign_thr + 1
overlaps = None
for i in range(split_length):
gt_bboxes_split = gt_bboxes[i * self.cpu_assign_thr : (i + 1) * self.cpu_assign_thr]
if overlaps is None:
overlaps = self.iou_calculator(gt_bboxes_split, bboxes)
else:
overlaps = torch.concat((overlaps, self.iou_calculator(gt_bboxes_split, bboxes)), dim=0)

else:
overlaps = self.iou_calculator(gt_bboxes, bboxes)
goodsong81 marked this conversation as resolved.
Show resolved Hide resolved

if (
self.ignore_iof_thr > 0
and gt_bboxes_ignore is not None
and gt_bboxes_ignore.numel() > 0
and bboxes.numel() > 0
):
if self.ignore_wrt_candidates:
ignore_overlaps = self.iou_calculator(bboxes, gt_bboxes_ignore, mode="iof")
ignore_max_overlaps, _ = ignore_overlaps.max(dim=1)
else:
ignore_overlaps = self.iou_calculator(gt_bboxes_ignore, bboxes, mode="iof")
ignore_max_overlaps, _ = ignore_overlaps.max(dim=0)
overlaps[:, ignore_max_overlaps > self.ignore_iof_thr] = -1

assign_result = self.assign_wrt_overlaps(overlaps, gt_labels)
if assign_on_cpu:
assign_result.gt_inds = assign_result.gt_inds.to(device)
assign_result.max_overlaps = assign_result.max_overlaps.to(device)
if assign_result.labels is not None:
assign_result.labels = assign_result.labels.to(device)
return assign_result
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@
train_cfg=dict(
rpn=dict(
assigner=dict(
type="MaxIoUAssigner",
type="CustomMaxIoUAssigner",
pos_iou_thr=0.7,
neg_iou_thr=0.3,
min_pos_iou=0.3,
Expand All @@ -97,7 +97,7 @@
),
rcnn=dict(
assigner=dict(
type="MaxIoUAssigner",
type="CustomMaxIoUAssigner",
pos_iou_thr=0.5,
neg_iou_thr=0.5,
min_pos_iou=0.5,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -91,7 +91,7 @@
train_cfg=dict(
rpn=dict(
assigner=dict(
type="MaxIoUAssigner",
type="CustomMaxIoUAssigner",
pos_iou_thr=0.7,
neg_iou_thr=0.3,
min_pos_iou=0.3,
Expand Down Expand Up @@ -119,7 +119,7 @@
),
rcnn=dict(
assigner=dict(
type="MaxIoUAssigner",
type="CustomMaxIoUAssigner",
pos_iou_thr=0.5,
neg_iou_thr=0.5,
min_pos_iou=0.5,
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,57 @@
"""Unit test for cusom max iou assigner."""
# Copyright (C) 2021-2023 Intel Corporation
# SPDX-License-Identifier: Apache-2.0
#

import pytest
import torch

from otx.algorithms.detection.adapters.mmdet.models.assigners import CustomMaxIoUAssigner
from tests.test_suite.e2e_test_system import e2e_pytest_unit


class TestCustomMaxIoUAssigner:
@pytest.fixture(autouse=True)
def setup(self):
"""Initial setup for unit tests."""
self.assigner = CustomMaxIoUAssigner(
pos_iou_thr=0.5,
neg_iou_thr=0.5,
min_pos_iou=0.5,
match_low_quality=True,
ignore_iof_thr=-1,
gpu_assign_thr=300,
)
self.assigner.cpu_assign_thr = 400

@e2e_pytest_unit
def test_assign_gpu(self):
"""Test custom assign function on gpu."""
gt_bboxes = torch.randn(200, 4)
bboxes = torch.randn(20000, 4)
assign_result = self.assigner.assign(bboxes, gt_bboxes)
assert assign_result.gt_inds.shape == torch.Size([20000])
assert assign_result.max_overlaps.shape == torch.Size([20000])

@e2e_pytest_unit
def test_assign_cpu(self):
"""Test custom assign function on cpu."""
gt_bboxes = torch.randn(350, 4)
bboxes = torch.randn(20000, 4)
assign_result = self.assigner.assign(bboxes, gt_bboxes)
assert assign_result.gt_inds.shape == torch.Size([20000])
assert assign_result.max_overlaps.shape == torch.Size([20000])

@e2e_pytest_unit
def test_assign_cpu_oom(self):
"""Test custom assign function on cpu in case of cpu oom."""
gt_bboxes = torch.randn(450, 4)
bboxes = torch.randn(20000, 4)
assign_result = self.assigner.assign(bboxes, gt_bboxes)
assert assign_result.gt_inds.shape == torch.Size([20000])
assert assign_result.max_overlaps.shape == torch.Size([20000])

self.assigner_cpu_assign_thr = 500
new_assign_result = self.assigner.assign(bboxes, gt_bboxes)
assert torch.all(new_assign_result.gt_inds == assign_result.gt_inds)
assert torch.all(new_assign_result.max_overlaps == assign_result.max_overlaps)