| |
| import torch |
|
|
|
|
| def fp16_clamp(x, min=None, max=None): |
| if not x.is_cuda and x.dtype == torch.float16: |
| |
| return x.float().clamp(min, max).half() |
|
|
| return x.clamp(min, max) |
|
|
|
|
| def bbox_overlaps(bboxes1, bboxes2, mode='iou', is_aligned=False, eps=1e-6): |
| """Calculate overlap between two set of bboxes. |
| |
| FP16 Contributed by https://github.com/open-mmlab/mmdetection/pull/4889 |
| Note: |
| Assume bboxes1 is M x 4, bboxes2 is N x 4, when mode is 'iou', |
| there are some new generated variable when calculating IOU |
| using bbox_overlaps function: |
| |
| 1) is_aligned is False |
| area1: M x 1 |
| area2: N x 1 |
| lt: M x N x 2 |
| rb: M x N x 2 |
| wh: M x N x 2 |
| overlap: M x N x 1 |
| union: M x N x 1 |
| ious: M x N x 1 |
| |
| Total memory: |
| S = (9 x N x M + N + M) * 4 Byte, |
| |
| When using FP16, we can reduce: |
| R = (9 x N x M + N + M) * 4 / 2 Byte |
| R large than (N + M) * 4 * 2 is always true when N and M >= 1. |
| Obviously, N + M <= N * M < 3 * N * M, when N >=2 and M >=2, |
| N + 1 < 3 * N, when N or M is 1. |
| |
| Given M = 40 (ground truth), N = 400000 (three anchor boxes |
| in per grid, FPN, R-CNNs), |
| R = 275 MB (one times) |
| |
| A special case (dense detection), M = 512 (ground truth), |
| R = 3516 MB = 3.43 GB |
| |
| When the batch size is B, reduce: |
| B x R |
| |
| Therefore, CUDA memory runs out frequently. |
| |
| Experiments on GeForce RTX 2080Ti (11019 MiB): |
| |
| | dtype | M | N | Use | Real | Ideal | |
| |:----:|:----:|:----:|:----:|:----:|:----:| |
| | FP32 | 512 | 400000 | 8020 MiB | -- | -- | |
| | FP16 | 512 | 400000 | 4504 MiB | 3516 MiB | 3516 MiB | |
| | FP32 | 40 | 400000 | 1540 MiB | -- | -- | |
| | FP16 | 40 | 400000 | 1264 MiB | 276MiB | 275 MiB | |
| |
| 2) is_aligned is True |
| area1: N x 1 |
| area2: N x 1 |
| lt: N x 2 |
| rb: N x 2 |
| wh: N x 2 |
| overlap: N x 1 |
| union: N x 1 |
| ious: N x 1 |
| |
| Total memory: |
| S = 11 x N * 4 Byte |
| |
| When using FP16, we can reduce: |
| R = 11 x N * 4 / 2 Byte |
| |
| So do the 'giou' (large than 'iou'). |
| |
| Time-wise, FP16 is generally faster than FP32. |
| |
| When gpu_assign_thr is not -1, it takes more time on cpu |
| but not reduce memory. |
| There, we can reduce half the memory and keep the speed. |
| |
| If ``is_aligned`` is ``False``, then calculate the overlaps between each |
| bbox of bboxes1 and bboxes2, otherwise the overlaps between each aligned |
| pair of bboxes1 and bboxes2. |
| |
| Args: |
| bboxes1 (Tensor): shape (B, m, 4) in <x1, y1, x2, y2> format or empty. |
| bboxes2 (Tensor): shape (B, n, 4) in <x1, y1, x2, y2> format or empty. |
| B indicates the batch dim, in shape (B1, B2, ..., Bn). |
| If ``is_aligned`` is ``True``, then m and n must be equal. |
| mode (str): "iou" (intersection over union), "iof" (intersection over |
| foreground) or "giou" (generalized intersection over union). |
| Default "iou". |
| is_aligned (bool, optional): If True, then m and n must be equal. |
| Default False. |
| eps (float, optional): A value added to the denominator for numerical |
| stability. Default 1e-6. |
| |
| Returns: |
| Tensor: shape (m, n) if ``is_aligned`` is False else shape (m,) |
| |
| Example: |
| >>> bboxes1 = torch.FloatTensor([ |
| >>> [0, 0, 10, 10], |
| >>> [10, 10, 20, 20], |
| >>> [32, 32, 38, 42], |
| >>> ]) |
| >>> bboxes2 = torch.FloatTensor([ |
| >>> [0, 0, 10, 20], |
| >>> [0, 10, 10, 19], |
| >>> [10, 10, 20, 20], |
| >>> ]) |
| >>> overlaps = bbox_overlaps(bboxes1, bboxes2) |
| >>> assert overlaps.shape == (3, 3) |
| >>> overlaps = bbox_overlaps(bboxes1, bboxes2, is_aligned=True) |
| >>> assert overlaps.shape == (3, ) |
| |
| Example: |
| >>> empty = torch.empty(0, 4) |
| >>> nonempty = torch.FloatTensor([[0, 0, 10, 9]]) |
| >>> assert tuple(bbox_overlaps(empty, nonempty).shape) == (0, 1) |
| >>> assert tuple(bbox_overlaps(nonempty, empty).shape) == (1, 0) |
| >>> assert tuple(bbox_overlaps(empty, empty).shape) == (0, 0) |
| """ |
|
|
| assert mode in ['iou', 'iof', 'giou'], f'Unsupported mode {mode}' |
| |
| assert (bboxes1.size(-1) == 4 or bboxes1.size(0) == 0) |
| assert (bboxes2.size(-1) == 4 or bboxes2.size(0) == 0) |
|
|
| |
| |
| assert bboxes1.shape[:-2] == bboxes2.shape[:-2] |
| batch_shape = bboxes1.shape[:-2] |
|
|
| rows = bboxes1.size(-2) |
| cols = bboxes2.size(-2) |
| if is_aligned: |
| assert rows == cols |
|
|
| if rows * cols == 0: |
| if is_aligned: |
| return bboxes1.new(batch_shape + (rows, )) |
| else: |
| return bboxes1.new(batch_shape + (rows, cols)) |
|
|
| area1 = (bboxes1[..., 2] - bboxes1[..., 0]) * ( |
| bboxes1[..., 3] - bboxes1[..., 1]) |
| area2 = (bboxes2[..., 2] - bboxes2[..., 0]) * ( |
| bboxes2[..., 3] - bboxes2[..., 1]) |
|
|
| if is_aligned: |
| lt = torch.max(bboxes1[..., :2], bboxes2[..., :2]) |
| rb = torch.min(bboxes1[..., 2:], bboxes2[..., 2:]) |
|
|
| wh = fp16_clamp(rb - lt, min=0) |
| overlap = wh[..., 0] * wh[..., 1] |
|
|
| if mode in ['iou', 'giou']: |
| union = area1 + area2 - overlap |
| else: |
| union = area1 |
| if mode == 'giou': |
| enclosed_lt = torch.min(bboxes1[..., :2], bboxes2[..., :2]) |
| enclosed_rb = torch.max(bboxes1[..., 2:], bboxes2[..., 2:]) |
| else: |
| lt = torch.max(bboxes1[..., :, None, :2], |
| bboxes2[..., None, :, :2]) |
| rb = torch.min(bboxes1[..., :, None, 2:], |
| bboxes2[..., None, :, 2:]) |
|
|
| wh = fp16_clamp(rb - lt, min=0) |
| overlap = wh[..., 0] * wh[..., 1] |
|
|
| if mode in ['iou', 'giou']: |
| union = area1[..., None] + area2[..., None, :] - overlap |
| else: |
| union = area1[..., None] |
| if mode == 'giou': |
| enclosed_lt = torch.min(bboxes1[..., :, None, :2], |
| bboxes2[..., None, :, :2]) |
| enclosed_rb = torch.max(bboxes1[..., :, None, 2:], |
| bboxes2[..., None, :, 2:]) |
|
|
| eps = union.new_tensor([eps]) |
| union = torch.max(union, eps) |
| ious = overlap / union |
| if mode in ['iou', 'iof']: |
| return ious |
| |
| enclose_wh = fp16_clamp(enclosed_rb - enclosed_lt, min=0) |
| enclose_area = enclose_wh[..., 0] * enclose_wh[..., 1] |
| enclose_area = torch.max(enclose_area, eps) |
| gious = ious - (enclose_area - union) / enclose_area |
| return gious |
|
|