repo stringlengths 1 99 | file stringlengths 13 215 | code stringlengths 12 59.2M | file_length int64 12 59.2M | avg_line_length float64 3.82 1.48M | max_line_length int64 12 2.51M | extension_type stringclasses 1
value |
|---|---|---|---|---|---|---|
gluon-cv | gluon-cv-master/gluoncv/model_zoo/ssd/presets.py | """SSD predefined models."""
from __future__ import absolute_import
import warnings
from .ssd import get_ssd
from .anchor import LiteAnchorGenerator
from .vgg_atrous import vgg16_atrous_300, vgg16_atrous_512
from ...data import VOCDetection
__all__ = ['ssd_300_vgg16_atrous_voc',
'ssd_300_vgg16_atrous_coco'... | 38,888 | 43.597477 | 98 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/ssd/anchor.py | # pylint: disable=unused-import,too-many-function-args
"""Anchor box generator for SSD detector."""
from __future__ import absolute_import
import numpy as np
from mxnet import gluon
class SSDAnchorGenerator(gluon.HybridBlock):
"""Bounding box anchor generator for Single-shot Object Detection.
Parameters
... | 5,184 | 42.208333 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/ssd/vgg_atrous.py | # pylint: disable=arguments-differ
"""VGG atrous network for object detection."""
from __future__ import division
import os
import mxnet as mx
from mxnet import gluon
from mxnet.gluon import nn
from mxnet.initializer import Xavier
__all__ = ['VGGAtrousExtractor', 'get_vgg_atrous_extractor', 'vgg16_atrous_300',
... | 7,829 | 36.285714 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/ssd/resnet_v1b_ssd.py | # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not u... | 4,701 | 44.211538 | 98 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/rcnn.py | """RCNN Model."""
from __future__ import absolute_import
import warnings
import mxnet as mx
from mxnet import gluon
from mxnet.gluon import nn
from ...nn.coder import NormalizedBoxCenterDecoder, MultiPerClassDecoder
from ...nn.feature import FPNFeatureExpander
class RCNN(gluon.HybridBlock):
"""RCNN network.
... | 16,602 | 45.63764 | 118 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/faster_rcnn/data_parallel.py | """Data parallel task for Faster RCNN Model."""
from mxnet import autograd
from gluoncv.utils.parallel import Parallelizable
class ForwardBackwardTask(Parallelizable):
""" Faster R-CNN training task that can be scheduled concurrently using Parallel.
Parameters
----------
net : gluon.HybridBlock
... | 3,970 | 43.617978 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/faster_rcnn/faster_rcnn.py | """Faster RCNN Model."""
# pylint: disable=not-callable
from __future__ import absolute_import
import os
import mxnet as mx
from mxnet import autograd
from mxnet.gluon import nn
from mxnet.gluon.contrib.nn import SyncBatchNorm
from .rcnn_target import RCNNTargetSampler, RCNNTargetGenerator
from ..rcnn import custom_... | 29,169 | 48.273649 | 109 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/faster_rcnn/rcnn_target.py | """RCNN Target Generator."""
from __future__ import absolute_import
from mxnet import autograd
from mxnet import gluon
from ....nn.coder import MultiClassEncoder, NormalizedPerClassBoxCenterEncoder
class RCNNTargetSampler(gluon.HybridBlock):
"""A sampler to choose positive/negative samples from RCNN Proposals
... | 9,106 | 45.464286 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/faster_rcnn/doublehead_rcnn.py | # pylint: disable=not-callable
"""
Double Head Faster RCNN Model. https://arxiv.org/abs/1904.06493
title={Rethinking Classification and Localization for Object Detection}
author={Wu, Yue and Chen, Yinpeng and Yuan}
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition}
year={2020}
The... | 31,849 | 48.303406 | 125 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/faster_rcnn/predefined_models.py | """Predefined Faster RCNN Model."""
from __future__ import absolute_import
import warnings
import mxnet as mx
from mxnet.gluon import nn
from mxnet.gluon.contrib.nn import SyncBatchNorm
from ..faster_rcnn import get_faster_rcnn
from .doublehead_rcnn import get_doublehead_rcnn
from ....nn.feature import FPNFeatureExp... | 41,542 | 49.724054 | 129 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/mask_rcnn/data_parallel.py | """Data parallel task for Mask R-CNN Model."""
import mxnet as mx
from mxnet import autograd
from gluoncv.utils.parallel import Parallelizable
class ForwardBackwardTask(Parallelizable):
""" Mask R-CNN training task that can be scheduled concurrently using Parallel.
Parameters
----------
net : gluon.... | 4,965 | 44.981481 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/mask_rcnn/mask_rcnn.py | """Mask R-CNN Model."""
from __future__ import absolute_import
import os
import warnings
import mxnet as mx
from mxnet import autograd
from mxnet.gluon import nn
from mxnet.gluon.contrib.nn import SyncBatchNorm
from .rcnn_target import MaskTargetGenerator
from ..faster_rcnn import FasterRCNN
from ..rcnn import custo... | 20,887 | 45.834081 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/mask_rcnn/rcnn_target.py | """Mask Target Generator."""
from __future__ import absolute_import
from mxnet import gluon, autograd
class MaskTargetGenerator(gluon.HybridBlock):
"""Mask RCNN target encoder to generate mask targets.
Parameters
----------
num_images : int
Number of input images.
num_rois : int
... | 4,236 | 38.971698 | 96 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/mask_rcnn/predefined_models.py | """Predefined Mask RCNN Model."""
from __future__ import absolute_import
import mxnet as mx
from mxnet.gluon import nn
from mxnet.gluon.contrib.nn import SyncBatchNorm
from ..mask_rcnn import get_mask_rcnn
from ....nn.feature import FPNFeatureExpander
__all__ = ['mask_rcnn_resnet50_v1b_coco',
'mask_rcnn_f... | 25,245 | 50.627812 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/rpn/rpn.py | """Region Proposal Networks Definition."""
from __future__ import absolute_import
import mxnet as mx
from mxnet import autograd
from mxnet import gluon
from mxnet.gluon import nn
from .anchor import RPNAnchorGenerator
from .proposal import RPNProposal
class RPN(gluon.HybridBlock):
r"""Region Proposal Network.
... | 11,065 | 43.441767 | 118 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/rpn/rpn_target.py | """Region Proposal Target Generator."""
from __future__ import absolute_import
import numpy as np
import mxnet as mx
from mxnet import autograd, gluon
from ....nn.bbox import BBoxSplit
from ....nn.coder import SigmoidClassEncoder, NumPyNormalizedBoxCenterEncoder
try:
import cython_rpn_target
except ImportError:
... | 7,021 | 40.064327 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/rpn/anchor.py | """RPN anchors."""
from __future__ import absolute_import
import numpy as np
from mxnet import gluon
class RPNAnchorGenerator(gluon.HybridBlock):
r"""Anchor generator for Region Proposal Networks.
Parameters
----------
stride : int
Feature map stride with respect to original image.
T... | 5,655 | 35.025478 | 95 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/rpn/proposal.py | """RPN proposals."""
from __future__ import absolute_import
from mxnet import autograd
from mxnet import gluon
from ....nn.bbox import BBoxCornerToCenter, BBoxClipToImage
from ....nn.coder import NormalizedBoxCenterDecoder
class RPNProposal(gluon.HybridBlock):
"""Proposal generator for RPN.
RPNProposal tak... | 3,673 | 42.223529 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/rcnn/rpn/bbox_clip.py | # pylint: disable=unused-argument
"""Custom OP: BBoxClipToImage, used to clip bbox to image edges."""
import mxnet as mx
class BBoxClipToImage(mx.operator.CustomOp):
"""Clip bounding box to image edges.
Parameters
----------
axis : int
The coordinate axis with length 4.
"""
def __ini... | 2,009 | 29 | 71 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/actionrec_inceptionv3.py | # pylint: disable=line-too-long,too-many-lines,missing-docstring,arguments-differ,unused-argument
import mxnet as mx
from mxnet import init
from mxnet.gluon import nn
from mxnet.gluon.nn import HybridBlock
from ..inception import inception_v3
__all__ = ['inceptionv3_ucf101', 'inceptionv3_hmdb51', 'inceptionv3_kinetics... | 11,413 | 45.210526 | 110 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/c3d.py | """C3D, implemented in Gluon. https://arxiv.org/abs/1412.0767"""
# pylint: disable=arguments-differ,unused-argument
__all__ = ['C3D', 'c3d_kinetics400']
from mxnet import init
from mxnet.context import cpu
from mxnet.gluon.block import HybridBlock
from mxnet.gluon import nn
class C3D(HybridBlock):
r"""
The C... | 6,326 | 39.299363 | 96 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/p3d.py | """P3D, implemented in Gluon. https://arxiv.org/abs/1711.10305.
Code adapted from https://github.com/qijiezhao/pseudo-3d-pytorch."""
# pylint: disable=arguments-differ,unused-argument,line-too-long
__all__ = ['P3D', 'p3d_resnet50_kinetics400', 'p3d_resnet101_kinetics400']
from mxnet import init
from mxnet.context imp... | 21,950 | 41.623301 | 103 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/non_local.py | """Non-local block for video action recognition"""
# pylint: disable=line-too-long,too-many-lines,missing-docstring,arguments-differ,unused-argument
from mxnet.gluon.block import HybridBlock
from mxnet import init
from mxnet.gluon import nn
from mxnet.gluon.nn import BatchNorm
def build_nonlocal_block(cfg):
""" Bu... | 7,121 | 46.48 | 136 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/slowfast.py | """SlowFast, implemented in Gluon. https://arxiv.org/abs/1812.03982.
Code adapted from https://github.com/r1ch88/SlowFastNetworks."""
# pylint: disable=line-too-long,too-many-lines,missing-docstring,arguments-differ,unused-argument
__all__ = ['SlowFast', 'slowfast_4x16_resnet50_kinetics400', 'slowfast_8x8_resnet50_kin... | 49,353 | 48.852525 | 160 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/actionrec_resnetv1b.py | # pylint: disable=line-too-long,too-many-lines,missing-docstring,arguments-differ,unused-argument
import mxnet as mx
from mxnet import init
from mxnet.gluon import nn
from mxnet.gluon.nn import HybridBlock
from ..resnetv1b import resnet18_v1b, resnet34_v1b, resnet50_v1b, resnet101_v1b, resnet152_v1b
__all__ = ['resnet... | 34,845 | 44.85 | 97 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/actionrec_vgg16.py | # pylint: disable=line-too-long,too-many-lines,missing-docstring,arguments-differ,unused-argument
import mxnet as mx
from mxnet import init
from mxnet.gluon import nn
from mxnet.gluon.nn import HybridBlock
from ..vgg import vgg16
__all__ = ['vgg16_ucf101', 'vgg16_hmdb51', 'vgg16_kinetics400', 'vgg16_sthsthv2']
class ... | 10,153 | 42.767241 | 119 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/i3d_slow.py | # pylint: disable=missing-function-docstring, line-too-long, unused-argument
"""
SlowFast Networks for Video Recognition
ICCV 2019, https://arxiv.org/abs/1812.03982
Code adapted from https://github.com/open-mmlab/mmaction and
https://github.com/decisionforce/TPN
"""
from mxnet import init
from mxnet.context import cpu
... | 17,514 | 41.512136 | 143 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/actionrec_inceptionv1.py | # pylint: disable=line-too-long,too-many-lines,missing-docstring,arguments-differ,unused-argument
import mxnet as mx
from mxnet import init
from mxnet.gluon import nn
from mxnet.gluon.nn import HybridBlock
from gluoncv.model_zoo.googlenet import googlenet
__all__ = ['inceptionv1_ucf101', 'inceptionv1_hmdb51', 'incepti... | 12,615 | 42.805556 | 110 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/i3d_inceptionv3.py | # pylint: disable=line-too-long,too-many-lines,missing-docstring,arguments-differ,unused-argument
__all__ = ['I3D_InceptionV3', 'i3d_inceptionv3_kinetics400']
from mxnet import nd
from mxnet import init
from mxnet.context import cpu
from mxnet.gluon.block import HybridBlock
from mxnet.gluon import nn
from mxnet.gluon... | 17,434 | 49.390173 | 139 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/i3d_resnet.py | # pylint: disable=line-too-long,too-many-lines,missing-docstring,arguments-differ,unused-argument
# Code adapted from https://github.com/open-mmlab/mmaction.
__all__ = ['I3D_ResNetV1', 'i3d_resnet50_v1_kinetics400', 'i3d_resnet101_v1_kinetics400',
'i3d_nl5_resnet50_v1_kinetics400', 'i3d_nl10_resnet50_v1_kin... | 57,330 | 44.28515 | 147 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/r2plus1d.py | # pylint: disable=arguments-differ,unused-argument,line-too-long
"""R2Plus1D, implemented in Gluon. https://arxiv.org/abs/1711.11248.
Code adapted from https://github.com/pytorch/vision/blob/master/torchvision/models/video/resnet.py."""
__all__ = ['R2Plus1D', 'r2plus1d_resnet18_kinetics400',
'r2plus1d_resn... | 25,164 | 40.595041 | 104 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/action_recognition/i3d_inceptionv1.py | # pylint: disable=line-too-long,too-many-lines,missing-docstring,arguments-differ,unused-argument
__all__ = ['I3D_InceptionV1', 'i3d_inceptionv1_kinetics400']
from mxnet import nd
from mxnet import init
from mxnet.context import cpu
from mxnet.gluon.block import HybridBlock
from mxnet.gluon import nn
from mxnet.gluon... | 19,083 | 48.697917 | 160 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/center_net/deconv_dla.py | """DLA network with Deconvolution layers for CenterNet object detection."""
# pylint: disable=arguments-differ,unused-argument
from __future__ import absolute_import
import warnings
import numpy as np
from mxnet.context import cpu
from mxnet.gluon import nn
from mxnet.gluon import contrib
from . deconv_resnet import ... | 10,850 | 36.417241 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/center_net/target_generator.py | """CenterNet training target generator."""
from __future__ import absolute_import
import numpy as np
from mxnet import nd
from mxnet import gluon
class CenterNetTargetGenerator(gluon.Block):
"""Target generator for CenterNet.
Parameters
----------
num_class : int
Number of categories.
o... | 5,184 | 31.816456 | 98 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/center_net/duc_mobilenet.py | """MobileNet variants with DUC upsampling layers for CenterNet object detection."""
# pylint: disable=unused-argument
from __future__ import absolute_import
import warnings
from mxnet.context import cpu
from mxnet.gluon import nn
from .. model_zoo import get_model
from ...nn.block import DUC
__all__ = ['DUCMobilenet... | 3,391 | 31.304762 | 97 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/center_net/deconv_resnet.py | """ResNet with Deconvolution layers for CenterNet object detection."""
# pylint: disable=unused-argument
from __future__ import absolute_import
import warnings
import math
import mxnet as mx
from mxnet.context import cpu
from mxnet.gluon import nn
from mxnet.gluon import contrib
from .. model_zoo import get_model
__... | 9,535 | 34.715356 | 93 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/center_net/center_net.py | """CenterNet object detector: Objects as Points, https://arxiv.org/abs/1904.07850"""
from __future__ import absolute_import
import os
import warnings
from collections import OrderedDict
import mxnet as mx
from mxnet.gluon import nn
from mxnet import autograd
from ...nn.coder import CenterNetDecoder
__all__ = ['Cente... | 39,660 | 41.738147 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/siamrpn/siam_rpn.py | """siamRPN RPN
Code adapted from https://github.com/STVIR/pysot"""
# pylint: disable=arguments-differ
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from mxnet.gluon.block import HybridBlock
from mxnet.gluon import nn
... | 4,286 | 37.276786 | 98 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/siamrpn/siam_net.py | """SiamRPN network
Code adapted from https://github.com/STVIR/pysot"""
# coding:utf-8
# pylint: disable=arguments-differ,unused-argument
from mxnet.gluon.block import HybridBlock
from mxnet.context import cpu
import mxnet as mx
from gluoncv.model_zoo.siamrpn.siam_alexnet import alexnetlegacy
from gluoncv.model_zoo.siam... | 3,168 | 30.69 | 88 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/siamrpn/siamrpn_tracker.py | """ siamrpn_tracker
Code adapted from https://github.com/STVIR/pysot """
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import math
from collections import namedtuple
import numpy as np
import mxnet as mx
from mxnet im... | 18,230 | 33.140449 | 96 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/siamrpn/siam_alexnet.py | """Alexnet, implemented in Gluon.
Code adapted from https://github.com/STVIR/pysot"""
# coding: utf-8
# pylint: disable=arguments-differ,unused-argument
from __future__ import division
from mxnet.gluon import nn
from mxnet.gluon.block import HybridBlock
from mxnet.context import cpu
class AlexNetLegacy(HybridBlock):
... | 1,989 | 43.222222 | 83 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/simple_pose/mobile_pose.py | # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not u... | 4,103 | 40.04 | 132 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/simple_pose/pose_target.py | """
Target generator for Simple Baselines for Human Pose Estimation and Tracking
(https://arxiv.org/abs/1804.06208)
---------------------------------------------
Copyright (c) Microsoft
Licensed under the MIT License.
Written by Bin Xiao (Bin.Xiao@microsoft.com)
---------------------------------------------
"""
import... | 3,637 | 38.543478 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/simple_pose/simple_pose_resnet.py | # ------------------------------------------------------------------------------
# Copyright (c) Microsoft
# Licensed under the MIT License.
# Written by Bin Xiao (Bin.Xiao@microsoft.com)
# ------------------------------------------------------------------------------
# coding: utf-8
# pylint: disable=missing-docstrin... | 9,511 | 39.649573 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/alpha_pose/utils.py | """Fast Pose Utils for loading parameters"""
from mxnet import random
from mxnet import ndarray
from mxnet.initializer import Initializer
def _try_load_parameters(self, filename=None, model=None, ctx=None, allow_missing=False,
ignore_extra=False):
def getblock(parent, name):
if le... | 3,282 | 31.50495 | 88 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/alpha_pose/fast_pose.py | """Fast pose network for alpha pose"""
# pylint: disable=arguments-differ
import math
import os
import mxnet as mx
from mxnet import initializer
from mxnet.gluon import nn
from mxnet.gluon.block import HybridBlock
from .utils import ZeroUniform, _try_load_parameters
__all__ = ['get_alphapose', 'alpha_pose_resnet101_... | 12,349 | 37.235294 | 96 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/monodepthv2/resnet_encoder.py | """Encoder module of Monodepth2
Code partially borrowed from
https://github.com/nianticlabs/monodepth2/blob/master/networks/resnet_encoder.py
"""
from __future__ import absolute_import, division, print_function
import os
import numpy as np
import mxnet as mx
from mxnet.gluon import nn
from mxnet.context import cpu
fr... | 4,457 | 40.663551 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/monodepthv2/pose_decoder.py | """Decoder module of Monodepth2
Code partially borrowed from
https://github.com/nianticlabs/monodepth2/blob/master/networks/pose_decoder.py
"""
# Copyright Niantic 2019. Patent Pending. All rights reserved.
#
# This software is licensed under the terms of the Monodepth2 licence
# which allows for non-commercial use onl... | 3,739 | 34.283019 | 94 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/monodepthv2/depth_decoder.py | """Decoder module of Monodepth2
Code partially borrowed from
https://github.com/nianticlabs/monodepth2/blob/master/networks/depth_decoder.py
"""
# Copyright Niantic 2019. Patent Pending. All rights reserved.
#
# This software is licensed under the terms of the Monodepth2 licence
# which allows for non-commercial use on... | 3,877 | 33.936937 | 96 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/monodepthv2/monodepth2.py | """Monodepth
Digging Into Self-Supervised Monocular Depth Estimation, ICCV 2019
https://arxiv.org/abs/1806.01260
"""
from mxnet.gluon import nn
from mxnet.context import cpu
from .resnet_encoder import ResnetEncoder
from .depth_decoder import DepthDecoder
class MonoDepth2(nn.HybridBlock):
r"""Monodepth2
Par... | 5,312 | 32.840764 | 97 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/monodepthv2/monodepth2_posenet.py | """Monodepth
Digging Into Self-Supervised Monocular Depth Estimation, ICCV 2019
https://arxiv.org/abs/1806.01260
"""
from mxnet.gluon import nn
from mxnet.context import cpu
from .resnet_encoder import ResnetEncoder
from .pose_decoder import PoseDecoder
class MonoDepth2PoseNet(nn.HybridBlock):
r"""Monodepth2
... | 5,784 | 36.564935 | 96 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/monodepthv2/layers.py | """The functions for monodepth2 model and loss
Code partially borrowed from
https://github.com/nianticlabs/monodepth2/blob/master/layers.py.
"""
# pylint: disable=unused-argument, missing-function-docstring
# Copyright Niantic 2019. Patent Pending. All rights reserved.
#
# This software is licensed under the terms of t... | 14,511 | 35.462312 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/yolo/yolo_target.py | """Target generators for YOLOs."""
# pylint: disable=arguments-differ, unsupported-assignment-operation
from __future__ import absolute_import
from __future__ import division
import numpy as np
from mxnet import gluon
from mxnet import nd
from mxnet import autograd
from ...nn.bbox import BBoxCornerToCenter, BBoxCenter... | 13,364 | 47.249097 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/yolo/yolo3.py | """You Only Look Once Object Detection v3"""
# pylint: disable=arguments-differ
from __future__ import absolute_import
from __future__ import division
import os
import warnings
import numpy as np
import mxnet as mx
from mxnet import gluon
from mxnet import autograd
from mxnet.gluon import nn
from mxnet.gluon.nn import... | 46,348 | 44.754195 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/model_zoo/yolo/darknet.py | """Darknet as YOLO backbone network."""
# pylint: disable=arguments-differ
from __future__ import absolute_import
import os
import mxnet as mx
from mxnet import gluon
from mxnet.gluon import nn
from mxnet.gluon.nn import BatchNorm
__all__ = ['DarknetV3', 'get_darknet', 'darknet53']
def _conv2d(channel, kernel, paddi... | 7,381 | 38.265957 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/nn/nms.py | """Modified NMS ops as static/dynamic layers"""
import numpy as np
import torch
from torchvision.ops import boxes as box_ops
from torchvision.ops import nms # BC-compat
__all__ = ['batched_nms', 'ml_nms']
def batched_nms(boxes: torch.Tensor, scores: torch.Tensor, idxs: torch.Tensor, iou_threshold: float):
"""
... | 5,323 | 32.068323 | 101 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/nn/batch_norm.py | """Custom BatchNorm implementation"""
# pylint: disable=unused-argument,bad-continuation
import logging
import torch
import torch.distributed as dist
from torch import nn
from torch.autograd.function import Function
from torch.nn import functional as F
from torch.nn import BatchNorm2d
from ..utils.comm import get_worl... | 9,885 | 41.982609 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/nn/deform_conv.py | """Custom and derived deformable convolution"""
# pylint: disable=missing-function-docstring,bad-continuation
import torch
from torch import nn
from torch.nn.modules.utils import _pair
from torchvision.ops import deform_conv2d
__all__ = ['DeformConvWithChangeableStride']
class _NewEmptyTensorOp(torch.autograd.Functi... | 3,468 | 32.355769 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/nn/iou_loss.py | """IOU loss"""
import torch
from torch import nn
class IOULoss(nn.Module):
"""
Intersetion Over Union (IoU) loss which supports three
different IoU computations:
* IoU
* Linear IoU
* gIoU
"""
def __init__(self, loc_loss_type='iou'):
super(IOULoss, self).__init__()
self... | 2,219 | 30.714286 | 65 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/nn/keypoint_loss.py | """Loss layers for keypoints that can be inserted to modules"""
import torch
import torch.nn as nn
__all__ = ['WeightedMSELoss', 'HMFocalLoss']
def _sigmoid(x):
y = torch.clamp(x.sigmoid_(), min=1e-4, max=1-1e-4)
return y
class WeightedMSELoss(nn.Module):
"""Weighted MSE loss layer"""
def __init__(se... | 1,312 | 26.93617 | 93 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/nn/smooth_l1_loss.py | """The smooth l1 loss layer for detection related targets"""
# pylint: disable=redefined-builtin
import torch
def smooth_l1_loss(input: torch.Tensor, target: torch.Tensor, beta: float, reduction: str = "none") -> torch.Tensor:
"""
Smooth L1 loss defined in the Fast R-CNN paper as:
| 0.5 * x... | 2,937 | 39.805556 | 116 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/nn/shape_spec.py | """Internal structure for shape"""
from collections import namedtuple
class ShapeSpec(namedtuple("_ShapeSpec", ["channels", "height", "width", "stride"])):
"""
A simple structure that contains basic shape specification about a tensor.
It is often used as the auxiliary inputs/outputs of models,
to obta... | 612 | 29.65 | 85 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/nn/focal_loss.py | import torch
from torch.nn import functional as F
def sigmoid_focal_loss(inputs: torch.Tensor,
targets: torch.Tensor,
alpha: float = -1,
gamma: float = 2,
reduction: str = "none") -> torch.Tensor:
"""
Loss used in Reti... | 1,805 | 35.857143 | 83 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/nn/group_norm.py | """Group Norm implementation"""
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.nn import Module, Parameter
from torch.nn import init
__all__ = ['GCN', 'NaiveGroupNorm']
class Conv2D(nn.Module):
"""Inline Conv2D for GroupNorm module"""
def __init__(self, in_channels, out_channel... | 5,610 | 35.914474 | 124 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/model_store.py | """Model store which provides pretrained models."""
from __future__ import print_function
import os
import logging
import portalocker
from ..utils.model_utils import download, check_sha1
__all__ = ['get_model_file', 'purge']
_model_sha1 = {name: checksum for checksum, name in [
('854b23e460ab1fbc5a5cd458b439a... | 6,523 | 44.622378 | 98 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/object_detection/model_utils.py | """Detection model utils"""
import torch
from torch import nn
from ...data.structures import Instances
from ...utils.utils import retry_if_cuda_oom
def build_top_module(cfg):
"""Build detection head/top module"""
top_type = cfg.CONFIG.MODEL.DETECTION_TOP_MODULE.NAME
if top_type == "conv":
inp = cf... | 2,779 | 32.493976 | 87 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/object_detection/fcos.py | """FCOS object detector"""
# pylint: disable=all
import logging
import math
from typing import List, Dict
import torch
from torch import nn
from torch.nn import functional as F
from ...nn.shape_spec import ShapeSpec
from ...nn.deform_conv import DeformConvWithChangeableStride
from ...nn.group_norm import NaiveGroupNor... | 29,515 | 38.302264 | 111 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/action_recognition/non_local.py | # pylint: disable=missing-function-docstring, unused-argument
"""
PyTorch implementation of the non-local module from Non-local Neural Networks
CVPR 2018, https://arxiv.org/abs/1711.07971
Code adapted from https://github.com/open-mmlab/mmaction
and https://github.com/AlexHex7/Non-local_pytorch
"""
import torch
import t... | 8,070 | 39.969543 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/action_recognition/slowfast.py | # pylint: disable=missing-function-docstring, line-too-long, unused-argument
"""
SlowFast Networks for Video Recognition
ICCV 2019, https://arxiv.org/abs/1812.03982
Code adapted from https://github.com/r1ch88/SlowFastNetworks
"""
import torch
import torch.nn as nn
from torch.nn import BatchNorm3d
__all__ = ['SlowFast... | 29,896 | 44.714067 | 119 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/action_recognition/r2plus1dv1.py | # pylint: disable=missing-function-docstring, missing-class-docstring, unused-argument
"""R2Plus1D, https://arxiv.org/abs/1711.11248. Code adapted from
https://github.com/pytorch/vision/blob/master/torchvision/models/video/resnet.py."""
import torch
import torch.nn as nn
from torch.nn import BatchNorm3d
__all__ = ['R2... | 14,346 | 38.963788 | 116 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/action_recognition/actionrec_resnetv1b.py | # pylint: disable=missing-function-docstring, unused-argument
"""
C2D video action recognition models based on ResNet
"""
import torch
import torch.nn as nn
import torchvision
__all__ = ['ActionRecResNetV1b', 'resnet18_v1b_kinetics400', 'resnet34_v1b_kinetics400',
'resnet50_v1b_kinetics400', 'resnet101_v1b... | 10,936 | 41.722656 | 108 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/action_recognition/i3d_slow.py | # pylint: disable=missing-function-docstring, line-too-long
"""
SlowFast Networks for Video Recognition
ICCV 2019, https://arxiv.org/abs/1812.03982
Code adapted from https://github.com/open-mmlab/mmaction and
https://github.com/decisionforce/TPN
"""
import torch
import torch.nn as nn
import torch.utils.checkpoint as cp... | 21,704 | 38.898897 | 123 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/action_recognition/tpn.py | # pylint: disable=missing-docstring, unused-argument, line-too-long, too-many-lines
"""
Temporal Pyramid Network for Action Recognition
CVPR 2020, https://arxiv.org/pdf/2004.03548.pdf
Code adapted from https://github.com/decisionforce/TPN
"""
import numpy as np
import torch
import torch.nn as nn
import torch.utils.che... | 41,567 | 35.916519 | 116 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/action_recognition/i3d_resnet.py | # pylint: disable=missing-function-docstring, missing-class-docstring, unused-argument, line-too-long
"""
PyTorch implementation of the I3D model.
Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset
CVPR 2017, https://arxiv.org/abs/1705.07750
"""
import torch
import torch.nn as nn
from torch.nn import B... | 32,983 | 45.131469 | 122 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/action_recognition/r2plus1dv2.py | # pylint: disable=missing-function-docstring, missing-class-docstring
"""
A Closer Look at Spatiotemporal Convolutions for Action Recognition
CVPR 2018, https://arxiv.org/abs/1711.11248
Large-scale weakly-supervised pre-training for video action recognition
CVPR 2019, https://arxiv.org/abs/1905.00561
"""
import torch
i... | 8,793 | 33.085271 | 97 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/action_recognition/ircsnv2.py | """
Video Classification with Channel-Separated Convolutional Networks
ICCV 2019, https://arxiv.org/abs/1904.02811
Large-scale weakly-supervised pre-training for video action recognition
CVPR 2019, https://arxiv.org/abs/1905.00561
"""
# pylint: disable=missing-function-docstring, missing-class-docstring
import torch
im... | 6,632 | 31.674877 | 114 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/pose/directpose_outputs.py | """Directpose outputs"""
# pylint: disable=line-too-long, redefined-builtin, missing-class-docstring, unused-variable, consider-using-enumerate,unused-argument
import logging
import os
from typing import List
import numpy as np
import torch
from torch import nn
import torch.nn.functional as F
from torchvision.ops impor... | 54,790 | 46.520382 | 157 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/pose/directpose.py | """Directpose implementation"""
# pylint: disable=line-too-long, redefined-builtin, missing-class-docstring, unused-variable,unnecessary-comprehension
import math
from typing import List, Dict
import torch
from torch import nn
from torch.nn import functional as F
from torch.autograd import Variable
import numpy as np
... | 18,490 | 43.236842 | 146 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/pose/directpose_resnet_fpn.py | """Directpose based on resnet + FPN backbones"""
# pylint: disable=line-too-long, redefined-builtin, missing-class-docstring, unused-variable,unnecessary-pass,arguments-differ
# adpated from https://github.com/aim-uofa/AdelaiDet/blob/master/adet/modeling/backbone
import math
from abc import ABCMeta, abstractmethod
imp... | 32,109 | 39.1375 | 143 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/model_zoo/video_language/coot_model.py | from collections import OrderedDict
import os
import numpy as np
import torch
from torch import nn
from gluoncv.torch.utils.coot_utils import truncated_normal_fill
class MultiModalTransformer:
def __init__(self,
cfg,
use_cuda: bool = True,
use_multi_gpu: bool = ... | 21,404 | 40.563107 | 90 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/engine/launch.py | """Multiprocessing distributed data parallel support"""
import torch
import torch.distributed as dist
import torch.backends.cudnn as cudnn
import torch.multiprocessing as mp
def get_local_ip_and_match(ip_list):
import socket
s = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
s.connect(("8.8.8.8", 80))
... | 1,897 | 36.215686 | 88 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/engine/config/directpose.py | """Default setting in training/testing for directpose"""
from yacs.config import CfgNode as CN
_C = CN()
# ---------------------------------------------------------------------------- #
# Distributed DataParallel setting: DDP_CONFIG
# ---------------------------------------------------------------------------- #
_C... | 13,822 | 42.468553 | 124 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/coot_utils.py | """
Utility functions for COOT model
"""
import ctypes
import datetime
import logging
import multiprocessing as mp
import os
from pathlib import Path
import random
import sys
from typing import Tuple, Dict
import numpy as np
import torch
import torch.backends.cudnn as cudnn
from torch import cuda
import torch.nn.funct... | 11,745 | 32.087324 | 93 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/comm.py | """Pytorch communication tools"""
import functools
import logging
import pickle
import numpy as np
import torch
import torch.distributed as dist
def get_world_size():
"""Get DDP world size"""
if not dist.is_available():
return 1
if not dist.is_initialized():
return 1
return dist.get_wor... | 8,165 | 28.374101 | 96 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/model_utils.py | """
Utility functions for model
"""
import os
import hashlib
import requests
from tqdm import tqdm
import torch
def deploy_model(model, cfg):
"""
Deploy model to multiple GPUs for DDP training.
"""
if cfg.DDP_CONFIG.DISTRIBUTED:
if cfg.DDP_CONFIG.GPU is not None:
torch.cuda.set_de... | 6,487 | 36.287356 | 110 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/loss.py | """
Loss functions used for training COOT model
"""
from torch import nn
import torch
__all__ = ['MaxMarginRankingLoss', 'CycleConsistencyCootLoss', 'IOULoss']
class MaxMarginRankingLoss(nn.Module):
"""MaxMarginRanking loss used in COOT paper (Eq. 1): https://arxiv.org/abs/2011.00597
Inputs:
- x_emb... | 8,918 | 39.540909 | 102 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/utils.py | # pylint: disable=missing-function-docstring, line-too-long, consider-using-with
"""
Utility functions, misc
"""
import os
import sys
import time
import functools
import logging
from contextlib import contextmanager
from functools import wraps
import numpy as np
import torch.nn as nn
import torch
def read_labelmap(l... | 9,803 | 30.728155 | 100 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/visualizer.py | """Visualization utils"""
# pylint: disable=line-too-long, redefined-builtin, missing-class-docstring, unused-variable
# adapted from https://github.com/facebookresearch/detectron2/blob/master/detectron2/utils/visualizer.py
import colorsys
import logging
import math
from enum import Enum, unique
import numpy as np
impo... | 52,762 | 38.761115 | 104 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/random.py | """Utils for random states"""
import os
from datetime import datetime
import random
import logging
import numpy as np
import torch
__all__ = ['seed_all_rng']
def seed_all_rng(seed=None):
"""
Set the random seed for the RNG in torch, numpy and python.
Args:
seed (int): if None, will use a strong... | 738 | 22.83871 | 68 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/lr_policy.py | """Learning rate policy
Gradually warm-up(increasing) learning rate for pytorch's optimizer.
Proposed in 'Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour'.
Code adapted from https://github.com/ildoonet/pytorch-gradual-warmup-lr
"""
# pylint: disable=missing-function-docstring, line-too-long, inconsistent-ret... | 8,103 | 38.921182 | 152 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/optimizer.py | """
optimizer RADAM
"""
import math
from enum import Enum
import torch
from torch.optim.optimizer import Optimizer
class RAdam(Optimizer):
"""RADam optimizer
"""
def __init__(self,
params,
lr=1e-3,
betas=(0.9, 0.999),
eps=1e-8,
... | 7,359 | 37.941799 | 81 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/task_utils/coot.py | # pylint: disable=line-too-long
"""
Utility functions for task
"""
from timeit import default_timer as timer
import torch
from torch.nn import functional as F
from .. import coot_utils
from ..coot_utils import compute_constrastive_loss, compute_cmc_loss
from ..coot_utils import unpack_data
def train_coot(config, bas... | 8,561 | 39.578199 | 99 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/task_utils/classification.py | # pylint: disable=line-too-long
"""
Utility functions for video action recognition task
"""
import os
import logging
import time
import numpy as np
import torch
from ..utils import AverageMeter, accuracy
logger = logging.getLogger(__name__)
def train_classification(base_iter,
model,
... | 9,299 | 41.272727 | 105 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/task_utils/pose.py | """Pose utils"""
import time
import logging
import datetime
import numpy as np
import torch
from .. import comm
from ..utils import AverageMeter
from ..model_utils import save_model
from ..optimizer import maybe_add_gradient_clipping
from ..eval_utils.coco_eval import COCOEvaluator
logger = logging.getLogger(__name_... | 9,957 | 41.016878 | 115 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/tvm_utils/nms.py | """A custom nms module for tvm with static shape"""
# pylint: disable=unused-argument
from tvm import relay
from tvm.relay.frontend.pytorch import _op, AttrCvt, get_relay_op
def nms(inputs, input_types):
"""A static NMS hack for torchvision.nms, it requires the nms to be the last layer of the network"""
boxes ... | 1,750 | 34.02 | 104 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/utils/eval_utils/coco_eval.py | """COCO evaluation utils"""
# pylint: disable=line-too-long, redefined-builtin
import contextlib
import copy
import io
import itertools
import json
import logging
import os
import pickle
from collections import OrderedDict
import numpy as np
import torch
import pandas as pd
from .. import comm
from ...data.registry.c... | 20,786 | 39.206963 | 168 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/data/video_cls/multigrid_helper.py | """Multigrid support to speed up training of video models"""
from functools import reduce
import numpy as np
from torch.utils.data import Sampler
_int_classes = int
__all__ = ['multiGridHelper', 'MultiGridBatchSampler']
sq2 = np.sqrt(2)
class multiGridHelper(object):
"""
A Multigrid Method for Efficiently ... | 6,795 | 38.057471 | 93 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/data/video_cls/dataset_classification.py | """Customized dataloader for general video classification tasks."""
import os
import warnings
import numpy as np
try:
from decord import VideoReader, cpu
except ImportError:
VideoReader = None
cpu = None
import torch
from torch.utils.data import Dataset
from ..transforms.videotransforms import video_trans... | 16,215 | 48.742331 | 117 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/data/registry/__init__.py | import os
from .catalog import DatasetCatalog, MetadataCatalog
from .metadata import _get_builtin_metadata
from .mscoco import register_coco_instances, register_coco_panoptic_separated
# ==== Predefined datasets and splits for COCO ==========
_PREDEFINED_SPLITS_COCO = {}
_PREDEFINED_SPLITS_COCO["coco"] = {
"coco_... | 4,678 | 40.40708 | 109 | py |
gluon-cv | gluon-cv-master/gluoncv/torch/data/pose/dataset_pose.py | """Build dataset for pose estimation"""
# pylint: disable=line-too-long
import itertools
import copy
import logging
import numpy as np
import torch
from ..registry.catalog import DatasetCatalog, MetadataCatalog
from ..detection.detection_dataset import load_proposals_into_dataset, filter_images_with_few_keypoints, fil... | 15,694 | 42.476454 | 175 | py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.