repo stringlengths 1 99 | file stringlengths 13 215 | code stringlengths 12 59.2M | file_length int64 12 59.2M | avg_line_length float64 3.82 1.48M | max_line_length int64 12 2.51M | extension_type stringclasses 1
value |
|---|---|---|---|---|---|---|
InDuDoNet | InDuDoNet-main/train.py | # !/usr/bin/env python
# -*- coding:utf-8 -*-
"""
Created on Tue Nov 5 11:56:06 2020
@author: hongwang (hongwang9209@hotmail.com)
MICCAI2021: ``InDuDoNet: An Interpretable Dual Domain Network for CT Metal Artifact Reduction''
paper link: https://arxiv.org/pdf/2109.05298.pdf
"""
from __future__ import print_function
imp... | 6,328 | 44.532374 | 182 | py |
InDuDoNet | InDuDoNet-main/test_deeplesion.py | import os
import os.path
import argparse
import numpy as np
import torch
import time
import matplotlib.pyplot as plt
import h5py
import PIL
from PIL import Image
from network.indudonet import InDuDoNet
from deeplesion.build_gemotry import initialization, build_gemotry
os.environ['CUDA_VISIBLE_DEVICES'] = '0'
parser = ... | 5,780 | 39.145833 | 117 | py |
InDuDoNet | InDuDoNet-main/deeplesion/Dataset.py | import os
import os.path
import numpy as np
import random
import h5py
import torch
import torch.utils.data as udata
import PIL.Image as Image
from numpy.random import RandomState
import scipy.io as sio
import PIL
from PIL import Image
from .build_gemotry import initialization, build_gemotry
param = initialization()
ra... | 2,795 | 34.846154 | 93 | py |
InDuDoNet | InDuDoNet-main/network/priornet.py | import torch
import torch.nn as nn
import torch.nn.functional as F
class UNet(nn.Module):
def __init__(self, n_channels=2, n_classes=1, n_filter=32):
super(UNet, self).__init__()
self.inc = inconv(n_channels, n_filter)
self.down1 = down(n_filter, n_filter*2)
self.down2 = down(n_filt... | 3,438 | 28.646552 | 122 | py |
InDuDoNet | InDuDoNet-main/network/indudonet.py | """
MICCAI2021: ``InDuDoNet: An Interpretable Dual Domain Network for CT Metal Artifact Reduction''
paper link: https://arxiv.org/pdf/2109.05298.pdf
"""
import os
import torch
import torch.nn as nn
import torch.nn.functional as F
from odl.contrib import torch as odl_torch
from .priornet import UNet
import sys
#sys.pat... | 8,679 | 41.54902 | 168 | py |
InDuDoNet | InDuDoNet-main/CLINIC_metal/preprocess_clinic/utils/torch.py | """Helper functions for torch
"""
__all__ = [
"get_device", "is_cuda", "copy_model", "find_layer", "to_npy", "get_last_checkpoint",
"print_model", "save_graph", "backprop_on", "backprop_off", "add_post", "flatten_model",
"FunctionModel"]
import os
import os.path as path
import numpy as np
import torch
imp... | 4,786 | 28.549383 | 100 | py |
InDuDoNet | InDuDoNet-main/CLINIC_metal/preprocess_clinic/utils/__init__.py | from .misc import *
from .torch import *
from .log import Logger | 64 | 20.666667 | 23 | py |
Input-Specific-Certification | Input-Specific-Certification-main/zipdata.py | import multiprocessing
import os.path as op
from threading import local
from zipfile import ZipFile, BadZipFile
from PIL import Image
from io import BytesIO
import torch.utils.data as data
_VALID_IMAGE_TYPES = ['.jpg', '.jpeg', '.tiff', '.bmp', '.png']
class ZipData(data.Dataset):
_IGNORE_ATTRS = {'_zip_file'}
... | 3,481 | 35.270833 | 110 | py |
Input-Specific-Certification | Input-Specific-Certification-main/certify_iss.py | # evaluate a smoothed classifier on a dataset
import argparse
from time import time
from model import resnet110
from datasets import get_dataset, DATASETS, get_num_classes
import numpy as np
from scipy.stats import norm
from statsmodels.stats.proportion import proportion_confint
import torch
from tqdm import tqdm
pars... | 8,532 | 38.50463 | 219 | py |
Input-Specific-Certification | Input-Specific-Certification-main/model.py | '''
ResNet110 for Cifar-10
References:
[1] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR, 2016.
[2] K. He, X. Zhang, S. Ren, and J. Sun. Identity mappings in deep residual networks. In ECCV, 2016.
'''
import torch.nn as nn
import torch.nn.functional as F
import math
def ... | 3,129 | 27.198198 | 101 | py |
Input-Specific-Certification | Input-Specific-Certification-main/datasets.py | import bisect
import os
import pickle
from PIL import Image
import numpy as np
import torch
from torch.utils.data import Dataset, DataLoader
from torchvision import transforms, datasets
from torchvision.datasets.utils import check_integrity
from typing import *
from zipdata import ZipData
# set this environment var... | 14,566 | 36.836364 | 109 | py |
ENCAS | ENCAS-main/evaluate.py | import time
from collections import defaultdict
import json
import torch
import numpy as np
from ofa.imagenet_classification.elastic_nn.utils import set_running_statistics
from networks.attentive_nas_dynamic_model import AttentiveNasDynamicModel
from networks.ofa_mbv3_my import OFAMobileNetV3My
from networks.proxyles... | 13,651 | 49.940299 | 143 | py |
ENCAS | ENCAS-main/nat.py | import itertools
import os
import time
from concurrent.futures.process import ProcessPoolExecutor
from pathlib import Path
import torch
import torch.nn.functional as F
import torchvision.transforms.functional
from torch.cuda.amp import GradScaler
from ofa.utils import AverageMeter, accuracy
from tqdm import tqdm
from... | 54,638 | 52.672888 | 223 | py |
ENCAS | ENCAS-main/utils.py | import atexit
import gzip
import logging
import math
import os
import random
import sys
import yaml
from ofa.utils import count_parameters, measure_net_latency
from pathlib import Path
from ptflops import get_model_complexity_info
from pymoo.factory import get_performance_indicator
from pymoo.util.nds.non_dominated_sor... | 23,267 | 34.577982 | 120 | py |
ENCAS | ENCAS-main/nat_run_many.py | import argparse
import glob
import os
from concurrent.futures.process import ProcessPoolExecutor
from pathlib import Path
import datetime
import torch
from matplotlib import pyplot as plt
import utils
from nat import default_kwargs, main
import yaml
from shutil import copy
import traceback
from concurrent.futures impo... | 9,599 | 41.105263 | 122 | py |
ENCAS | ENCAS-main/dynamic_resolution_collator.py | import random
import copy
import ctypes
import torch
import multiprocessing as mp
import numpy as np
from torchvision import transforms
from utils import onehot, rand_bbox, show_im_from_torch_tensor
class DynamicResolutionCollator:
def __init__(self, n_resolutions_max, if_return_target_idx=True, if_cutmix=Fal... | 4,319 | 43.081633 | 125 | py |
ENCAS | ENCAS-main/utils_train.py | import random
import numpy as np
import torch
from torch.nn.modules.module import Module
# implementation of CutMixCrossEntropyLoss taken from https://github.com/ildoonet/cutmix
class CutMixCrossEntropyLoss(Module):
def __init__(self, size_average=True):
super().__init__()
self.size_average = siz... | 3,420 | 32.213592 | 112 | py |
ENCAS | ENCAS-main/networks/attentive_nas_dynamic_model.py | # taken from https://github.com/facebookresearch/AttentiveNAS
# Difference: images not resized in forward, but beforehand, in the collator (which is faster)
import copy
import random
import collections
import math
import torch
import torch.nn as nn
from torch.utils.checkpoint import checkpoint_sequential, checkpoint
... | 20,925 | 41.189516 | 142 | py |
ENCAS | ENCAS-main/networks/ofa_mbv3_my.py | import copy
import torch
from ofa.imagenet_classification.elastic_nn.modules import DynamicMBConvLayer, DynamicConvLayer, DynamicLinearLayer
from ofa.imagenet_classification.elastic_nn.networks import OFAMobileNetV3
from ofa.imagenet_classification.networks import MobileNetV3
from ofa.utils import val2list, make_divis... | 9,234 | 48.12234 | 148 | py |
ENCAS | ENCAS-main/networks/attentive_nas_static_model.py | # taken from https://github.com/facebookresearch/AttentiveNAS
# Difference: images not resized in forward, but beforehand, in the collator (which is faster)
import torch
import torch.nn as nn
from .modules_alphanet.nn_base import MyNetwork
class AttentiveNasStaticModel(MyNetwork):
def __init__(self, first_conv, ... | 2,694 | 31.083333 | 113 | py |
ENCAS | ENCAS-main/networks/modules_alphanet/dynamic_layers.py | # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved
# adapted from OFA: https://github.com/mit-han-lab/once-for-all
from collections import OrderedDict
import copy
import torch
import torch.nn as nn
import torch.nn.functional as F
from .static_layers import MBInvertedConvLayer, ConvBnActLayer, Lin... | 15,686 | 38.916031 | 152 | py |
ENCAS | ENCAS-main/networks/modules_alphanet/static_layers.py | # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved
# adapted from OFA: https://github.com/mit-han-lab/once-for-all
from collections import OrderedDict
import torch.nn as nn
from .nn_utils import get_same_padding, build_activation, make_divisible, drop_connect
from .nn_base import MyModule
from .act... | 12,039 | 30.76781 | 131 | py |
ENCAS | ENCAS-main/networks/modules_alphanet/activations.py | # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved
# adapted from OFA: https://github.com/mit-han-lab/once-for-all
import torch
import torch.nn as nn
import torch.nn.functional as F
# A memory-efficient implementation of Swish function
class SwishImplementation(torch.autograd.Function):
@stati... | 1,405 | 24.107143 | 70 | py |
ENCAS | ENCAS-main/networks/modules_alphanet/dynamic_ops.py | # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved
# adapted from OFA: https://github.com/mit-han-lab/once-for-all
from torch.autograd.function import Function
import torch.nn.functional as F
from torch.nn.parameter import Parameter
import torch.nn as nn
import torch
from torch.nn.modules._function... | 10,816 | 39.211896 | 106 | py |
ENCAS | ENCAS-main/networks/modules_alphanet/nn_base.py | # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved
# adapted from OFA: https://github.com/mit-han-lab/once-for-all
import math
import torch
import torch.nn as nn
try:
from fvcore.common.file_io import PathManager
except:
pass
class MyModule(nn.Module):
def forward(self, x):
... | 4,767 | 30.576159 | 113 | py |
ENCAS | ENCAS-main/networks/modules_alphanet/nn_utils.py | # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved
# adapted from OFA: https://github.com/mit-han-lab/once-for-all
import torch.nn as nn
from .activations import *
def make_divisible(v, divisor=8, min_value=1):
"""
forked from slim:
https://github.com/tensorflow/models/blob/\
0344... | 3,431 | 29.918919 | 96 | py |
ENCAS | ENCAS-main/run_manager/run_manager_my.py | from collections import defaultdict
import time
import torch.nn as nn
import torch.nn.parallel
import torch.optim
from sklearn.metrics import balanced_accuracy_score
from tqdm import tqdm
import torchvision
from ofa.utils import AverageMeter, accuracy
class RunManagerMy:
def __init__(self, net, run_config, no_gpu... | 9,055 | 50.748571 | 155 | py |
ENCAS | ENCAS-main/after_search/store_outputs_timm.py | import copy
import utils
from collections import defaultdict
import pandas as pd
import timm
import json
import os
from concurrent.futures import ProcessPoolExecutor
from pathlib import Path
from os.path import join
import numpy as np
from ofa.utils import AverageMeter, accuracy
from timm.data import create_dataset... | 18,030 | 109.619632 | 10,476 | py |
ENCAS | ENCAS-main/after_search/evaluate_stored_outputs.py | import copy
import json
import numpy as np
import os
import torch
import glob
import gzip
import yaml
from matplotlib import pyplot as plt
import utils
from utils import execute_func_for_all_runs_and_combine
labels_path_prefix = utils.NAT_DATA_PATH
def evaluate_stored_one_run(run_path, dataset_type, path_labels, *... | 10,795 | 50.409524 | 269 | py |
ENCAS | ENCAS-main/after_search/average_weights.py | import os
import torch
import utils
def swa(run_path, iters, supernet_name_in, supernet_name_out):
checkpoint_paths = [os.path.join(run_path, f'iter_{i}', supernet_name_in) for i in iters]
# read checkpoints
checkpoints = [torch.load(p, map_location='cpu') for p in checkpoint_paths]
state_dicts = [c[... | 2,015 | 39.32 | 125 | py |
ENCAS | ENCAS-main/encas/encas_api.py | import glob
import pickle
import gzip
import numpy as np
import os
import torch
from utils import threshold_gene_to_value_moregranular as threshold_gene_to_value
class EncasAPI:
def __init__(self, filename):
self.use_cache = True
kwargs = pickle.load(open(filename, 'rb'))
self.if_allow_... | 6,361 | 42.278912 | 172 | py |
ENCAS | ENCAS-main/encas/greedy_search.py | # implementation of the algorithm from the paper http://proceedings.mlr.press/v80/streeter18a/streeter18a.pdf
import time
from concurrent.futures import ProcessPoolExecutor
import numpy as np
import torch
import utils
class GreedySearchWrapperEnsembleClassification:
def __init__(self, alphabet, subnet_to_output_d... | 9,745 | 50.294737 | 148 | py |
ENCAS | ENCAS-main/data_utils/store_labels.py | import numpy as np
import torch
import yaml
from run_manager import get_run_config
from utils import NAT_PATH, NAT_DATA_PATH
import os
'''
Store labels of val & test sets for all the datasets.
To not specify all data-related information again, simply use any NAT config that uses the appropriate dataset
'''
def get_v... | 2,011 | 48.073171 | 110 | py |
ENCAS | ENCAS-main/data_providers/auto_augment_tf.py | # taken verbatim from https://github.com/facebookresearch/AttentiveNAS
""" Auto Augment
Implementation adapted from timm: https://github.com/rwightman/pytorch-image-models
"""
import random
import math
from PIL import Image, ImageOps, ImageEnhance
import PIL
_PIL_VER = tuple([int(x) for x in PIL.__version__.split('... | 13,406 | 32.185644 | 105 | py |
ENCAS | ENCAS-main/data_providers/cifar.py | import functools
import math
import numpy as np
import torchvision
import torch.utils.data
import torchvision.transforms as transforms
from ofa.imagenet_classification.data_providers.imagenet import DataProvider
from timm.data import rand_augment_transform
import utils_train
from utils import _pil_interp
from dynam... | 11,902 | 40.618881 | 126 | py |
ENCAS | ENCAS-main/data_providers/imagenet.py | import functools
import warnings
import os
import math
import numpy as np
import torch.utils.data
import torchvision
import torchvision.transforms as transforms
import torchvision.datasets as datasets
from ofa.imagenet_classification.data_providers.imagenet import DataProvider
import utils
import utils_train
from dy... | 14,256 | 46.052805 | 131 | py |
Likelihood-free_OICA | Likelihood-free_OICA-master/LFOICA.py | import numpy as np
import torch
import torch.nn as nn
from torch import Tensor
import torch.optim as optim
import torch.nn.functional as F
from libs.distance_measure.mmd import mix_rbf_mmd2
from libs.common_utils import cos_act, normalize_mixing_matrix
from scipy.stats import semicircular
from scipy.stats import hypsec... | 6,446 | 41.695364 | 118 | py |
Likelihood-free_OICA | Likelihood-free_OICA-master/libs/pytorch_pgm.py | from torch.optim.sgd import SGD
import torch
from torch.optim.optimizer import required
class PGM(SGD):
def __init__(self, params, proxs, lr=required, reg_lambda=0, momentum=0, dampening=0,
nesterov=False):
kwargs = dict(lr=lr, momentum=momentum, dampening=dampening, weight_decay=0, nestero... | 1,798 | 34.98 | 125 | py |
Likelihood-free_OICA | Likelihood-free_OICA-master/libs/common_utils.py | import torch
import numpy as np
def gumbel_softmax(p, device, sample_num=1, tau=0.2):
g = np.random.uniform(0.0001, 0.9999, [sample_num, p.shape[0]])
g = -torch.log(-torch.log(torch.Tensor(g)))
numerator = torch.exp((torch.log(p).repeat(sample_num, 1) + g.to(device)) / tau)
denominator = torch.sum(num... | 1,497 | 33.045455 | 104 | py |
Likelihood-free_OICA | Likelihood-free_OICA-master/libs/distance_measure/kernel_density.py | import torch
def kernel_density_loss(X, Y, sigma_list):
assert (X.size(0) == Y.size(0))
m = X.size(0)
n = X.size(1)
expand_X = X.unsqueeze(2).expand(-1, -1, m).permute(0, 2, 1)
expand_Y = Y.unsqueeze(2).expand(-1, -1, m).permute(2, 0, 1)
X_Y = expand_X - expand_Y
X_Y_norm = torch.zeros((m, ... | 729 | 37.421053 | 113 | py |
Likelihood-free_OICA | Likelihood-free_OICA-master/libs/distance_measure/mmd.py | #!/usr/bin/env python
# encoding: utf-8
import torch
min_var_est = 1e-8
# Consider linear time MMD with a linear kernel:
# K(f(x), f(y)) = f(x)^Tf(y)
# h(z_i, z_j) = k(x_i, x_j) + k(y_i, y_j) - k(x_i, y_j) - k(x_j, y_i)
# = [f(x_i) - f(y_i)]^T[f(x_j) - f(y_j)]
#
# f_of_X: batch_size * k
# f_of_Y: batch... | 7,282 | 37.739362 | 131 | py |
DSLA-DSLA | DSLA-DSLA/setup.py | #!/usr/bin/env python
# Copyright (c) OpenMMLab. All rights reserved.
import os
import os.path as osp
import platform
import shutil
import sys
import warnings
from setuptools import find_packages, setup
import torch
from torch.utils.cpp_extension import (BuildExtension, CppExtension,
... | 8,008 | 34.914798 | 125 | py |
DSLA-DSLA | DSLA-DSLA/tools/test.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import os
import os.path as osp
import time
import warnings
import mmcv
import torch
from mmcv import Config, DictAction
from mmcv.cnn import fuse_conv_bn
from mmcv.parallel import MMDataParallel, MMDistributedDataParallel
from mmcv.runner import (get_dis... | 9,774 | 37.789683 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tools/train.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import copy
import os
import os.path as osp
import time
import warnings
import mmcv
import torch
from mmcv import Config, DictAction
from mmcv.runner import get_dist_info, init_dist
from mmcv.utils import get_git_hash
from mmdet import __version__
from m... | 7,317 | 36.147208 | 115 | py |
DSLA-DSLA | DSLA-DSLA/tools/deployment/test_torchserver.py | from argparse import ArgumentParser
import numpy as np
import requests
from mmdet.apis import inference_detector, init_detector, show_result_pyplot
from mmdet.core import bbox2result
def parse_args():
parser = ArgumentParser()
parser.add_argument('img', help='Image file')
parser.add_argument('config', h... | 2,357 | 30.44 | 77 | py |
DSLA-DSLA | DSLA-DSLA/tools/deployment/mmdet2torchserve.py | # Copyright (c) OpenMMLab. All rights reserved.
from argparse import ArgumentParser, Namespace
from pathlib import Path
from tempfile import TemporaryDirectory
import mmcv
try:
from model_archiver.model_packaging import package_model
from model_archiver.model_packaging_utils import ModelExportUtils
except Imp... | 3,693 | 32.279279 | 78 | py |
DSLA-DSLA | DSLA-DSLA/tools/deployment/onnx2tensorrt.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import os
import os.path as osp
import warnings
import numpy as np
import onnx
import torch
from mmcv import Config
from mmcv.tensorrt import is_tensorrt_plugin_loaded, onnx2trt, save_trt_engine
from mmdet.core.export import preprocess_example_input
from... | 8,516 | 32.4 | 78 | py |
DSLA-DSLA | DSLA-DSLA/tools/deployment/mmdet_handler.py | # Copyright (c) OpenMMLab. All rights reserved.
import base64
import os
import mmcv
import torch
from ts.torch_handler.base_handler import BaseHandler
from mmdet.apis import inference_detector, init_detector
class MMdetHandler(BaseHandler):
threshold = 0.5
def initialize(self, context):
properties ... | 2,560 | 34.569444 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tools/deployment/pytorch2onnx.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import os.path as osp
import warnings
from functools import partial
import numpy as np
import onnx
import torch
from mmcv import Config, DictAction
from mmdet.core.export import build_model_from_cfg, preprocess_example_input
from mmdet.core.export.model_... | 11,781 | 33.052023 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tools/model_converters/selfsup2mmdet.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
from collections import OrderedDict
import torch
def moco_convert(src, dst):
"""Convert keys in pycls pretrained moco models to mmdet style."""
# load caffe model
moco_model = torch.load(src)
blobs = moco_model['state_dict']
# conver... | 1,243 | 27.930233 | 74 | py |
DSLA-DSLA | DSLA-DSLA/tools/model_converters/publish_model.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import subprocess
import torch
def parse_args():
parser = argparse.ArgumentParser(
description='Process a checkpoint to be published')
parser.add_argument('in_file', help='input checkpoint filename')
parser.add_argument('out_file', h... | 1,301 | 28.590909 | 78 | py |
DSLA-DSLA | DSLA-DSLA/tools/model_converters/regnet2mmdet.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
from collections import OrderedDict
import torch
def convert_stem(model_key, model_weight, state_dict, converted_names):
new_key = model_key.replace('stem.conv', 'conv1')
new_key = new_key.replace('stem.bn', 'bn1')
state_dict[new_key] = mode... | 3,063 | 32.67033 | 77 | py |
DSLA-DSLA | DSLA-DSLA/tools/model_converters/upgrade_model_version.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import re
import tempfile
from collections import OrderedDict
import torch
from mmcv import Config
def is_head(key):
valid_head_list = [
'bbox_head', 'mask_head', 'semantic_head', 'grid_head', 'mask_iou_head'
]
return any(key.starts... | 6,848 | 31.459716 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tools/model_converters/upgrade_ssd_version.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import tempfile
from collections import OrderedDict
import torch
from mmcv import Config
def parse_config(config_strings):
temp_file = tempfile.NamedTemporaryFile()
config_path = f'{temp_file.name}.py'
with open(config_path, 'w') as f:
... | 1,789 | 29.338983 | 78 | py |
DSLA-DSLA | DSLA-DSLA/tools/model_converters/detectron2pytorch.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
from collections import OrderedDict
import mmcv
import torch
arch_settings = {50: (3, 4, 6, 3), 101: (3, 4, 23, 3)}
def convert_bn(blobs, state_dict, caffe_name, torch_name, converted_names):
# detectron replace bn with affine channel layer
sta... | 3,578 | 41.607143 | 78 | py |
DSLA-DSLA | DSLA-DSLA/tools/analysis_tools/benchmark.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import copy
import os
import time
import torch
from mmcv import Config, DictAction
from mmcv.cnn import fuse_conv_bn
from mmcv.parallel import MMDistributedDataParallel
from mmcv.runner import init_dist, load_checkpoint, wrap_fp16_model
from mmdet.datase... | 6,416 | 33.132979 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tools/analysis_tools/optimize_anchors.py | # Copyright (c) OpenMMLab. All rights reserved.
"""Optimize anchor settings on a specific dataset.
This script provides two method to optimize YOLO anchors including k-means
anchor cluster and differential evolution. You can use ``--algorithm k-means``
and ``--algorithm differential_evolution`` to switch two method.
... | 13,161 | 34.477089 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tools/analysis_tools/get_flops.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import numpy as np
import torch
from mmcv import Config, DictAction
from mmdet.models import build_detector
try:
from mmcv.cnn import get_model_complexity_info
except ImportError:
raise ImportError('Please upgrade mmcv to >0.6.2')
def parse_ar... | 2,995 | 29.571429 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tools/analysis_tools/test_robustness.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import copy
import os
import os.path as osp
import mmcv
import torch
from mmcv import DictAction
from mmcv.parallel import MMDataParallel, MMDistributedDataParallel
from mmcv.runner import (get_dist_info, init_dist, load_checkpoint,
... | 15,222 | 38.234536 | 79 | py |
DSLA-DSLA | DSLA-DSLA/.dev_scripts/benchmark_filter.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import os
import os.path as osp
def parse_args():
parser = argparse.ArgumentParser(description='Filter configs to train')
parser.add_argument(
'--basic-arch',
action='store_true',
help='to train models in basic arch')
... | 7,101 | 41.27381 | 92 | py |
DSLA-DSLA | DSLA-DSLA/.dev_scripts/gather_models.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import glob
import json
import os.path as osp
import shutil
import subprocess
from collections import OrderedDict
import mmcv
import torch
import yaml
def ordered_yaml_dump(data, stream=None, Dumper=yaml.SafeDumper, **kwds):
class OrderedDumper(Dum... | 10,489 | 34.924658 | 79 | py |
DSLA-DSLA | DSLA-DSLA/.dev_scripts/benchmark_inference_fps.py | # Copyright (c) OpenMMLab. All rights reserved.
import argparse
import os
import os.path as osp
import mmcv
from mmcv import Config, DictAction
from mmcv.runner import init_dist
from terminaltables import GithubFlavoredMarkdownTable
from tools.analysis_tools.benchmark import repeat_measure_inference_speed
def parse_... | 6,763 | 38.788235 | 79 | py |
DSLA-DSLA | DSLA-DSLA/.dev_scripts/batch_test_list.py | # Copyright (c) OpenMMLab. All rights reserved.
# yapf: disable
atss = dict(
config='configs/atss/atss_r50_fpn_1x_coco.py',
checkpoint='atss_r50_fpn_1x_coco_20200209-985f7bd0.pth',
eval='bbox',
metric=dict(bbox_mAP=39.4),
)
autoassign = dict(
config='configs/autoassign/autoassign_r50_fpn_8x2_1x_coco... | 12,707 | 34.3 | 117 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_runtime/async_benchmark.py | # Copyright (c) OpenMMLab. All rights reserved.
import asyncio
import os
import shutil
import urllib
import mmcv
import torch
from mmdet.apis import (async_inference_detector, inference_detector,
init_detector)
from mmdet.utils.contextmanagers import concurrent
from mmdet.utils.profiling impor... | 3,215 | 30.223301 | 77 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_runtime/test_async.py | # Copyright (c) OpenMMLab. All rights reserved.
"""Tests for async interface."""
import asyncio
import os
import sys
import asynctest
import mmcv
import torch
from mmdet.apis import async_inference_detector, init_detector
if sys.version_info >= (3, 7):
from mmdet.utils.contextmanagers import concurrent
class ... | 2,608 | 30.059524 | 75 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_runtime/test_config.py | # Copyright (c) OpenMMLab. All rights reserved.
from os.path import dirname, exists, join
from unittest.mock import Mock
import pytest
from mmdet.core import BitmapMasks, PolygonMasks
from mmdet.datasets.builder import DATASETS
from mmdet.datasets.utils import NumClassCheckHook
def _get_config_directory():
"""F... | 15,152 | 39.733871 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_runtime/test_eval_hook.py | # Copyright (c) OpenMMLab. All rights reserved.
import os.path as osp
import tempfile
import unittest.mock as mock
from collections import OrderedDict
from unittest.mock import MagicMock, patch
import pytest
import torch
import torch.nn as nn
from mmcv.runner import EpochBasedRunner, build_optimizer
from mmcv.utils im... | 8,590 | 32.956522 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_runtime/test_fp16.py | # Copyright (c) OpenMMLab. All rights reserved.
import numpy as np
import pytest
import torch
import torch.nn as nn
from mmcv.runner import auto_fp16, force_fp32
from mmcv.runner.fp16_utils import cast_tensor_type
def test_cast_tensor_type():
inputs = torch.FloatTensor([5.])
src_type = torch.float32
dst_t... | 9,746 | 31.274834 | 75 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_loss.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmcv.utils import digit_version
from mmdet.models.losses import (BalancedL1Loss, CrossEntropyLoss, DiceLoss,
DistributionFocalLoss, FocalLoss,
GaussianFocalLoss,
... | 7,860 | 35.393519 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_forward.py | # Copyright (c) OpenMMLab. All rights reserved.
"""pytest tests/test_forward.py."""
import copy
from os.path import dirname, exists, join
import numpy as np
import pytest
import torch
def _get_config_directory():
"""Find the predefined detector config directory."""
try:
# Assume we are running in the... | 23,230 | 32.092593 | 110 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_necks.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from torch.nn.modules.batchnorm import _BatchNorm
from mmdet.models.necks import (FPN, YOLOXPAFPN, ChannelMapper, CTResNetNeck,
DilatedEncoder, SSDNeck, YOLOV3Neck)
def test_fpn():
"""Tests fpn."""
s = ... | 13,273 | 31.614251 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_plugins.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.plugins import DropBlock
def test_dropblock():
feat = torch.rand(1, 1, 11, 11)
drop_prob = 1.0
dropblock = DropBlock(drop_prob, block_size=11, warmup_iters=0)
out_feat = dropblock(feat)
assert (out_feat =... | 840 | 27.033333 | 67 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_loss_compatibility.py | # Copyright (c) OpenMMLab. All rights reserved.
"""pytest tests/test_loss_compatibility.py."""
import copy
from os.path import dirname, exists, join
import numpy as np
import pytest
import torch
def _get_config_directory():
"""Find the predefined detector config directory."""
try:
# Assume we are run... | 6,361 | 30.49505 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_pvt.py | import pytest
import torch
from mmdet.models.backbones.pvt import (PVTEncoderLayer,
PyramidVisionTransformer,
PyramidVisionTransformerV2)
def test_pvt_block():
# test PVT structure and forward
block = PVTEncoderLayer(
emb... | 3,332 | 31.048077 | 69 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_hourglass.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.backbones.hourglass import HourglassNet
def test_hourglass_backbone():
with pytest.raises(AssertionError):
# HourglassNet's num_stacks should larger than 0
HourglassNet(num_stacks=0)
with pytest.rais... | 1,464 | 28.3 | 65 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_res2net.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.backbones import Res2Net
from mmdet.models.backbones.res2net import Bottle2neck
from .utils import is_block
def test_res2net_bottle2neck():
with pytest.raises(AssertionError):
# Style must be in ['pytorch', 'caff... | 1,976 | 30.380952 | 72 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_swin.py | import pytest
import torch
from mmdet.models.backbones.swin import SwinBlock, SwinTransformer
def test_swin_block():
# test SwinBlock structure and forward
block = SwinBlock(embed_dims=64, num_heads=4, feedforward_channels=256)
assert block.ffn.embed_dims == 64
assert block.attn.w_msa.num_heads == 4
... | 2,648 | 30.915663 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_resnet.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmcv import assert_params_all_zeros
from mmcv.ops import DeformConv2dPack
from torch.nn.modules import AvgPool2d, GroupNorm
from torch.nn.modules.batchnorm import _BatchNorm
from mmdet.models.backbones import ResNet, ResNetV1d
from mmdet.m... | 22,380 | 34.35703 | 78 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/utils.py | # Copyright (c) OpenMMLab. All rights reserved.
from torch.nn.modules import GroupNorm
from torch.nn.modules.batchnorm import _BatchNorm
from mmdet.models.backbones.res2net import Bottle2neck
from mmdet.models.backbones.resnet import BasicBlock, Bottleneck
from mmdet.models.backbones.resnext import Bottleneck as Bottl... | 1,026 | 30.121212 | 77 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_mobilenet_v2.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from torch.nn.modules import GroupNorm
from torch.nn.modules.batchnorm import _BatchNorm
from mmdet.models.backbones.mobilenet_v2 import MobileNetV2
from .utils import check_norm_state, is_block, is_norm
def test_mobilenetv2_backbone():
w... | 6,546 | 36.626437 | 77 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_hrnet.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.backbones.hrnet import HRModule, HRNet
from mmdet.models.backbones.resnet import BasicBlock, Bottleneck
@pytest.mark.parametrize('block', [BasicBlock, Bottleneck])
def test_hrmodule(block):
# Test multiscale forward
... | 3,089 | 26.589286 | 68 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_csp_darknet.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from torch.nn.modules.batchnorm import _BatchNorm
from mmdet.models.backbones.csp_darknet import CSPDarknet
from .utils import check_norm_state, is_norm
def test_csp_darknet_backbone():
with pytest.raises(ValueError):
# frozen_sta... | 4,117 | 34.196581 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_renext.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.backbones import ResNeXt
from mmdet.models.backbones.resnext import Bottleneck as BottleneckX
from .utils import is_block
def test_renext_bottleneck():
with pytest.raises(AssertionError):
# Style must be in ['pyt... | 3,528 | 32.292453 | 73 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_trident_resnet.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.backbones import TridentResNet
from mmdet.models.backbones.trident_resnet import TridentBottleneck
def test_trident_resnet_bottleneck():
trident_dilations = (1, 2, 3)
test_branch_idx = 1
concat_output = True
... | 6,372 | 34.209945 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_resnest.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.backbones import ResNeSt
from mmdet.models.backbones.resnest import Bottleneck as BottleneckS
def test_resnest_bottleneck():
with pytest.raises(AssertionError):
# Style must be in ['pytorch', 'caffe']
Bot... | 1,473 | 29.708333 | 76 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_regnet.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.backbones import RegNet
regnet_test_data = [
('regnetx_400mf',
dict(w0=24, wa=24.48, wm=2.54, group_w=16, depth=22,
bot_mul=1.0), [32, 64, 160, 384]),
('regnetx_800mf',
dict(w0=56, wa=35.73, wm=2.2... | 2,177 | 35.915254 | 73 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_backbones/test_detectors_resnet.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
from mmdet.models.backbones import DetectoRS_ResNet
def test_detectorrs_resnet_backbone():
detectorrs_cfg = dict(
depth=50,
num_stages=4,
out_indices=(0, 1, 2, 3),
frozen_stages=1,
norm_cfg=dict(type='BN', requi... | 1,611 | 32.583333 | 77 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_utils/test_position_encoding.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.utils import (LearnedPositionalEncoding,
SinePositionalEncoding)
def test_sine_positional_encoding(num_feats=16, batch_size=2):
# test invalid type of scale
with pytest.raises(Assertio... | 1,437 | 34.95 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_utils/test_model_misc.py | # Copyright (c) OpenMMLab. All rights reserved.
import numpy as np
import torch
from mmdet.models.utils import interpolate_as
def test_interpolate_as():
source = torch.rand((1, 5, 4, 4))
target = torch.rand((1, 1, 16, 16))
# Test 4D source and target
result = interpolate_as(source, target)
asser... | 805 | 27.785714 | 54 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_utils/test_se_layer.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.utils import SELayer
def test_se_layer():
with pytest.raises(AssertionError):
# act_cfg sequence length must equal to 2
SELayer(channels=32, act_cfg=(dict(type='ReLU'), ))
with pytest.raises(Assertio... | 674 | 26 | 76 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_utils/test_inverted_residual.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmcv.cnn import is_norm
from torch.nn.modules import GroupNorm
from mmdet.models.utils import InvertedResidual, SELayer
def test_inverted_residual():
with pytest.raises(AssertionError):
# stride must be in [1, 2]
Inv... | 2,635 | 33.233766 | 71 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_utils/test_conv_upsample.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmdet.models.utils import ConvUpsample
@pytest.mark.parametrize('num_layers', [0, 1, 2])
def test_conv_upsample(num_layers):
num_upsample = num_layers if num_layers > 0 else 0
num_layers = num_layers if num_layers > 0 else 1
... | 628 | 24.16 | 54 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_utils/test_transformer.py | # Copyright (c) OpenMMLab. All rights reserved.
import pytest
import torch
from mmcv.utils import ConfigDict
from mmdet.models.utils.transformer import (AdaptivePadding,
DetrTransformerDecoder,
DetrTransformerEncoder, PatchEmbed,
... | 16,994 | 28.815789 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_utils/test_brick_wrappers.py | from unittest.mock import patch
import torch
import torch.nn as nn
import torch.nn.functional as F
from mmdet.models.utils import AdaptiveAvgPool2d, adaptive_avg_pool2d
if torch.__version__ != 'parrots':
torch_version = '1.7'
else:
torch_version = 'parrots'
@patch('torch.__version__', torch_version)
def te... | 2,931 | 30.191489 | 69 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_dense_heads/test_lad_head.py | # Copyright (c) OpenMMLab. All rights reserved.
import mmcv
import numpy as np
import torch
from mmdet.models.dense_heads import LADHead, lad_head
from mmdet.models.dense_heads.lad_head import levels_to_images
def test_lad_head_loss():
"""Tests lad head loss when truth is empty and non-empty."""
class mock_... | 5,294 | 34.3 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_dense_heads/test_anchor_head.py | # Copyright (c) OpenMMLab. All rights reserved.
import mmcv
import torch
from mmdet.models.dense_heads import AnchorHead
def test_anchor_head_loss():
"""Tests anchor head loss when truth is empty and non-empty."""
s = 256
img_metas = [{
'img_shape': (s, s, 3),
'scale_factor': 1,
'... | 2,548 | 34.901408 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_dense_heads/test_centernet_head.py | # Copyright (c) OpenMMLab. All rights reserved.
import numpy as np
import torch
from mmcv import ConfigDict
from mmdet.models.dense_heads import CenterNetHead
def test_center_head_loss():
"""Tests center head loss when truth is empty and non-empty."""
s = 256
img_metas = [{
'img_shape': (s, s, 3)... | 4,385 | 39.611111 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_dense_heads/test_ga_anchor_head.py | # Copyright (c) OpenMMLab. All rights reserved.
import mmcv
import torch
from mmdet.models.dense_heads import GuidedAnchorHead
def test_ga_anchor_head_loss():
"""Tests anchor head loss when truth is empty and non-empty."""
s = 256
img_metas = [{
'img_shape': (s, s, 3),
'scale_factor': 1,
... | 3,410 | 36.076087 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_dense_heads/test_yolof_head.py | # Copyright (c) OpenMMLab. All rights reserved.
import mmcv
import torch
from mmdet.models.dense_heads import YOLOFHead
def test_yolof_head_loss():
"""Tests yolof head loss when truth is empty and non-empty."""
s = 256
img_metas = [{
'img_shape': (s, s, 3),
'scale_factor': 1,
'pad... | 2,716 | 34.285714 | 79 | py |
DSLA-DSLA | DSLA-DSLA/tests/test_models/test_dense_heads/test_vfnet_head.py | # Copyright (c) OpenMMLab. All rights reserved.
import mmcv
import torch
from mmdet.models.dense_heads import VFNetHead
def test_vfnet_head_loss():
"""Tests vfnet head loss when truth is empty and non-empty."""
s = 256
img_metas = [{
'img_shape': (s, s, 3),
'scale_factor': 1,
'pad... | 2,561 | 39.03125 | 79 | py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.