repo stringlengths 1 99 | file stringlengths 13 215 | code stringlengths 12 59.2M | file_length int64 12 59.2M | avg_line_length float64 3.82 1.48M | max_line_length int64 12 2.51M | extension_type stringclasses 1
value |
|---|---|---|---|---|---|---|
GNS-Modeling | GNS-Modeling-master/gns/inference/optimization/optimize_parallel.py | import numpy as np
import torch
import torch.nn as nn
import torch.optim as optim
from .parse import Parse
def optimize_parselist_mt(parse_list, image, loss_fn, iterations=3000,
optimizer_init=None, stopper_init=None,
clip_grad=None):
"""
Parallel version ... | 2,183 | 30.2 | 74 | py |
GNS-Modeling | GNS-Modeling-master/gns/inference/optimization/soft_constraints.py | import torch
import torch.autograd as autograd
import torch.nn.functional as F
__all__ = ['soft_lb', 'soft_ub', 'soft_ub_lb', 'passthrough_lb',
'passthrough_ub', 'passthrough_ub_lb']
# ---- soft bounds ----
def soft_lb(x, vmin):
"""
Activation is approximately linear everywhere except near lower ... | 2,544 | 29.662651 | 77 | py |
GNS-Modeling | GNS-Modeling-master/gns/inference/optimization/optimizers.py | """
Stochastic gradient descent with noisy gradients (instead of minibatch).
"""
import math
import random
import torch
import torch.optim as optim
class SGD(optim.SGD):
"""
This version uses multiplicative gradient noise.
"""
def __init__(self, params, noise=0.1, **kwargs):
self.noise = nois... | 5,339 | 35.326531 | 86 | py |
GNS-Modeling | GNS-Modeling-master/gns/inference/optimization/objectives.py | import torch
import torch.distributions as dist
from pybpl.util import nested_map
def to_cuda(x):
return x.cuda()
class FullModel:
"""
Full graphical model with the following joint distribution:
P(type, token, image) = P(type) * P(token | type) * P(image | token)
"""
def __init__(self, ... | 2,914 | 35.898734 | 97 | py |
GNS-Modeling | GNS-Modeling-master/gns/inference/optimization/optimize.py | from tqdm import tqdm
import numpy as np
import torch
import torch.nn as nn
import torch.optim as optim
from .parse import Parse, ParseWithToken
from .early_stopping import EarlyStopper
def optimize_parse(parse, image, loss_fn, iterations=3000, optimizer=None,
tune_blur=True, tune_fn=None, stoppe... | 5,518 | 36.290541 | 85 | py |
GNS-Modeling | GNS-Modeling-master/gns/inference/optimization/tune.py | import numpy as np
import torch
def get_param_grid(pblur, peps, nbins_blur, nbins_eps):
blur_grid = np.linspace(-50, pblur, nbins_blur)
eps_grid = np.linspace(-2, 1, nbins_eps-1)
eps_grid = np.append(eps_grid, peps)
param_grid = np.meshgrid(blur_grid, eps_grid)
param_grid = np.stack(param_grid, a... | 1,624 | 46.794118 | 90 | py |
GNS-Modeling | GNS-Modeling-master/gns/inference/optimization/parse.py | import torch
import torch.nn as nn
from pybpl.splines import get_stk_from_bspline
from pybpl.util.affine import apply_warp
from . import soft_constraints as C
def param_state(p):
return p.detach().cpu().clone()
class Parse(nn.Module):
def __init__(self, init_parse, init_blur=16., init_epsilon=0.5, bound_m... | 3,607 | 32.407407 | 87 | py |
GNS-Modeling | GNS-Modeling-master/experiments/classification/get_base_parses.py | import argparse
import os
import time
import torch
from pybpl.util import nested_map
from pybpl.splines import get_stk_from_bspline
from gns.inference.parsing import get_topK_parses
from gns.omniglot.classification import ClassificationDataset
from gns.type import TypeModel
from gns.utils.experiments import mkdir, time... | 2,925 | 34.253012 | 75 | py |
GNS-Modeling | GNS-Modeling-master/experiments/classification/refit_parses_single.py | import argparse
import os
import time
import torch
from gns.rendering import Renderer
from gns.token import TokenModel
from gns.inference import optimization as opt
from gns.omniglot.classification import ClassificationDataset
from gns.utils.experiments import mkdir, time_string
def config_for_refit(parse):
pars... | 5,095 | 35.4 | 95 | py |
GNS-Modeling | GNS-Modeling-master/experiments/classification/optimize_parses.py | import argparse
import os
import time
import torch
from gns.rendering import Renderer
from gns.token import TokenModel
from gns.type import TypeModel
from gns.inference import optimization as opt
from gns.omniglot.classification import ClassificationDataset
from gns.utils.experiments import mkdir, time_string
def lo... | 4,603 | 34.415385 | 88 | py |
GNS-Modeling | GNS-Modeling-master/experiments/generate_concepts/generate.py | from tqdm.notebook import tqdm
import matplotlib.pylab as plt
import torch
from pybpl.splines import get_stk_from_bspline
from gns.type import TypeModel
from gns.utils import render_strokes
from gns.viz import plot_image
MEAN = torch.tensor([50., -50.])
SCALE = torch.tensor([20., 20.])
def convert_space(x, inverse=... | 2,233 | 28.394737 | 91 | py |
GNS-Modeling | GNS-Modeling-master/experiments/generate_exemplars/generate.py | import os
import matplotlib.pyplot as plt
import matplotlib.gridspec as gridspec
import numpy as np
import torch
import torch.distributions as dist
from pybpl.splines import get_stk_from_bspline
from gns.rendering import Renderer
from gns.token import TokenModel
from gns.inference import optimization as opt
from gns.v... | 3,749 | 29.737705 | 87 | py |
GNS-Modeling | GNS-Modeling-master/experiments/generate_exemplars/get_parses.py | import argparse
import os
import time
import numpy as np
from imageio import imread
import torch
from pybpl.util import nested_map
from pybpl.splines import get_stk_from_bspline
from gns.inference.parsing import get_topK_parses
from gns.rendering import Renderer
from gns.token import TokenModel
from gns.type import Typ... | 5,670 | 32.755952 | 86 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/gui/ui_model.py | from gui.ui.ui_window import Ui_MainWindow
from gui.ui_draw import *
from PIL import Image, ImageQt
import random, io, os
import numpy as np
import torch
import torchvision.transforms as transforms
from util import task, util
from dataloader.image_folder import make_dataset
from model import create_model
from util.visu... | 18,116 | 38.643326 | 212 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/options/base_options.py | import argparse
import os
import torch
import model
from util import util
class BaseOptions():
def __init__(self):
self.parser = argparse.ArgumentParser()
self.initialized = False
def initialize(self, parser):
# base define
parser.add_argument('--name', type=str, default='expe... | 4,701 | 44.211538 | 160 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/evaluations/inception_score_torch.py | import torch
from torch import nn
from torch.autograd import Variable
from torch.nn import functional as F
import torch.utils.data
import torchvision.transforms as transforms
from torchvision.models.inception import inception_v3
import numpy as np
import math
from PIL import Image
from dataloader.image_folder import m... | 3,468 | 32.355769 | 165 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/evaluations/fid_score_torch.py | #!/usr/bin/env python3
"""Calculates the Frechet Inception Distance (FID) to evalulate GANs
The FID metric calculates the distance between two distributions of images.
Typically, we have summary statistics (mean & covariance matrix) of one
of these distributions, while the 2nd distribution is given by a GAN.
When run a... | 9,605 | 38.368852 | 79 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/util/task.py | import torch
import torch.nn.functional as F
import torchvision.transforms as transforms
from random import randint
import numpy as np
import cv2
from PIL import Image
import random
###################################################################
# random mask generation
############################################... | 3,531 | 29.713043 | 83 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/util/util.py | import numpy as np
import os
import imageio
import math
import torch
# convert a tensor into a numpy array
def tensor2im(image_tensor, bytes=255.0, imtype=np.uint8):
if image_tensor.dim() == 3:
image_numpy = image_tensor.cpu().float().numpy()
else:
image_numpy = image_tensor[0].cpu().float().n... | 1,147 | 23.956522 | 85 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/model/base_model.py | import os, ntpath
import torch
from collections import OrderedDict
from util import util
from . import base_function
class BaseModel():
def __init__(self, opt):
self.opt = opt
self.gpu_ids = opt.gpu_ids
self.isTrain = opt.isTrain
self.save_dir = os.path.join(opt.checkpoints_dir, op... | 6,411 | 39.582278 | 117 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/model/pluralistic_model.py | import torch
from .base_model import BaseModel
from . import network, base_function, external_function
from util import task
import itertools
class Pluralistic(BaseModel):
"""This class implements the pluralistic image completion, for 256*256 resolution image inpainting"""
def name(self):
return "Plur... | 11,583 | 48.293617 | 136 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/model/network.py | from .base_function import *
from .external_function import SpectralNorm
import torch.nn.functional as F
##############################################################################################################
# Network function
###################################################################################... | 13,524 | 40.360856 | 141 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/model/base_function.py | import torch
import torch.nn as nn
from torch.nn import init
import functools
from torch.optim import lr_scheduler
from .external_function import SpectralNorm
######################################################################################
# base function for network structure
###################################... | 13,230 | 36.481586 | 141 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/model/external_function.py | import torch
from torch import nn
from torch.nn import Parameter
import torch.nn.functional as F
import copy
####################################################################################################
# spectral normalization layer to decouple the magnitude of a weight tensor
################################... | 9,261 | 39.269565 | 132 | py |
Pluralistic-Inpainting | Pluralistic-Inpainting-master/dataloader/data_loader.py | from PIL import Image, ImageFile
import torchvision.transforms as transforms
import torch.utils.data as data
from .image_folder import make_dataset
from util import task
import random
class CreateDataset(data.Dataset):
def __init__(self, opt):
self.opt = opt
self.img_paths, self.img_size = make_da... | 3,932 | 35.757009 | 124 | py |
multinav2 | multinav2-main/multinav/algorithms/policy_net.py | """Policies for agents with composite observations: environment + automaton states."""
from typing import Any, Callable, Dict, Optional, Sequence
import numpy as np
import tensorflow as tf
from gym import spaces
from ray.rllib.models.tf.tf_modelv2 import TFModelV2
from tensorflow import keras, summary
from tensorflow... | 8,393 | 30.675472 | 94 | py |
ARC-Solvers | ARC-Solvers-main/arc_solvers/nn/util.py | import torch
from allennlp.common.checks import ConfigurationError
from allennlp.nn.util import replace_masked_values
from allennlp.nn.util import get_text_field_mask
import allennlp
from typing import Union, Dict
import torch
from allennlp.modules import MatrixAttention, Seq2SeqEncoder
def masked_mean(tensor, dim, m... | 7,076 | 44.954545 | 145 | py |
ARC-Solvers | ARC-Solvers-main/arc_solvers/modules/single_time_distributed.py | """
=====================================================================
Decomposable Graph Entailment Model code replicated from SciTail repo
https://github.com/allenai/scitail
=====================================================================
"""
import torch
class SingleTimeDistributed(torch.nn.Module):
"... | 3,268 | 49.292308 | 98 | py |
ARC-Solvers | ARC-Solvers-main/arc_solvers/models/qa/multi_choice/qa_multi_choice_max_att.py | from allennlp.modules.matrix_attention import MatrixAttention
from typing import Dict, Optional, AnyStr, List, Any
import torch
from allennlp.common import Params
from allennlp.common.checks import ConfigurationError
from allennlp.data import Vocabulary
from allennlp.models.model import Model
from allennlp.modules imp... | 12,504 | 48.039216 | 163 | py |
ARC-Solvers | ARC-Solvers-main/arc_solvers/models/entailment/tree_attention.py | """
=====================================================================
Decomposable Graph Entailment Model code replicated from SciTail repo
https://github.com/allenai/scitail
=====================================================================
"""
from typing import Dict, List, Any, Tuple
import numpy
import tor... | 16,217 | 50.814696 | 99 | py |
bert-japanese | bert-japanese-main/convert_tf2_ckpt_for_all_frameworks.py | # Copyright 2023 Masatoshi Suzuki (@singletongue)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or a... | 12,201 | 41.964789 | 114 | py |
understanding-momentum | understanding-momentum-master/resnet_on_cifar/main.py | from __future__ import print_function
import argparse
import os
import datetime
import sys
import time
import torch
import torch.nn.functional as F
import torch.backends.cudnn as cudnn
from tensorboardX import SummaryWriter
from torchvision import datasets, transforms
from shb import SHB
from utils import get_git_di... | 10,043 | 33.754325 | 81 | py |
understanding-momentum | understanding-momentum-master/resnet_on_cifar/shb.py | import torch
from torch.optim import Optimizer
class SHB(Optimizer):
def __init__(self, params, lr, momentum=0, weight_decay=0):
if lr < 0.0:
raise ValueError("Invalid learning rate: {}".format(lr))
if momentum < 0.0:
raise ValueError("Invalid momentum value: {}".format(momentum))
if weight_... | 1,443 | 28.469388 | 77 | py |
understanding-momentum | understanding-momentum-master/resnet_on_cifar/models.py | '''
resnet for cifar in pytorch
Reference:
[1] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR, 2016.
[2] K. He, X. Zhang, S. Ren, and J. Sun. Identity mappings in deep residual networks. In ECCV, 2016.
'''
import torch
import torch.nn as nn
import math
import torch.nn.functi... | 9,557 | 28.140244 | 109 | py |
understanding-momentum | understanding-momentum-master/lr_on_mnist/main.py | from __future__ import print_function
import argparse
import os
import datetime
import sys
import torch
import torch.nn as nn
import torch.nn.functional as F
from tensorboardX import SummaryWriter
from torchvision import datasets, transforms
from qhm import QHM
from utils import get_git_diff, get_git_hash, Logger
... | 8,417 | 33.929461 | 85 | py |
understanding-momentum | understanding-momentum-master/lr_on_mnist/qhm.py | # Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
import torch
from torch.optim.optimizer import Optimizer, required
import param_conv
class QHM(Optimizer):
r"""Implements the quasi-hyp... | 11,422 | 40.689781 | 120 | py |
access | access-main/access/preprocess.py | # Copyright (c) Facebook, Inc. and its affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
#
from functools import wraps
import multiprocessing
import random
import re
from joblib import Parallel, delayed
import t... | 5,957 | 34.676647 | 116 | py |
ANNA-PALM | ANNA-PALM-master/AnetLib/options/base_options.py | import argparse
import os
class BaseOptions():
def __init__(self):
self.parser = argparse.ArgumentParser()
self.initialized = False
self.isTrain = False
def initialize(self):
self.parser.add_argument('--workdir', type=str, required=True, help='work directory')
self.pars... | 8,793 | 76.140351 | 228 | py |
ANNA-PALM | ANNA-PALM-master/AnetLib/data/folder_dataset.py | ################################################################################
# Code from
# https://github.com/pytorch/vision/blob/master/torchvision/datasets/folder.py
# Modified the original code so that it also loads images from the current
# directory as well as the subdirectories
###############################... | 6,594 | 36.050562 | 106 | py |
ANNA-PALM | ANNA-PALM-master/AnetLib/data/image_folder.py | ###############################################################################
# Code from
# https://github.com/pytorch/vision/blob/master/torchvision/datasets/folder.py
# Modified the original code so that it also loads images from the current
# directory as well as the subdirectories
################################... | 1,900 | 27.80303 | 79 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/import_task.py | '''
Imports data and models for the five tasks of the paper (cfr. Table 1).
'''
import os
from experiment_utils import import_cifar
from rotation_rate_utils import get_kernel_layer_names
from models import VGG, resnet_v1, VGG_pytorchBlogStyle, WideResNet
def import_task(experiment, mode = ''):
if experiment not i... | 6,060 | 37.852564 | 114 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/rotation_rate_utils.py | '''
Utilities related to layer-wise angle deviation curves
'''
import numpy as np
from scipy.spatial.distance import cosine
import matplotlib
import matplotlib.pyplot as plt
from keras.callbacks import Callback
from keras.engine.training import _make_batches, _slice_arrays
import keras.backend as K
from keras.losses... | 3,473 | 35.1875 | 130 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/models.py | '''
Functions to create the 5 models used in the paper.
'''
import os
import numpy as np
import keras
import keras.backend as K
from keras.models import Model
from keras.layers import Input, Dense, BatchNormalization, Activation, Dropout, Flatten, add
from keras.layers import Conv2D, ZeroPadding2D, AveragePooling2D,... | 17,068 | 41.779449 | 141 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/experiment_utils.py | import warnings
import numpy as np
import matplotlib
import matplotlib.pyplot as plt
from keras.datasets import cifar10, cifar100
from keras.callbacks import History
def one_hot_encoding(labels,nb_labels = None):
n = len(labels)
if not nb_labels:
nb_labels = labels.max() + 1
if nb_labels... | 5,083 | 31.589744 | 117 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/layca_optimizers.py | '''
Code for applying Layca on SGD, Adam, RMSprop and Adagrad.
Source: keras' implementation of the original optimization methods.
'''
from keras.optimizers import Optimizer
import keras.backend as K
from keras.legacy import interfaces
import tensorflow as tf
from keras.optimizers import Optimizer
import keras.backe... | 14,697 | 39.827778 | 158 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/get_training_utils.py | '''
Get training parameters such as learning rate schedules and stopping criteria for training of the five tasks
'''
import numpy as np
from experiment_utils import lr_schedule
from layca_optimizers import SGD,Adam,RMSprop,Adagrad
from keras.callbacks import Callback, LearningRateScheduler
class StoppingCriteria(Cal... | 5,278 | 41.572581 | 137 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/Supplementary Results/training_monitoring.py | '''
Keras callbacks for storing signals of interest (Adam's 2nd raw moment estimation, layer rotation rate per epoch,..) during training.
'''
import sys
sys.path.insert(0, "../")
import numpy as np
from scipy.spatial.distance import cosine
import matplotlib
import matplotlib.pyplot as plt
from keras.callbacks impor... | 9,381 | 41.645455 | 136 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/Supplementary Results/LARS.py | from keras.optimizers import Optimizer
import keras.backend as K
from keras.legacy import interfaces
import numpy as np
import tensorflow as tf
class LARS(Optimizer):
"""
LARS optimizer (ref: https://arxiv.org/abs/1708.03888)
coded by modifying Keras' SGD optimizer code
"""
def ... | 3,482 | 37.274725 | 153 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/MNIST toy example/training_memory.py | '''
Keras callbacks for storing input and feedbacks of layers during training
'''
import numpy as np
from numpy import random
from keras.callbacks import Callback
import keras.backend as K
from keras.models import Model
def create_gradient_function(model, tensors, loss, add_batchaxis = False):
'''
create gra... | 8,340 | 39.100962 | 183 | py |
An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization | An-experimental-study-of-layer-level-training-speed-and-its-impact-on-generalization-master/MNIST toy example/mnist_experiment_utils.py | import os
import numpy as np
import keras
from keras.datasets import mnist
from keras.layers import Input, Dense
from keras.models import Model
def get_split(nb,x,y):
'''
divides data in two sets, one with nb samples per class, the other contains the rest
'''
train_selection = set()
test_selection... | 3,220 | 35.191011 | 111 | py |
id_prediction_intervals | id_prediction_intervals-main/important_directions/imp_dirs.py | import numpy as np
import scipy
import scipy.stats
import torch
from tqdm.auto import tqdm
from .datasets.regression_datamodule import safe_numpy
def flatten_grad(g):
return torch.cat([p.view(-1) for p in g])
def numel(m: torch.nn.Module, only_trainable: bool = True):
"""
returns the total number of p... | 8,192 | 36.240909 | 120 | py |
id_prediction_intervals | id_prediction_intervals-main/important_directions/method_ens.py | import logging
import os
import numpy as np
import scipy
import scipy.stats
import torch.cuda
import pytorch_lightning as pl
from pytorch_lightning.callbacks import EarlyStopping
from pytorch_lightning import loggers as pl_loggers
from important_directions.datasets.regression_datamodule import RegressionDataModule,... | 4,285 | 36.596491 | 115 | py |
id_prediction_intervals | id_prediction_intervals-main/important_directions/method_dropout.py | import logging
import os
import numpy as np
import pytorch_lightning as pl
import scipy
import torch
from pytorch_lightning import loggers as pl_loggers
from scipy.special import logsumexp
from tqdm.auto import tqdm
from .datasets.regression_datamodule import RegressionDataModule, safe_numpy
from .network import Regr... | 6,222 | 37.652174 | 113 | py |
id_prediction_intervals | id_prediction_intervals-main/important_directions/experiment.py | import argparse
import os
import numpy as np
import pandas as pd
import torch.multiprocessing as mp
import yaml
from scipy.stats.stats import pearsonr
from sklearn.metrics import mean_squared_error, r2_score
from important_directions.datasets.regression_datamodule import RegressionDataModule, safe_numpy
from importan... | 4,617 | 38.135593 | 103 | py |
id_prediction_intervals | id_prediction_intervals-main/important_directions/network.py | import numpy as np
import torch
from torch import nn
import torch.nn.functional as F
from tqdm.auto import tqdm
import pytorch_lightning as pl
from torchmetrics.regression import ExplainedVariance, MeanSquaredError
from .datasets.regression_datamodule import safe_numpy
class RegressionNet(pl.LightningModule):
... | 3,955 | 30.903226 | 102 | py |
id_prediction_intervals | id_prediction_intervals-main/important_directions/method_id.py | import os
import pytorch_lightning as pl
from pytorch_lightning import loggers as pl_loggers
from pytorch_lightning.callbacks import EarlyStopping
from important_directions.datasets.regression_datamodule import RegressionDataModule, safe_numpy
from important_directions.imp_dirs import ImportantDirectionsPytorch
from ... | 3,620 | 47.28 | 111 | py |
id_prediction_intervals | id_prediction_intervals-main/important_directions/method_id_exact.py | import os
import numpy as np
import scipy
import scipy.stats
from tqdm.auto import tqdm
import pytorch_lightning as pl
import torch
from pytorch_lightning import loggers as pl_loggers
from pytorch_lightning.callbacks import EarlyStopping
from important_directions.datasets.regression_datamodule import RegressionData... | 3,811 | 35.653846 | 107 | py |
id_prediction_intervals | id_prediction_intervals-main/important_directions/datasets/dataloader.py | import numpy as np
import torch
from torch.utils.data import DataLoader
class BatchXYDataLoader:
def __init__(self, X, y, batch_size=128, shuffle=False):
self.X = X
self.y = y
self.batch_size = batch_size
self.shuffle = shuffle
self.n = self.X.shape[0]
self.len =... | 856 | 23.485714 | 60 | py |
id_prediction_intervals | id_prediction_intervals-main/important_directions/datasets/regression_datamodule.py | import numbers
import numpy as np
import torch
from torch.utils.data import TensorDataset, DataLoader
from .dataloader import BatchXYDataLoader
def max_k_n_ratio(k, n, ratio):
return max(k, int(n * ratio))
def get_index_range(a, idx, start, stop):
return a[idx[start:stop]]
def scale_inplace(t, mean, std... | 5,073 | 30.515528 | 120 | py |
DOCRED-FE | DOCRED-FE-main/code/REBEL/src/test.py | import omegaconf
import hydra
import pytorch_lightning as pl
from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint
from pl_data_modules import BasePLDataModule
from pl_modules import BasePLModule
from transformers import AutoConfig, AutoModelForSeq2SeqLM, AutoTokenizer
from pytorch_lightning.loggers... | 6,098 | 50.252101 | 1,961 | py |
DOCRED-FE | DOCRED-FE-main/code/REBEL/src/model_saving.py | from pl_modules import BasePLModule
from transformers import AutoConfig, AutoModelForSeq2SeqLM, AutoTokenizer
import torch
import omegaconf
config = AutoConfig.from_pretrained(
'facebook/bart-large',
decoder_start_token_id = 0,
early_stopping = False,
no_repeat_ngram_size = 0,
)
tokenizer = AutoTokeniz... | 1,014 | 34 | 172 | py |
DOCRED-FE | DOCRED-FE-main/code/REBEL/src/pl_modules.py | from typing import Any
import nltk
import json
import pytorch_lightning as pl
import torch
import numpy as np
import pandas as pd
from score import score, re_score
from transformers import AutoConfig, AutoModelForSeq2SeqLM, AutoTokenizer
from transformers.optimization import (
Adafactor,
AdamW,
get_constant... | 37,880 | 65.225524 | 3,629 | py |
DOCRED-FE | DOCRED-FE-main/code/REBEL/src/utils.py |
import math
from typing import Callable, Dict, Iterable, List, Tuple, Union
import numpy as np
import torch
import torch.distributed as dist
from torch import nn
from torch import Tensor
from torch.nn.parameter import Parameter
from transformers.models.bart.modeling_bart import shift_tokens_right
def label_smoothed... | 11,000 | 40.513208 | 165 | py |
DOCRED-FE | DOCRED-FE-main/code/REBEL/src/generate_samples.py | from typing import Sequence
import torch
from pytorch_lightning import Callback, LightningModule, Trainer
from torch import nn
import numpy as np
import pandas as pd
from torch.nn.utils.rnn import pad_sequence
import wandb
class GenerateTextSamplesCallback(Callback): # pragma: no cover
"""
PL Callback to gene... | 3,357 | 42.051282 | 130 | py |
DOCRED-FE | DOCRED-FE-main/code/REBEL/src/scheduler.py |
from torch.optim.lr_scheduler import LambdaLR
def get_inverse_square_root_schedule_with_warmup(
optimizer, num_warmup_steps, warmup_init_lr=-1, last_epoch=-1
):
"""
Create a schedule with a learning rate that decreases as a polynomial decay from the initial lr set in the
optimizer to end lr defined by... | 1,873 | 35.038462 | 115 | py |
DOCRED-FE | DOCRED-FE-main/code/REBEL/src/modeling_bart.py | # coding=utf-8
# Copyright 2021 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/... | 85,205 | 43.916183 | 239 | py |
DOCRED-FE | DOCRED-FE-main/code/REBEL/src/pl_data_modules.py | from pyexpat import model
from typing import Any, Union, List, Optional
from omegaconf import DictConfig
import torch
from torch.utils.data import DataLoader
import pytorch_lightning as pl
from datasets import load_dataset, set_caching_enabled
from transformers import (
AutoConfig,
AutoModelForSeq2SeqLM,
... | 9,058 | 45.219388 | 185 | py |
DOCRED-FE | DOCRED-FE-main/code/REBEL/src/train.py | import omegaconf
import hydra
import pytorch_lightning as pl
from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint
from pl_data_modules import BasePLDataModule
from pl_modules import BasePLModule
from transformers import AutoConfig, AutoModelForSeq2SeqLM, AutoTokenizer
from pytorch_lightning.loggers... | 7,637 | 47.037736 | 1,961 | py |
DOCRED-FE | DOCRED-FE-main/code/LSTM/config/Config.py | # coding: utf-8
import torch
import torch.nn as nn
from torch.autograd import Variable
import torch.optim as optim
import numpy as np
import os
import time
import datetime
import json
import sys
import sklearn.metrics
from tqdm import tqdm
import matplotlib
matplotlib.use('Agg')
import matplotlib.pyplot as plt
import r... | 25,660 | 31.07625 | 271 | py |
DOCRED-FE | DOCRED-FE-main/code/LSTM/config/EviConfig.py | # coding: utf-8
import torch
import torch.nn as nn
from torch.autograd import Variable
import torch.optim as optim
import numpy as np
import os
import time
import datetime
import json
import sys
import sklearn.metrics
from tqdm import tqdm
import matplotlib
matplotlib.use('Agg')
import matplotlib.pyplot as plt
import r... | 18,017 | 29.435811 | 147 | py |
DOCRED-FE | DOCRED-FE-main/code/LSTM/models/BiLSTM.py | import torch
import torch.autograd as autograd
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
from torch.autograd import Variable
from torch import nn
import numpy as np
import math
from torch.nn import init
from torch.nn.utils import rnn
class BiLSTM(nn.Module):
def __init__(self,... | 11,205 | 36.604027 | 132 | py |
DOCRED-FE | DOCRED-FE-main/code/LSTM/models/CNN3.py | import torch
import torch.autograd as autograd
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
from torch.autograd import Variable
class CNN3(nn.Module):
def __init__(self, config):
super(CNN3, self).__init__()
self.config = config
self.word_emb = nn.Embedding(config.data_word_... | 3,178 | 37.768293 | 143 | py |
DOCRED-FE | DOCRED-FE-main/code/LSTM/models/LSTM.py | import torch
import torch.autograd as autograd
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
from torch.autograd import Variable
from torch import nn
import numpy as np
import math
from torch.nn import init
from torch.nn.utils import rnn
class LSTM(nn.Module):
def __init__(self, c... | 8,656 | 34.048583 | 132 | py |
DOCRED-FE | DOCRED-FE-main/code/GAIN/code/test.py | import sklearn.metrics
import torch
from config import *
from data import DGLREDataset, DGLREDataloader, BERTDGLREDataset
from models.GAIN import GAIN_GloVe, GAIN_BERT
from utils import get_cuda, logging, print_params
# for ablation
# from models.GCNRE_nomention import GAIN_GloVe, GAIN_BERT
def test(model, dataloa... | 8,393 | 34.719149 | 136 | py |
DOCRED-FE | DOCRED-FE-main/code/GAIN/code/evaluate.py | import sklearn.metrics
import torch
from config import *
from data import DGLREDataset, DGLREDataloader, BERTDGLREDataset
from models.GAIN import GAIN_GloVe, GAIN_BERT
from utils import get_cuda, logging, print_params
# for ablation
# from models.GCNRE_nomention import GAIN_GloVe, GAIN_BERT
def eval(model, dataloa... | 8,545 | 36.156522 | 165 | py |
DOCRED-FE | DOCRED-FE-main/code/GAIN/code/utils.py | from datetime import datetime
import numpy as np
import torch
# 返回一个tensor的cuda版本
def get_cuda(tensor):
if torch.cuda.is_available():
return tensor.cuda()
# return tensor # 宁愿报错,也拒绝返回CPU
def logging(s):
print(datetime.now(), s)
class Accuracy(object):
def __init__(self):
self.corre... | 884 | 20.071429 | 109 | py |
DOCRED-FE | DOCRED-FE-main/code/GAIN/code/data.py | import json
import math
import os
import pickle
import random
from collections import defaultdict
import dgl
import numpy as np
import torch
from torch.utils.data import IterableDataset, DataLoader
from tqdm import tqdm
from transformers import *
from models.GAIN import Bert
from utils import get_cuda
IGNORE_INDEX =... | 37,359 | 47.082368 | 186 | py |
DOCRED-FE | DOCRED-FE-main/code/GAIN/code/train.py | import time
import matplotlib
import matplotlib.pyplot as plt
import torch
from torch import nn
from torch import optim
from tqdm import tqdm
from config import *
from data import DGLREDataset, DGLREDataloader, BERTDGLREDataset # 这是从data.py 中引入几个类
from models.GAIN import GAIN_GloVe, GAIN_BERT
from test import test
f... | 9,725 | 37.595238 | 154 | py |
DOCRED-FE | DOCRED-FE-main/code/GAIN/code/models/GAIN.py | import dgl
import dgl.nn.pytorch as dglnn
import numpy as np
import torch
import torch.nn as nn
from transformers import *
from utils import get_cuda
class GAIN_GloVe(nn.Module):
def __init__(self, config):
super(GAIN_GloVe, self).__init__()
self.config = config
word_emb_size = config.... | 41,175 | 47.78673 | 180 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex_train.py | import hydra
from hydra.core.config_store import ConfigStore
from omegaconf import OmegaConf
from torch import seed
import warnings
warnings.filterwarnings("ignore")
from configs import TrainConfig
from jerex import model, util
cs = ConfigStore.instance()
cs.store(name="train", node=TrainConfig)
def seed_everything(... | 1,039 | 25.666667 | 97 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/loss.py | from abc import ABC
import torch
class Loss(ABC):
def compute(self, *args, **kwargs):
pass
class JointLoss(Loss):
def __init__(self, task_weights=None):
self._mention_criterion = torch.nn.BCEWithLogitsLoss(reduction='none')
self._entity_criterion = torch.nn.CrossEntropyLoss(reductio... | 5,942 | 33.754386 | 103 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/model.py | import os
import pickle
import warnings
from multiprocessing import Lock
import pytorch_lightning as pl
import torch
import transformers
from pytorch_lightning import loggers
from pytorch_lightning.callbacks import ModelCheckpoint, LearningRateMonitor
from transformers import AdamW, BertConfig, BertTokenizer
from con... | 20,682 | 51.628499 | 119 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/datasets.py | import json
from collections import OrderedDict
import torch
from typing import List
from torch.utils.data import Dataset as TorchDataset
from tqdm import tqdm
from jerex import util
from jerex.entities import Relation, Document, Entity, Token, Sentence, EntityMention
from jerex.task_types import TaskType
from jerex... | 10,283 | 37.373134 | 117 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/util.py | import os
import torch
from hydra.utils import to_absolute_path
from jerex.entities import TokenSpan
def create_directories_file(f):
d = os.path.dirname(f)
if d and not os.path.exists(d):
os.makedirs(d)
return f
def create_directories_dir(d):
if d and not os.path.exists(d):
os.ma... | 3,134 | 24.696721 | 104 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/data_module.py | import json
from collections import OrderedDict
import pytorch_lightning as pl
from torch.utils.data import DataLoader
from transformers import BertTokenizer
from jerex.datasets import DocREDDataset
from jerex.entities import EntityType, RelationType
from jerex.sampling.sampling_common import collate_fn_padding
cla... | 6,985 | 43.782051 | 113 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/evaluation/conversion.py | import torch
def convert_gt_cluster(entity, include_entity_type=False):
t = set([m.orig_span for m in entity.entity_mentions])
if include_entity_type:
t = (t, entity.entity_type)
return t
def convert_gt_relation(relation, include_entity_type=False):
head = convert_gt_cluster(relation.head_... | 3,258 | 31.59 | 110 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/evaluation/joint_evaluator.py | import json
import os
import jinja2
import torch
from typing import List, Tuple, Dict
import jerex.evaluation.scoring
from jerex import util
from jerex.entities import Document
from jerex.evaluation import conversion, scoring
from jerex.evaluation.evaluator import Evaluator
SCRIPT_PATH = os.path.dirname(os.path.real... | 12,679 | 42.129252 | 121 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/classification_models.py | import torch
from transformers import BertConfig
from transformers import BertModel
from transformers import BertPreTrainedModel
from jerex import util
from jerex.evaluation.classification_evaluator import MentionLocalizationEvaluator, RelClassificationEvaluator, \
EntityClassificationEvaluator, CoreferenceResolut... | 13,412 | 48.677778 | 155 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/misc.py | import torch
from sklearn.cluster import AgglomerativeClustering
from jerex import util
from jerex.sampling import sampling_common
def create_coref_mention_pairs(valid_mentions, mention_spans, encodings, tokenizer):
""" Creates pairs of of mentions associated edit distances and sample masks for coreference resol... | 12,736 | 47.064151 | 113 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/joint_models.py | from abc import abstractmethod
import torch
from transformers import BertConfig, BertTokenizer
from transformers import BertModel
from transformers import BertPreTrainedModel
from jerex import util
from jerex.evaluation.joint_evaluator import JointEvaluator
from jerex.loss import JointLoss
from jerex.task_types impor... | 19,226 | 59.462264 | 155 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/__init__.py | import torch
from torch import nn
from transformers import BertPreTrainedModel, BertConfig, BertTokenizer
from jerex.models.classification_models import RelClassificationMultiInstanceModel, \
EntityClassificationModel, CoreferenceResolutionModel, MentionLocalizationModel, RelClassificationGlobal
from jerex.models.... | 3,443 | 46.833333 | 138 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/modules/relation_classification_global.py | from torch import nn as nn
class RelationClassificationGlobal(nn.Module):
def __init__(self, hidden_size, relation_types):
super().__init__()
self.rel_classifier = nn.Linear(hidden_size, relation_types)
def forward(self, entity_pair_reprs):
rel_clf = self.rel_classifier(entity_pair_r... | 350 | 24.071429 | 68 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/modules/mention_localization.py | import torch
from torch import nn as nn
class MentionLocalization(nn.Module):
def __init__(self, hidden_size, meta_embedding_size, size_embeddings_count, prop_drop):
super().__init__()
self.linear = nn.Linear(hidden_size + meta_embedding_size, hidden_size)
self.mention_classifier = nn.Lin... | 915 | 38.826087 | 91 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/modules/relation_classification_multi_instance.py | import torch
from torch import nn as nn
from jerex import util
import numpy as np
class RelationClassificationMultiInstance(nn.Module):
def __init__(self, hidden_size, entity_types, relation_types, meta_embedding_size,
token_dist_embeddings_count, sentence_dist_embeddings_count, prop_drop, use_f... | 8,504 | 54.227273 | 154 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/modules/entity_pair_representation.py | import torch
from torch import nn as nn
from jerex import util
class EntityPairRepresentation(nn.Module):
def __init__(self, hidden_size, entity_types, meta_embedding_size, prop_drop):
super().__init__()
self.entity_pair_linear = nn.Linear(hidden_size * 2 + meta_embedding_size * 2, hidden_size)
... | 1,437 | 34.073171 | 106 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/modules/entity_classification.py | import torch
from torch import nn as nn
class EntityClassification(nn.Module):
def __init__(self, hidden_size, entity_types, prop_drop):
super().__init__()
self.linear = nn.Linear(hidden_size, hidden_size)
self.entity_classifier = nn.Linear(hidden_size, entity_types)
self.dropout ... | 542 | 29.166667 | 74 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/modules/entity_representation.py | from torch import nn as nn
from jerex import util
class EntityRepresentation(nn.Module):
def __init__(self, prop_drop):
super().__init__()
self.dropout = nn.Dropout(prop_drop)
def forward(self, mention_reprs, entities, entity_masks):
mention_clusters = util.batch_index(mention_reprs... | 652 | 28.681818 | 68 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/modules/relation_classification_multi_instance copy.py | import torch
from torch import nn as nn
from jerex import util
import numpy as np
class RelationClassificationMultiInstance(nn.Module):
def __init__(self, hidden_size, entity_types, relation_types, meta_embedding_size,
token_dist_embeddings_count, sentence_dist_embeddings_count, prop_drop):
... | 7,586 | 53.582734 | 150 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/modules/mention_representation.py | import torch
from torch import nn as nn
class MentionRepresentation(nn.Module):
def __init__(self):
super().__init__()
def forward(self, h, mention_masks, max_spans=None):
mention_reprs = torch.zeros([mention_masks.shape[0], mention_masks.shape[1],
h.shape... | 1,097 | 33.3125 | 84 | py |
DOCRED-FE | DOCRED-FE-main/code/JEREX/jerex/models/modules/coreference_resolution.py | import torch
from torch import nn as nn
from jerex import util
class CoreferenceResolution(nn.Module):
def __init__(self, hidden_size, meta_embedding_size, ed_embeddings_count, prop_drop):
super().__init__()
self.coref_linear = nn.Linear(hidden_size * 2 + meta_embedding_size, hidden_size)
... | 2,196 | 39.685185 | 97 | py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.