code
stringlengths
42
43.2k
apis
list
extract_api
stringlengths
115
61.9k
from pathlib import Path import sys sys.path.append(str(Path().absolute())) import logging log_level = "INFO" logging.basicConfig( filename=str(snakemake.log), filemode="w", level=log_level, format="[%(asctime)s]:%(levelname)s: %(message)s", datefmt="%d/%m/%Y %I:%M:%S %p", ) from evaluate.calculator...
[ "evaluate.calculator.RecallCalculator", "evaluate.report.RecallReport.from_files" ]
[((918, 981), 'pathlib.Path', 'Path', (['snakemake.output.recall_file_for_one_sample_vs_nb_samples'], {}), '(snakemake.output.recall_file_for_one_sample_vs_nb_samples)\n', (922, 981), False, 'from pathlib import Path\n'), ((996, 1027), 'logging.info', 'logging.info', (['f"""Loading report"""'], {}), "(f'Loading report'...
import argparse import os import random import shutil import time import warnings import pickle from collections import OrderedDict from datetime import datetime import numpy as np import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.parallel import torch.backends.cudnn as cudnn import to...
[ "evaluate.Evaluator", "evaluate.ClassifierGenerator" ]
[((901, 965), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""PyTorch ImageNet Training"""'}), "(description='PyTorch ImageNet Training')\n", (924, 965), False, 'import argparse\n'), ((7204, 7229), 'torch.cuda.device_count', 'torch.cuda.device_count', ([], {}), '()\n', (7227, 7229), False...
# repo originally forked from https://github.com/Confusezius/Deep-Metric-Learning-Baselines """to do: clean all of the files - particularly the main.py and also the losses and dataset files and the file for doing the dataloading -- fast loading etc need to change all of the copyrights at the top of all of the files...
[ "evaluate.DistanceMeasure", "evaluate.evaluate", "evaluate.GradientMeasure" ]
[((399, 432), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (422, 432), False, 'import warnings\n'), ((576, 597), 'matplotlib.use', 'matplotlib.use', (['"""agg"""'], {}), "('agg')\n", (590, 597), False, 'import os, numpy as np, argparse, random, matplotlib, datetime\n'), ...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Mon Sep 3 14:33:25 2018 @author: iswariya """ import numpy as np import torch import torch.nn as nn import torch.optim as optim import torchvision.models as models from dataloader import SoccerDataset, get_dataloader from evaluate import predict from met...
[ "evaluate.predict" ]
[((890, 1016), 'dataloader.SoccerDataset', 'SoccerDataset', (['PATH', 'image_folder', 'mask_folder', 'downsampled_resolution'], {'transform': "['Horizontal_Flip', 'Brightness_adjust']"}), "(PATH, image_folder, mask_folder, downsampled_resolution,\n transform=['Horizontal_Flip', 'Brightness_adjust'])\n", (903, 1016),...
from argparse import ArgumentParser from pathlib import Path import time import os import gin import numpy as np import torch import torch.nn as nn import torch.optim.lr_scheduler as lr_scheduler from torch.nn import CrossEntropyLoss from datasets import get_dataset from models.gan import get_architecture from torch....
[ "evaluate.classifier.test_classifier", "evaluate.classifier.accuracy", "evaluate.AverageMeter" ]
[((1217, 1244), 'gin.configurable', 'gin.configurable', (['"""options"""'], {}), "('options')\n", (1233, 1244), False, 'import gin\n'), ((699, 762), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'description': '"""Testing script: Linear evaluation"""'}), "(description='Testing script: Linear evaluation')\n", (713,...
import os from generate_libfm_data import generate_libfm_data, tfidf2str from evaluate import scoring if not os.path.exists('news_tfidf.pkl') or not os.path.exists('user_tfidf.pkl'): os.system('python generate_tf_idf_feature_file.py') if not os.path.exists('train.libfm') or not os.path.exists('dev.libfm') or not ...
[ "evaluate.scoring" ]
[((189, 240), 'os.system', 'os.system', (['"""python generate_tf_idf_feature_file.py"""'], {}), "('python generate_tf_idf_feature_file.py')\n", (198, 240), False, 'import os\n'), ((354, 375), 'generate_libfm_data.generate_libfm_data', 'generate_libfm_data', ([], {}), '()\n', (373, 375), False, 'from generate_libfm_data...
''' Usage: quote_detection.py <model-name> -t <corpus-type> -c <corpus-path> -e <embedding-path> Options: -t Corpus type (either parc, rwg, or stop) -c Path to corpus -e Path to embeddings file ''' import random from docopt import docopt import torch import torch.nn as nn impo...
[ "evaluate.report", "evaluate.evaluate" ]
[((15507, 15529), 'random.shuffle', 'random.shuffle', (['zipped'], {}), '(zipped)\n', (15521, 15529), False, 'import random\n'), ((20864, 20934), 'evaluate.evaluate', 'evaluate', (['predicted_processed', 'raw_labels'], {'roles': 'roles', 'mode': 'eval_mode'}), '(predicted_processed, raw_labels, roles=roles, mode=eval_m...
import evaluate evaluate.dump_aggregated_period()
[ "evaluate.dump_aggregated_period" ]
[((17, 50), 'evaluate.dump_aggregated_period', 'evaluate.dump_aggregated_period', ([], {}), '()\n', (48, 50), False, 'import evaluate\n')]
from data_processing import preprocess_28D import autoencoders.autoencoder as ae import autoencoders.variational_autoencoder as vae import autoencoders.sparse_autoencoder as sae from create_plots import plot_initial_data, plot_test_pred_data, correlation_plot from evaluate import evaluate_model import pandas as pd impo...
[ "evaluate.evaluate_model" ]
[((444, 469), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (467, 469), False, 'import argparse\n'), ((1854, 1898), 'data_loader.load_cms_data', 'load_cms_data', ([], {'filename': '"""open_cms_data.root"""'}), "(filename='open_cms_data.root')\n", (1867, 1898), False, 'from data_loader import l...
import os import sys sys.path.append(os.path.join(os.path.dirname(__file__))) import yaml import logging import torch import torch.optim as optim import torch.optim.lr_scheduler as lr_scheduler from torch.cuda import amp from tools.datasets import create_dataloader, preprocess from tqdm import tqdm import math import ...
[ "evaluate.test" ]
[((674, 701), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (691, 701), False, 'import logging\n'), ((1182, 1214), 'os.makedirs', 'os.makedirs', (['wdir'], {'exist_ok': '(True)'}), '(wdir, exist_ok=True)\n', (1193, 1214), False, 'import os\n'), ((1689, 1709), 'tools.general.init_seeds', ...
import json from typing import List from resource_mapping.result_analyzer import ResultAnalyzer from execution_handler.execution_handler import ExecutionHandler from aggregator.aggregator import Aggregator, AggregatorResults from datetime import datetime, date from quantum_execution_job import QuantumExecutionJob from ...
[ "evaluate.circuit_gen.circ_gen", "evaluate.util.sv_to_probability" ]
[((1883, 1897), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (1895, 1897), False, 'from datetime import datetime, date\n'), ((2802, 2822), 'config.load_config.load_or_create', 'cfg.load_or_create', ([], {}), '()\n', (2820, 2822), True, 'import config.load_config as cfg\n'), ((2827, 2867), 'logger.set_log_...
#!/usr/bin/env python3 import argparse import os import time import numpy as np import pandas as pd import torch from torch import optim import callbacks as cb import evaluate import utils import visual_plt from continual_learner import ContinualLearner from data import get_multitask_experiment from encoder import Cl...
[ "evaluate.intial_accuracy", "evaluate.show_reconstruction", "evaluate.show_samples", "evaluate.initiate_metrics_dict" ]
[((580, 682), 'argparse.ArgumentParser', 'argparse.ArgumentParser', (['"""./main.py"""'], {'description': '"""Run individual continual learning experiment."""'}), "('./main.py', description=\n 'Run individual continual learning experiment.')\n", (603, 682), False, 'import argparse\n'), ((11920, 11959), 'torch.device...
import argparse import os import sys import json sys.path.append(os.path.abspath(os.path.join(os.path.dirname(__file__), ".."))) from datasets.ct_dataset import get_dataloaders from evaluate import evaluate from config import * from metrics import VolumeLoss from models import get_model def test(train_data_root, mo...
[ "evaluate.evaluate" ]
[((1011, 1098), 'metrics.VolumeLoss', 'VolumeLoss', (['config.dice_loss_weight', 'config.wce_loss_weight', 'config.ce_loss_weight'], {}), '(config.dice_loss_weight, config.wce_loss_weight, config.\n ce_loss_weight)\n', (1021, 1098), False, 'from metrics import VolumeLoss\n'), ((1226, 1316), 'evaluate.evaluate', 'eva...
# -*- coding: utf-8 -*- import time import numpy as np from evaluate import get_idcg_list, evaluate, evaluate_ranking, evaluate_alpha from clicks import * import StartStatus as ss class SingleSimulation(object): def __init__(self, sim_args, output_queue, click_model, datafold): self.train_only = sim_args.train...
[ "evaluate.evaluate_ranking", "evaluate.get_idcg_list", "evaluate.evaluate_alpha", "evaluate.evaluate" ]
[((765, 869), 'evaluate.get_idcg_list', 'get_idcg_list', (['self.datafold.train_label_vector', 'self.datafold.train_doclist_ranges', 'self.n_results'], {}), '(self.datafold.train_label_vector, self.datafold.\n train_doclist_ranges, self.n_results)\n', (778, 869), False, 'from evaluate import get_idcg_list, evaluate,...
import os import time import json import logging import argparse import sys sys.path.append("libs") import numpy as np import pandas as pd import tensorflow as tf from tensorflow.keras import backend as K from data import CollaborativeVAEDataGenerator from train_vae import get_collabo_vae from eval...
[ "evaluate.EvaluateModel" ]
[((84, 107), 'sys.path.append', 'sys.path.append', (['"""libs"""'], {}), "('libs')\n", (99, 107), False, 'import sys\n'), ((476, 501), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (499, 501), False, 'import argparse\n'), ((1188, 1204), 'tensorflow.ConfigProto', 'tf.ConfigProto', ([], {}), '()...
import operator import numpy as np import tensorflow as tf from data import get_best_span, get_phrase from evaluate import evaluate class Evaluation(object): pass class SquadEvaluation(Evaluation): def __init__(self, data, inputs=None, outputs=None, loss=None, global_step=None): self.inputs = inpu...
[ "evaluate.evaluate" ]
[((3054, 3096), 'evaluate.evaluate', 'evaluate', (["self.data.squad['data']", 'answers'], {}), "(self.data.squad['data'], answers)\n", (3062, 3096), False, 'from evaluate import evaluate\n'), ((2549, 2597), 'data.get_best_span', 'get_best_span', (['logits1', 'logits2'], {'op': 'operator.add'}), '(logits1, logits2, op=o...
"""Train an encoder using Contrastive Learning.""" import argparse import os import subprocess import torch import torch.backends.cudnn as cudnn import torch.nn as nn import torch.optim as optim from optimizers import LARS # from torchlars import LARS from tqdm import tqdm from dataset import get_datasets from crit...
[ "evaluate.test", "evaluate.encode_train_set", "evaluate.train_clf" ]
[((655, 682), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (672, 682), False, 'import logging\n'), ((686, 738), 'hydra.main', 'hydra.main', ([], {'config_path': '"""conf"""', 'config_name': '"""config"""'}), "(config_path='conf', config_name='config')\n", (696, 738), False, 'import hydr...
# from rdkit import Chem # from torch_geometric.data import Data, Batch from train import train from data import ModalRetriever # import torch import os # from train_utils import setup_common # from utils import mkdir # from model.load_model import get # from transformers import BertTokenizer # from transformers impor...
[ "evaluate.evaluate" ]
[((758, 769), 'options.read_args', 'read_args', ([], {}), '()\n', (767, 769), False, 'from options import read_args\n'), ((3010, 3084), 'data.ModalRetriever', 'ModalRetriever', (["(data_dir + 'mention2ent.json')", "(data_dir + 'cmpd_info.json')"], {}), "(data_dir + 'mention2ent.json', data_dir + 'cmpd_info.json')\n", (...
# python train.py anchor-based --model-dir ../models/ab_mobilenet --splits ../splits/tvsum.yml ../splits/summe.yml --max-epoch 50 --cnn mobilenet --base-model attention --num-feature 1280 --num-head 10 # python train.py anchor-based --model-dir ../models/ab_mobilenet_lstm --splits ../splits/tvsum.yml ../splits/summe.y...
[ "evaluate.evaluate" ]
[((770, 789), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (787, 789), False, 'import logging\n'), ((1138, 1292), 'anchor_based.dsnet.DSNet', 'DSNet', ([], {'base_model': 'args.base_model', 'num_feature': 'args.num_feature', 'num_hidden': 'args.num_hidden', 'anchor_scales': 'args.anchor_scales', 'num_hea...
from __future__ import print_function, division import sys sys.path.append('core') import copy from datetime import datetime import argparse import os import time import numpy as np import matplotlib matplotlib.use('Agg') import matplotlib.pyplot as plt import torch import torch.nn as nn import torch.optim as optim i...
[ "evaluate.validate_chairs", "evaluate.validate_kitti", "evaluate.validate_viper", "evaluate.validate_sintel", "evaluate.validate_things" ]
[((59, 82), 'sys.path.append', 'sys.path.append', (['"""core"""'], {}), "('core')\n", (74, 82), False, 'import sys\n'), ((201, 222), 'matplotlib.use', 'matplotlib.use', (['"""Agg"""'], {}), "('Agg')\n", (215, 222), False, 'import matplotlib\n'), ((2289, 2467), 'torch.optim.lr_scheduler.OneCycleLR', 'optim.lr_scheduler....
# python train.py anchor-free --model-dir ../models/af_mobilenet --splits ../splits/tvsum.yml ../splits/summe.yml --max-epoch 50 --cnn mobilenet --base-model attention --num-feature 1280 --num-head 10 --nms-thresh 0.4 # python train.py anchor-free --model-dir ../models/af_default --splits ../splits/tvsum.yml ../spli...
[ "evaluate.evaluate" ]
[((1476, 1495), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (1493, 1495), False, 'import logging\n'), ((1545, 1666), 'anchor_free.dsnet_af.DSNetAF', 'DSNetAF', ([], {'base_model': 'args.base_model', 'num_feature': 'args.num_feature', 'num_hidden': 'args.num_hidden', 'num_head': 'args.num_head'}), '(base...
import pickle import os import sys import pathlib import numpy as np from torch import optim from torch.utils.data import DataLoader import torch from torch.nn.utils import clip_grad_norm_ from tqdm import tqdm from tensorboardX import SummaryWriter abs_path = pathlib.Path(__file__).parent.absolute() sys.path.append(...
[ "evaluate.evaluate" ]
[((320, 345), 'sys.path.append', 'sys.path.append', (['abs_path'], {}), '(abs_path)\n', (335, 345), False, 'import sys\n'), ((908, 957), 'torch.device', 'torch.device', (["('cuda' if config.is_cuda else 'cpu')"], {}), "('cuda' if config.is_cuda else 'cpu')\n", (920, 957), False, 'import torch\n'), ((971, 977), 'model.P...
""" Evaluate the output files to get the numbers reported in ACL18""" import argparse from os.path import join, abspath, dirname, exists from evaluate import eval_meteor, eval_rouge _REF_DIR = join(abspath(dirname(__file__)), 'acl18_results') def main(args): dec_dir = args.decode_dir ref_dir = join(_REF_DI...
[ "evaluate.eval_rouge", "evaluate.eval_meteor" ]
[((308, 335), 'os.path.join', 'join', (['_REF_DIR', '"""reference"""'], {}), "(_REF_DIR, 'reference')\n", (312, 335), False, 'from os.path import join, abspath, dirname, exists\n'), ((706, 722), 'os.path.exists', 'exists', (['_REF_DIR'], {}), '(_REF_DIR)\n', (712, 722), False, 'from os.path import join, abspath, dirnam...
from torch.utils.data.dataloader import default_collate import visual_visdom import evaluate ######################################################### ## Callback-functions for evaluating model-performance ## ######################################################### def _sample_cb(log, config, visdom=None, test_data...
[ "evaluate.precision" ]
[((2402, 2672), 'evaluate.precision', 'evaluate.precision', (['classifier', 'test_datasets', 'task', 'iteration'], {'classes_per_task': 'classes_per_task', 'scenario': 'scenario', 'precision_dict': 'precision_dict', 'collate_fn': 'collate_fn', 'test_size': 'test_size', 'visdom': 'visdom', 'summary_graph': 'summary_grap...
import torch from torch.utils.data import DataLoader from torch.utils.tensorboard import SummaryWriter import argparse import numpy as np import os from data import build_train_dataset from gmflow.gmflow import GMFlow from loss import flow_loss_func from evaluate import (validate_chairs, validate_things, v...
[ "evaluate.validate_chairs", "evaluate.create_sintel_submission", "evaluate.validate_kitti", "evaluate.create_kitti_submission", "evaluate.inference_on_dir", "evaluate.validate_sintel", "evaluate.validate_things" ]
[((623, 648), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (646, 648), False, 'import argparse\n'), ((6394, 6417), 'torch.manual_seed', 'torch.manual_seed', (['seed'], {}), '(seed)\n', (6411, 6417), False, 'import torch\n'), ((6423, 6443), 'numpy.random.seed', 'np.random.seed', (['seed'], {})...
#!/usr/bin/env python2 from __future__ import print_function import os import sys from joblib import Parallel, delayed import numpy as np from tqdm import tqdm sys.path.insert(0, '../prog_data/') import evaluate sys.path.append('%s/../prog_common' % os.path.dirname(os.path.realpath(__file__))) from cmd_args import...
[ "evaluate.tokenize", "evaluate.parse", "evaluate.get_parser", "evaluate.eval_at" ]
[((164, 199), 'sys.path.insert', 'sys.path.insert', (['(0)', '"""../prog_data/"""'], {}), "(0, '../prog_data/')\n", (179, 199), False, 'import pdb, traceback, sys, code\n'), ((606, 626), 'numpy.random.seed', 'np.random.seed', (['seed'], {}), '(seed)\n', (620, 626), True, 'import numpy as np\n'), ((699, 713), 'att_model...
import itertools import math import os import random import time import datetime from statistics import mean import torch import torch.nn as nn import torch.backends.cudnn as cudnn import torchvision.transforms as transforms from torch.utils.tensorboard import SummaryWriter # tensorboard --logdir /home/isaeng/Exjobb/...
[ "evaluate.evaluate" ]
[((989, 1040), 'os.path.join', 'os.path.join', (['EXTR_PATH_SSD', '"""final_data_info.json"""'], {}), "(EXTR_PATH_SSD, 'final_data_info.json')\n", (1001, 1040), False, 'import os\n'), ((1056, 1093), 'os.path.join', 'os.path.join', (['EXTR_PATH_SSD', '"""final/"""'], {}), "(EXTR_PATH_SSD, 'final/')\n", (1068, 1093), Fal...
# @yifan # 2021.12.05 # Hierarchical VQ # return label and depth import numpy as np import faiss from myKMeans import myKMeans from evaluate import MSE class HKM(): def __init__(self, n_nodes_p_level): self.n_nodes_p_level = n_nodes_p_level self.depth = len(n_nodes_p_level) self.init_clust...
[ "evaluate.MSE" ]
[((1879, 1911), 'faiss.IndexFlatL2', 'faiss.IndexFlatL2', (['cent.shape[1]'], {}), '(cent.shape[1])\n', (1896, 1911), False, 'import faiss\n'), ((674, 718), 'myKMeans.myKMeans', 'myKMeans', ([], {'n_clusters': 'self.n_nodes_p_level[i]'}), '(n_clusters=self.n_nodes_p_level[i])\n', (682, 718), False, 'from myKMeans impor...
from flask import Flask from flask import render_template from flask import request from time import process_time from game import Game import evaluate app = Flask(__name__) game = Game() @app.route("/") def main(): global game game = Game() game.depth = 4 game.timeout = 60 return render_template('index....
[ "evaluate.evaluate" ]
[((160, 175), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (165, 175), False, 'from flask import Flask\n'), ((184, 190), 'game.Game', 'Game', ([], {}), '()\n', (188, 190), False, 'from game import Game\n'), ((243, 249), 'game.Game', 'Game', ([], {}), '()\n', (247, 249), False, 'from game import Game\n'),...
import os import sys sys.path.insert(1, os.path.join(sys.path[0], '../utils')) import numpy as np import argparse import h5py import math import time import logging import matplotlib.pyplot as plt import torch import torch.nn as nn import torch.nn.functional as F import torch.optim as optim import pickle from utilities...
[ "evaluate.StatisticsContainer", "evaluate.Evaluator" ]
[((40, 77), 'os.path.join', 'os.path.join', (['sys.path[0]', '"""../utils"""'], {}), "(sys.path[0], '../utils')\n", (52, 77), False, 'import os\n'), ((3402, 3496), 'os.path.join', 'os.path.join', (['workspace', '"""cross_validation_metadata"""', '"""train_curated_cross_validation.csv"""'], {}), "(workspace, 'cross_vali...
#!/usr/bin/env python # coding: utf-8 from __future__ import division, print_function, unicode_literals import argparse import json import os import shutil import time from datetime import datetime from io import open from itertools import product from pathlib import Path import submitit import torch import wandb fro...
[ "evaluate.MultiWozEvaluator" ]
[((9406, 9431), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (9429, 9431), False, 'import argparse\n'), ((1397, 1506), 'model.model.Model', 'Model', (['args', 'input_lang_index2word', 'output_lang_index2word', 'input_lang_word2index', 'output_lang_word2index'], {}), '(args, input_lang_index2w...
from evaluate import evaluate_activity_detection from app import main, init_activity_detection, init_classificator, init_feature_extraction, init_pre_processing from models import Waveform from utils import load_groundtruth, load_annotation, read_csv, plot_metrics_boxplot import numpy as np import os import csv ''' Ac...
[ "evaluate.evaluate_activity_detection" ]
[((1249, 1274), 'os.path.exists', 'os.path.exists', (['tests_dir'], {}), '(tests_dir)\n', (1263, 1274), False, 'import os\n'), ((1280, 1302), 'os.makedirs', 'os.makedirs', (['tests_dir'], {}), '(tests_dir)\n', (1291, 1302), False, 'import os\n'), ((1746, 1768), 'models.Waveform', 'Waveform', ([], {'path': 'wav_dir'}), ...
import sys import os import keras import random as rn import numpy as np import tensorflow as tf from keras.optimizers import Adam from keras.regularizers import l2 from keras import backend as K from keras.models import Model from evaluate import evaluate config = tf.ConfigProto() config.gpu_options.allow_growth = Tru...
[ "evaluate.evaluate" ]
[((266, 282), 'tensorflow.ConfigProto', 'tf.ConfigProto', ([], {}), '()\n', (280, 282), True, 'import tensorflow as tf\n'), ((330, 355), 'tensorflow.Session', 'tf.Session', ([], {'config': 'config'}), '(config=config)\n', (340, 355), True, 'import tensorflow as tf\n'), ((358, 396), 'keras.backend.tensorflow_backend.set...
import glob import logging import os import pickle import random import re import shutil from typing import Dict, List, Tuple import pandas as pd import numpy as np import torch from sklearn.model_selection import train_test_split from torch.nn.utils.rnn import pad_sequence from torch.utils.data impo...
[ "evaluate.evaluate" ]
[((974, 1001), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (991, 1001), False, 'import logging\n'), ((1033, 1066), 'transformers.MODEL_WITH_LM_HEAD_MAPPING.keys', 'MODEL_WITH_LM_HEAD_MAPPING.keys', ([], {}), '()\n', (1064, 1066), False, 'from transformers import MODEL_WITH_LM_HEAD_MAPP...
import pandas as pd from pandas.testing import assert_frame_equal from evaluate.report import ( PrecisionReport, RecallReport, Report, DelimNotFoundError, ReturnTypeDoesNotMatchError ) from evaluate.classification import AlignmentAssessment import pytest from io import StringIO import math from test...
[ "evaluate.report.RecallReport", "evaluate.report.PrecisionReport", "evaluate.report.PrecisionReport.from_files", "evaluate.report.RecallReport.from_files", "evaluate.report.Report.get_value_from_header_fast" ]
[((23973, 24045), 'unittest.mock.patch.object', 'patch.object', (['RecallReport', 'RecallReport._create_helper_columns.__name__'], {}), '(RecallReport, RecallReport._create_helper_columns.__name__)\n', (23985, 24045), False, 'from unittest.mock import patch\n'), ((24742, 24814), 'unittest.mock.patch.object', 'patch.obj...
#!/usr/bin/env python import math import config import configspark import ml_parse import evaluate sc = configspark.SPARK_CONTEXT def clean(): config.clean_path(config.ML_MODEL) def main(): clean() ratings_train_text = sc.textFile(config.ML_RATINGS_TRAIN) ratings_train = ( ratings_train_t...
[ "evaluate.evaluate" ]
[((151, 185), 'config.clean_path', 'config.clean_path', (['config.ML_MODEL'], {}), '(config.ML_MODEL)\n', (168, 185), False, 'import config\n'), ((679, 755), 'evaluate.evaluate', 'evaluate.evaluate', (['ratings_train', 'ratings_validation', 'config.ML_RESULTS_FILE'], {}), '(ratings_train, ratings_validation, config.ML_...
""" Evaluate trained models and print errors on test and dev datasets """ import logging import os from os.path import basename from evaluate import evaluate import config def run_evaluate(): """ Run evaluation on 2 models: * baseline * baseline + lexical Save evaluation and error analy...
[ "evaluate.evaluate" ]
[((617, 666), 'logging.info', 'logging.info', (['f"""Reports directory: {reports_dir}"""'], {}), "(f'Reports directory: {reports_dir}')\n", (629, 666), False, 'import logging\n'), ((671, 710), 'os.makedirs', 'os.makedirs', (['reports_dir'], {'exist_ok': '(True)'}), '(reports_dir, exist_ok=True)\n', (682, 710), False, '...
import argparse, torch, gc, os, random, json from data import MyDataset, load_data, my_collate_fn, device, word2feature, label_multihot, mask_noFeature from model import MCEncoder from lcmodel import LCEncoder from tqdm import tqdm from evaluate import evaluate, evaluate_test import numpy as np import os def main(freq...
[ "evaluate.evaluate_test", "evaluate.evaluate" ]
[((486, 523), 'os.makedirs', 'os.makedirs', (['directory'], {'exist_ok': '(True)'}), '(directory, exist_ok=True)\n', (497, 523), False, 'import os\n'), ((627, 647), 'data.load_data', 'load_data', (['frequency'], {}), '(frequency)\n', (636, 647), False, 'from data import MyDataset, load_data, my_collate_fn, device, word...
#!/usr/bin/env python3 from pathlib import Path import argparse import listing # Version of the format where I save the models, cases, etc. If in the future # I change the format I can just change the string, so a new folder will be made # and old things will be left ignored in old folder DATA_VERSION = "v5" CASES_P...
[ "evaluate.evaluate" ]
[((326, 358), 'pathlib.Path', 'Path', (['f"""./cases/{DATA_VERSION}/"""'], {}), "(f'./cases/{DATA_VERSION}/')\n", (330, 358), False, 'from pathlib import Path\n'), ((401, 426), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (424, 426), False, 'import argparse\n'), ((636, 659), 'train.train', 't...
# Entry point for a script to attempt vocal and music separation using Tensorflow # Accompanying thesis: Vocal and Audio separation using deep learning for Riga Technical University. # Author: <NAME> <<EMAIL>>, Student ID: 161RDB280 import logging import argparse import os import sys from dataset import Dataset from m...
[ "evaluate.Evaluator" ]
[((488, 516), 'config.prepare_config', 'prepare_config', (['"""config.ini"""'], {}), "('config.ini')\n", (502, 516), False, 'from config import prepare_config\n'), ((1009, 1043), 'logging.addLevelName', 'logging.addLevelName', (['(55)', '"""Hello!"""'], {}), "(55, 'Hello!')\n", (1029, 1043), False, 'import logging\n'),...
import os import glob import numpy as np import caffe import evaluate import config caffe.set_mode_gpu() caffe.set_device(0) model_weights = os.path.join(config.LRCN_MODELS_DIR, 'RGB_lstm_model_iter_30000.caffemodel') h5Files = glob.glob(os.path.join(config.DATASET_DIR, 'extracted_features_lstm_RGB/*.h5')) model_lst...
[ "evaluate.evaluate" ]
[((86, 106), 'caffe.set_mode_gpu', 'caffe.set_mode_gpu', ([], {}), '()\n', (104, 106), False, 'import caffe\n'), ((107, 126), 'caffe.set_device', 'caffe.set_device', (['(0)'], {}), '(0)\n', (123, 126), False, 'import caffe\n'), ((144, 220), 'os.path.join', 'os.path.join', (['config.LRCN_MODELS_DIR', '"""RGB_lstm_model_...
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import, division, print_function import os import sys log_level_index = sys.argv.index('--log_level') + 1 if '--log_level' in sys.argv else 0 os.environ['TF_CPP_MIN_LOG_LEVEL'] = sys.argv[log_level_index] if log_level_index > 0 and log_leve...
[ "evaluate.evaluate" ]
[((3177, 3212), 'tensorflow.transpose', 'tf.transpose', (['batch_x', '[1, 0, 2, 3]'], {}), '(batch_x, [1, 0, 2, 3])\n', (3189, 3212), True, 'import tensorflow as tf\n'), ((3274, 3360), 'tensorflow.reshape', 'tf.reshape', (['batch_x', '[-1, Config.n_input + 2 * Config.n_input * Config.n_context]'], {}), '(batch_x, [-1, ...
import misc, model, eval, evaluate import numpy as np import torch, time, random from torch import nn import TD_RvNN device = 'cuda' tree_file = 'resource/data.TD_RvNN.vol_5000.txt' label_file = 'resource/Twitter15_label_All.txt' train_file = 'nfold/RNNtrainSet_Twitter152_tree.txt' test_file = 'nfold/RNNtestSet_Tw...
[ "evaluate.evaluation_4class" ]
[((756, 774), 'TD_RvNN.loadData', 'TD_RvNN.loadData', ([], {}), '()\n', (772, 774), False, 'import TD_RvNN\n'), ((1309, 1336), 'torch.nn.MSELoss', 'nn.MSELoss', ([], {'reduction': '"""sum"""'}), "(reduction='sum')\n", (1319, 1336), False, 'from torch import nn\n'), ((1575, 1586), 'time.time', 'time.time', ([], {}), '()...
import os import util import time import math import torch import logging import evaluate import coref_model format = '%(asctime)s - %(levelname)s - %(name)s - %(message)s' logging.basicConfig(format=format) logger = logging.getLogger(__name__) logger.setLevel(logging.INFO) if __name__ == "__main__": os.environ[...
[ "evaluate.evaluate" ]
[((174, 208), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': 'format'}), '(format=format)\n', (193, 208), False, 'import logging\n'), ((218, 245), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (235, 245), False, 'import logging\n'), ((357, 382), 'torch.cuda.is_available', ...
import time import numpy as np import torch import os from evaluate import angle_acc import argparse from model import DenseNet_GCN from loss import JSD_Loss from dataset import Sphere_Dataset from train import train from torch.utils.data import DataLoader def test(epoch, device, data_loader, model, criterion, vis, s...
[ "evaluate.angle_acc" ]
[((610, 621), 'time.time', 'time.time', ([], {}), '()\n', (619, 621), False, 'import time\n'), ((4085, 4110), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (4108, 4110), False, 'import argparse\n'), ((4717, 4771), 'dataset.Sphere_Dataset', 'Sphere_Dataset', ([], {'root': 'test_opts.data_path',...
import os import time import numpy as np import torch from torch.autograd import Variable import torchvision.utils as vutils from options import TestOptions from edges2shoes_data import DataLoader, load_edges2shoes, AlignedIterator, UnalignedIterator from model import StochCycleGAN, AugmentedCycleGAN import cPickle as ...
[ "evaluate.eval_mse_A", "evaluate.eval_ubo_B" ]
[((748, 779), 'os.path.join', 'os.path.join', (['opt.res_dir', 'name'], {}), '(opt.res_dir, name)\n', (760, 779), False, 'import os\n'), ((1208, 1239), 'os.path.join', 'os.path.join', (['opt.res_dir', 'name'], {}), '(opt.res_dir, name)\n', (1220, 1239), False, 'import os\n'), ((2036, 2067), 'os.path.join', 'os.path.joi...
from __future__ import print_function # from network import VGGNet from dirtorch.utils import common import dirtorch.nets as nets import pandas as pd import faiss import torch import torchvision.transforms as transforms import torch.nn as nn from six.moves import cPickle import numpy as np import image...
[ "evaluate.evaluate_class" ]
[((487, 523), 'dirtorch.utils.common.load_checkpoint', 'common.load_checkpoint', (['path', 'iscuda'], {}), '(path, iscuda)\n', (509, 523), False, 'from dirtorch.utils import common\n'), ((535, 598), 'dirtorch.nets.create_model', 'nets.create_model', ([], {'pretrained': '""""""'}), "(pretrained='', **checkpoint['model_o...
from config import Arguments as args import os os.environ["CUDA_DEVICE_ORDER"]="PCI_BUS_ID" os.environ["CUDA_VISIBLE_DEVICES"]=args.train_visible_devices import sys, random import numpy as np import torch import torch.nn as nn from torch.optim import Adam from torch.utils.data import DataLoader from torch.utils.t...
[ "evaluate.evaluate" ]
[((2638, 2793), 'utils.scheduler.WarmupScheduler', 'WarmupScheduler', (['optimizer'], {'warmup_epochs': 'args.warmup_steps', 'initial_lr': 'args.init_lr', 'max_lr': 'args.max_lr', 'milestones': 'args.milestones', 'gamma': 'args.gamma'}), '(optimizer, warmup_epochs=args.warmup_steps, initial_lr=args\n .init_lr, max_l...
from evaluate import evaluate from os_utils import OSUtils from save_slot_checkpoint import save_slot from torch import nn from tqdm import tqdm def train(model, train_dl, val_dl, args, optimizer, scheduler=None): print(f"training on {len(train_dl.dataset)}, validating on {len(val_dl.dataset)} examples.") c...
[ "evaluate.evaluate" ]
[((331, 352), 'torch.nn.CrossEntropyLoss', 'nn.CrossEntropyLoss', ([], {}), '()\n', (350, 352), False, 'from torch import nn\n'), ((1128, 1175), 'evaluate.evaluate', 'evaluate', (['model', 'val_dl', 'criterion', 'args.device'], {}), '(model, val_dl, criterion, args.device)\n', (1136, 1175), False, 'from evaluate import...
import os import json import matplotlib matplotlib.use('TkAgg') import matplotlib.pyplot as plt from evaluate import evaluate if __name__ == '__main__': training_commands, predict_commands = [], [] seq2vec_name_to_last_layer = {"dan": 4, "gru": 4} probing_accuracies = {} for seq2vec_name, layer in ...
[ "evaluate.evaluate" ]
[((40, 63), 'matplotlib.use', 'matplotlib.use', (['"""TkAgg"""'], {}), "('TkAgg')\n", (54, 63), False, 'import matplotlib\n'), ((4198, 4220), 'matplotlib.pyplot.ylabel', 'plt.ylabel', (['"""Accuracy"""'], {}), "('Accuracy')\n", (4208, 4220), True, 'import matplotlib.pyplot as plt\n'), ((4225, 4288), 'matplotlib.pyplot....
""" Baseline training script """ import os import sys import copy import glob import math import json import shutil import argparse import logging from typing import Any, Dict, Optional from types import SimpleNamespace from itertools import cycle from contextlib import ExitStack from comet_ml import Experiment # mus...
[ "evaluate.Evaluator" ]
[((2586, 2607), 'metrics.MetricStore', 'metrics.MetricStore', ([], {}), '()\n', (2605, 2607), False, 'import metrics\n'), ((3138, 3230), 'experiment.initialize_experiment', 'initialize_experiment', (['self.args', "('data', 'model', 'optim')", 'self.args.experiment_name'], {}), "(self.args, ('data', 'model', 'optim'), s...
import pickle import numpy import glob import os import sys from keras.callbacks import ModelCheckpoint from keras.layers import LSTM from kutilities.callbacks import MetricsCallback, PlottingCallback from kutilities.helpers.data_preparation import get_labels_to_categories_map, \ get_class_weights2, onehot_to_categ...
[ "evaluate.evaluate.performance_analysis" ]
[((742, 765), 'numpy.random.seed', 'numpy.random.seed', (['(1337)'], {}), '(1337)\n', (759, 765), False, 'import numpy\n'), ((2296, 2340), 'utilities_nn.data_loader.get_embeddings', 'get_embeddings', ([], {'corpus': 'WV_CORPUS', 'dim': 'WV_DIM'}), '(corpus=WV_CORPUS, dim=WV_DIM)\n', (2310, 2340), False, 'from utilities...
import argparse import functools import itertools import os.path import time from pprint import pprint import torch import numpy as np from benepar import char_lstm from benepar import decode_chart from benepar import nkutil from benepar import parse_chart import evaluate import learning_rates import treebanks def...
[ "evaluate.evalb" ]
[((732, 1457), 'benepar.nkutil.HParams', 'nkutil.HParams', ([], {'max_len_train': '(0)', 'max_len_dev': '(0)', 'batch_size': '(32)', 'learning_rate': '(5e-05)', 'learning_rate_warmup_steps': '(160)', 'clip_grad_norm': '(0.0)', 'checks_per_epoch': '(4)', 'step_decay_factor': '(0.5)', 'step_decay_patience': '(5)', 'max_c...
import chess import random from evaluate import evaluate #import copy #import os #import psutil from multiprocessing import Pool count = 0 # TODO Learn about castling properly! # TODO Eliminate 3-fold repetition! See code of main.py # TODO Implement time constraints to avoid "Black forfeits on time" def search(board...
[ "evaluate.evaluate" ]
[((799, 820), 'evaluate.evaluate', 'evaluate', (['board', 'turn'], {}), '(board, turn)\n', (807, 820), False, 'from evaluate import evaluate\n')]
""" WORKER OF HPBANDSTER OPTIMIZER """ import numpy import time import ConfigSpace as CS from hpbandster.core.worker import Worker import os from evaluate import evaluate from set_up_run_folder import set_up_run_folder class MyWorker(Worker): def __init__(self, *args, **kwargs): super().__init__(*args,...
[ "evaluate.evaluate" ]
[((1576, 1595), 'set_up_run_folder.set_up_run_folder', 'set_up_run_folder', ([], {}), '()\n', (1593, 1595), False, 'from set_up_run_folder import set_up_run_folder\n'), ((2613, 2636), 'ConfigSpace.ConfigurationSpace', 'CS.ConfigurationSpace', ([], {}), '()\n', (2634, 2636), True, 'import ConfigSpace as CS\n'), ((1973, ...
import sys import time from torch.utils.data import DataLoader from torch.utils.tensorboard import SummaryWriter from torchvision import transforms from ray import tune from ray.tune.schedulers import HyperBandScheduler from datasets import dataset_factory from datasets.augmentation import * from datasets.graph impor...
[ "evaluate.evaluate" ]
[((612, 626), 'utils.AverageMeter', 'AverageMeter', ([], {}), '()\n', (624, 626), False, 'from utils import AverageMeter\n'), ((643, 657), 'utils.AverageMeter', 'AverageMeter', ([], {}), '()\n', (655, 657), False, 'from utils import AverageMeter\n'), ((671, 685), 'utils.AverageMeter', 'AverageMeter', ([], {}), '()\n', ...
import datetime import os import json import yaml import visdom import torch import numpy as np from gamified_squad import GamifiedSquad from gamified_newsqa import GamifiedNewsQA from agent import Agent, HistoryScoreCache import generic import evaluate def train(): time_1 = datetime.datetime.now() wit...
[ "evaluate.evaluate" ]
[((284, 307), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (305, 307), False, 'import datetime\n'), ((565, 572), 'agent.Agent', 'Agent', ([], {}), '()\n', (570, 572), False, 'from agent import Agent, HistoryScoreCache\n'), ((597, 612), 'visdom.Visdom', 'visdom.Visdom', ([], {}), '()\n', (610, 612...
import os import sys import math import argparse import numpy as np from tqdm import tqdm import torch from torch.multiprocessing import Queue, Process sys.path.insert(0, '../lib') sys.path.insert(0, '../model') from data.WiderPerson import WiderPerson from utils import misc_utils, nms_utils from evaluate import comp...
[ "evaluate.compute_JI.evaluation_all", "evaluate.compute_APMR.compute_APMR" ]
[((154, 182), 'sys.path.insert', 'sys.path.insert', (['(0)', '"""../lib"""'], {}), "(0, '../lib')\n", (169, 182), False, 'import sys\n'), ((183, 213), 'sys.path.insert', 'sys.path.insert', (['(0)', '"""../model"""'], {}), "(0, '../model')\n", (198, 213), False, 'import sys\n'), ((410, 468), 'os.path.join', 'os.path.joi...
import os import sys import time import random import argparse import shutil import logging import torch import numpy as np from allennlp.data.iterators import BasicIterator from preprocess import build_tasks from models import build_model from trainer import build_trainer from evaluate import evaluate from util impor...
[ "evaluate.evaluate" ]
[((405, 444), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '""""""'}), "(description='')\n", (428, 444), False, 'import argparse\n'), ((5828, 5871), 'os.makedirs', 'os.makedirs', (['args.store_root'], {'exist_ok': '(True)'}), '(args.store_root, exist_ok=True)\n', (5839, 5871), False, 'impo...
import config from utils import squad_evaluate, load_and_cache_examples import glob import json import logging import os import random from evaluate import evaluate import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler from torch.utils.data.distributed import DistributedSampler from tqd...
[ "evaluate.evaluate" ]
[((429, 456), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (446, 456), False, 'import logging\n'), ((555, 577), 'random.seed', 'random.seed', (['args.seed'], {}), '(args.seed)\n', (566, 577), False, 'import random\n'), ((582, 607), 'numpy.random.seed', 'np.random.seed', (['args.seed'], ...
"""Train the model""" import argparse import logging import os import random import numpy as np import torch import torch.optim as optim from torch.autograd import Variable from tqdm import tqdm import sys sys.path.append('') #os.chdir("..") import utils from data.data_utils import fetch_data, FERDataset, DataLoade...
[ "evaluate.evaluate" ]
[((209, 228), 'sys.path.append', 'sys.path.append', (['""""""'], {}), "('')\n", (224, 228), False, 'import sys\n'), ((1398, 1428), 'train_utils.ConfusionMatrix', 'ConfusionMatrix', ([], {'num_classes': '(8)'}), '(num_classes=8)\n', (1413, 1428), False, 'from train_utils import MarginCalibratedCELoss, ConfusionMatrix\n'...
import numpy as np import torch import torch.nn as nn import os import matplotlib.pyplot as plt import torchvision.transforms as transforms from utils.opt import parse_option from torch.utils.tensorboard import SummaryWriter from torchvision.datasets import mnist from torch.nn import CrossEntropyLoss from torch.optim i...
[ "evaluate.validate" ]
[((520, 534), 'utils.opt.parse_option', 'parse_option', ([], {}), '()\n', (532, 534), False, 'from utils.opt import parse_option\n'), ((995, 1036), 'torch.utils.data.DataLoader', 'DataLoader', (['train_dataset', 'opt.batch_size'], {}), '(train_dataset, opt.batch_size)\n', (1005, 1036), False, 'from torch.utils.data imp...
import argparse from tools import ParsingCfg import numpy as np from tools import InitDataDir # ### # import os # os.environ["CUDA_VISIBLE_DEVICES"] = "0" # ### if __name__ == '__main__': parser=argparse.ArgumentParser() parser.add_argument("cfg_path",help="config file path",type=str) parse...
[ "evaluate.cocoeval.COCOEval", "evaluate.Evaluation", "evaluate.FramePerSecond" ]
[((214, 239), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (237, 239), False, 'import argparse\n'), ((1771, 1791), 'tools.ParsingCfg', 'ParsingCfg', (['cfg_path'], {}), '(cfg_path)\n', (1781, 1791), False, 'from tools import ParsingCfg\n'), ((2112, 2144), 'numpy.array', 'np.array', (["cfg_dic...
from training import train from testing import test from evaluate import evaluate def random_forest(path_train, path_test, name_identifiers, name_targets, features, delimiter, num_cores=1): ''' this met...
[ "evaluate.evaluate" ]
[((1489, 1618), 'training.train', 'train', (['path_train', 'name_identifiers', 'name_targets', 'features'], {'output_path_model': 'None', 'cores': 'num_cores', 'the_delimiter': 'delimiter'}), '(path_train, name_identifiers, name_targets, features,\n output_path_model=None, cores=num_cores, the_delimiter=delimiter)\n...
import numpy as np from scipy import sparse import torch import time from tqdm import tqdm from evaluate import eval_func, euclidean_dist def calculate_V(initial_rank, all_feature_len, dis_i_qg, i, k1): # dis_i_qg = euclidean_dist(torch.tensor([all_feature[i].numpy()]), all_feature).numpy() forward_k_nei...
[ "evaluate.euclidean_dist" ]
[((1426, 1465), 'numpy.unique', 'np.unique', (['k_reciprocal_expansion_index'], {}), '(k_reciprocal_expansion_index)\n', (1435, 1465), True, 'import numpy as np\n'), ((1521, 1568), 'numpy.exp', 'np.exp', (['(-dis_i_qg[k_reciprocal_expansion_index])'], {}), '(-dis_i_qg[k_reciprocal_expansion_index])\n', (1527, 1568), Tr...
#%% -*- coding: utf-8 -*- # Plotting import matplotlib matplotlib.use('agg') import os import time import argparse import numpy as np import torch import torch.nn as nn import torch.optim as optim # Internal imports from utils.data import load_dataset, get_external_sounds from models.vae.ae import AE, RegressionAE, Di...
[ "evaluate.evaluate_latent_space", "evaluate.evaluate_semantic_parameters", "evaluate.evaluate_projection", "evaluate.evaluate_synthesis", "evaluate.evaluate_latent_neighborhood", "evaluate.evaluate_reconstruction", "evaluate.evaluate_params", "evaluate.evaluate_meta_parameters" ]
[((56, 77), 'matplotlib.use', 'matplotlib.use', (['"""agg"""'], {}), "('agg')\n", (70, 77), False, 'import matplotlib\n'), ((695, 720), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (718, 720), False, 'import argparse\n'), ((6309, 6320), 'time.time', 'time.time', ([], {}), '()\n', (6318, 6320)...
from pathlib import Path import sys sys.path.append(str(Path().absolute())) import logging log_level = "INFO" logging.basicConfig( filename=str(snakemake.log), filemode="w", level=log_level, format="[%(asctime)s]:%(levelname)s: %(message)s", datefmt="%d/%m/%Y %I:%M:%S %p", ) from evaluate.calcula...
[ "evaluate.report.PrecisionReport.from_files", "evaluate.calculator.PrecisionCalculator" ]
[((554, 606), 'pathlib.Path', 'Path', (['snakemake.output.precision_file_for_one_sample'], {}), '(snakemake.output.precision_file_for_one_sample)\n', (558, 606), False, 'from pathlib import Path\n'), ((932, 963), 'logging.info', 'logging.info', (['f"""Loading report"""'], {}), "(f'Loading report')\n", (944, 963), False...
#!/usr/bin/env python import sys import os import config import evaluate import ml_parse from pyspark.mllib.recommendation import MatrixFactorizationModel, ALS def main(): import configspark sc = configspark.SPARK_CONTEXT print("\nLoading MovieLens test dataset\n") test_text = sc.textFile(config.ML...
[ "evaluate.evaluate_model" ]
[((437, 468), 'os.path.exists', 'os.path.exists', (['config.ML_MODEL'], {}), '(config.ML_MODEL)\n', (451, 468), False, 'import os\n'), ((748, 792), 'evaluate.evaluate_model', 'evaluate.evaluate_model', (['model', 'ratings_test'], {}), '(model, ratings_test)\n', (771, 792), False, 'import evaluate\n'), ((589, 639), 'pys...
"""Train the model""" import argparse import logging from tensorboardX import SummaryWriter import os, shutil import numpy as np import pandas as pd from sklearn.utils.class_weight import compute_class_weight import torch import torch.optim as optim import torchvision.models as models from torch.autograd import Varia...
[ "evaluate.evaluate" ]
[((533, 558), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (556, 558), False, 'import argparse\n'), ((3886, 3908), 'utils.RunningAverage', 'utils.RunningAverage', ([], {}), '()\n', (3906, 3908), False, 'import utils\n'), ((13597, 13616), 'torch.cat', 'torch.cat', (['Zhats', '(0)'], {}), '(Zha...
#!/usr/bin/env python import sys import os from shutil import copyfile import time import itertools import torch from torch.autograd import Variable import torchvision.utils as vutils from options import TrainOptions, create_sub_dirs from edges2shoes_data import DataLoader, load_edges2shoes, AlignedIterator, Unaligne...
[ "evaluate.eval_mse_A", "evaluate.eval_ubo_B" ]
[((596, 634), 'os.path.join', 'os.path.join', (['expr_dir', '"""results.json"""'], {}), "(expr_dir, 'results.json')\n", (608, 634), False, 'import os\n'), ((824, 855), 'glob.glob', 'glob.glob', (["('%s/*.py' % dir_path)"], {}), "('%s/*.py' % dir_path)\n", (833, 855), False, 'import glob\n'), ((1615, 1676), 'os.path.joi...
### modified from https://github.com/qiuqiangkong/audioset_tagging_cnn import os import sys sys.path.insert(1, os.path.join(sys.path[0], '../utils')) import numpy as np import argparse import h5py import math import time import logging import matplotlib.pyplot as plt from sklearn import metrics import _pickle as cPickl...
[ "evaluate.Evaluator" ]
[((372, 392), 'torch.manual_seed', 'torch.manual_seed', (['(0)'], {}), '(0)\n', (389, 392), False, 'import torch\n'), ((111, 148), 'os.path.join', 'os.path.join', (['sys.path[0]', '"""../utils"""'], {}), "(sys.path[0], '../utils')\n", (123, 148), False, 'import os\n'), ((2465, 2489), 'losses.get_loss_func', 'get_loss_f...
import argparse import logging import os from types import SimpleNamespace import falcon import pandas import torch from falcon_cors import CORS import json import waitress from data import Data from torch.utils.data import DataLoader from utils import load_torch_model from model import BertForClassification, CharCNN f...
[ "evaluate.evaluate" ]
[((441, 517), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO', 'format': '"""%(asctime)-18s %(message)s"""'}), "(level=logging.INFO, format='%(asctime)-18s %(message)s')\n", (460, 517), False, 'import logging\n'), ((527, 546), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (544,...
''' Created on Aug 9, 2016 Keras Implementation of Neural Matrix Factorization (NeuMF) recommender model in: He Xiangnan et al. Neural Collaborative Filtering. In WWW 2017. @author: <NAME> (<EMAIL>) ''' import argparse from time import time import numpy as np from keras.layers import Embedding, Input, Dens...
[ "evaluate.evaluate_model" ]
[((662, 710), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Run NNCF."""'}), "(description='Run NNCF.')\n", (685, 710), False, 'import argparse\n'), ((3347, 3398), 'keras.layers.Input', 'Input', ([], {'shape': '(1,)', 'dtype': '"""int32"""', 'name': '"""user_input"""'}), "(shape=(1,), d...
#!/usr/bin/env python # -*- encoding: utf-8 -*- # @Version : Python 3.6 import os import json import torch import torch.optim as optim import torch.nn as nn import numpy as np from tqdm import tqdm from config import Config from utils import EmbeddingLoader, RelationLoader, NYTDataLoader from model import DS_Model f...
[ "evaluate.Eval" ]
[((3172, 3180), 'config.Config', 'Config', ([], {}), '()\n', (3178, 3180), False, 'from config import Config\n'), ((3518, 3557), 'utils.NYTDataLoader', 'NYTDataLoader', (['rel2id', 'token2id', 'config'], {}), '(rel2id, token2id, config)\n', (3531, 3557), False, 'from utils import EmbeddingLoader, RelationLoader, NYTDat...
import warnings warnings.filterwarnings("ignore") import os, numpy as np, argparse, random, matplotlib, datetime os.chdir(os.path.dirname(os.path.realpath(__file__))) from pathlib import Path matplotlib.use('agg') from tqdm import tqdm import auxiliaries as aux import datasets as data import netlib as netlib import los...
[ "evaluate.GradientMeasure", "evaluate.DistanceMeasure", "evaluate.evaluate" ]
[((16, 49), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (39, 49), False, 'import warnings\n'), ((192, 213), 'matplotlib.use', 'matplotlib.use', (['"""agg"""'], {}), "('agg')\n", (206, 213), False, 'import os, numpy as np, argparse, random, matplotlib, datetime\n'), ((53...
#!/usr/bin/env python # -*- coding: utf-8 -*- ############################################################################### # Git-based CTF ############################################################################### # # Author: <NAME> <<EMAIL>> # <NAME> <<EMAIL>> # <NAME> <<EMAIL>> # # Copyright (...
[ "evaluate.evaluate" ]
[((3346, 3398), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': 'desc', 'prog': 'prog'}), '(description=desc, prog=prog)\n', (3369, 3398), False, 'import argparse\n'), ((3561, 3649), 'verify_service.verify_service', 'verify_service', (['args.team', 'args.branch', 'args.service_port', 'args.ho...
from ignite.engine.engine import Engine, State, Events from dataset import get_iterator from model import get_model from loss import get_loss from optimizer import get_optimizer from logger import get_logger, log_results from utils import prepare_batch from metric import get_metrics from evaluate import get_evaluator...
[ "evaluate.evaluate_once", "evaluate.get_evaluator" ]
[((853, 873), 'ignite.engine.engine.Engine', 'Engine', (['update_model'], {}), '(update_model)\n', (859, 873), False, 'from ignite.engine.engine import Engine, State, Events\n'), ((1315, 1333), 'dataset.get_iterator', 'get_iterator', (['args'], {}), '(args)\n', (1327, 1333), False, 'from dataset import get_iterator\n')...
# coding:utf-8 """ Filename: train.py Author: @DvdNss Created on 12/17/2021 """ import argparse import os import torch from sklearn.metrics import accuracy_score from torch.utils.data import DataLoader from tqdm import tqdm from transformers import PerceiverForSequenceClassification, AdamW from evaluate import evalu...
[ "evaluate.evaluate" ]
[((471, 496), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (494, 496), False, 'import argparse\n'), ((1556, 1705), 'transformers.PerceiverForSequenceClassification.from_pretrained', 'PerceiverForSequenceClassification.from_pretrained', (['"""deepmind/language-perceiver"""'], {'problem_type': ...
import torch import gym from config import CartPole, Pong from train import train from evaluate import evaluate_policy import os os.environ['KMP_DUPLICATE_LIB_OK'] = 'True' # needed locally ENV_CONFIGS = { 'CartPole-v0': CartPole, 'Pong-v0': Pong } def training(env): """ function that is used to ca...
[ "evaluate.evaluate_policy" ]
[((412, 422), 'train.train', 'train', (['env'], {}), '(env)\n', (417, 422), False, 'from train import train\n'), ((841, 854), 'gym.make', 'gym.make', (['env'], {}), '(env)\n', (849, 854), False, 'import gym\n'), ((1112, 1229), 'evaluate.evaluate_policy', 'evaluate_policy', (['dqn', 'env', 'env_config', 'env_name', 'n_e...
""" Evaluate the baselines ont ROUGE/METEOR""" import argparse import json import os from os.path import join, exists from evaluate import eval_meteor, eval_rouge try: _DATA_DIR = os.environ['DATA'] except KeyError: print('please use environment variable to specify data directories') def main(args): de...
[ "evaluate.eval_meteor", "evaluate.eval_rouge" ]
[((328, 359), 'os.path.join', 'join', (['args.decode_dir', '"""output"""'], {}), "(args.decode_dir, 'output')\n", (332, 359), False, 'from os.path import join, exists\n'), ((475, 505), 'os.path.join', 'join', (['_DATA_DIR', '"""refs"""', 'split'], {}), "(_DATA_DIR, 'refs', split)\n", (479, 505), False, 'from os.path im...
import os import time import json import logging import argparse import sys sys.path.append("libs") from utils import Init_logging from utils import PiecewiseSchedule import numpy as np import pandas as pd import tensorflow as tf from tensorflow.keras import optimizers from tensorflow.keras import call...
[ "evaluate.EvaluateCold", "evaluate.EvaluateModel" ]
[((84, 107), 'sys.path.append', 'sys.path.append', (['"""libs"""'], {}), "('libs')\n", (99, 107), False, 'import sys\n'), ((888, 921), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (911, 921), False, 'import warnings\n'), ((952, 973), 'numpy.random.seed', 'np.random.seed'...
import pandas as pd import matplotlib.pyplot as plt import keras from keras import layers from keras.utils import plot_model from keras.constraints import non_neg from keras.callbacks import ModelCheckpoint from sklearn.model_selection import train_test_split from sklearn.metrics import mean_squared_error from sklearn....
[ "evaluate.mean_average_precision", "evaluate.roc_auc", "evaluate.normalized_dcg" ]
[((651, 675), 'load_ratings.load', 'load_ratings.load', (['"""../"""'], {}), "('../')\n", (668, 675), False, 'import load_ratings\n'), ((751, 807), 'sklearn.model_selection.train_test_split', 'train_test_split', (['dataset'], {'test_size': '(0.3)', 'random_state': '(0)'}), '(dataset, test_size=0.3, random_state=0)\n', ...
#保存结果与建议至json文件 import time import json from evaluate import evaluate,getTimes from getdata import get_alldata def get_advice(proposal,times): standard_proposal=[x/times for x in proposal] if max(standard_proposal)<1.5: #print("动作标准,请继续保持") return "动作标准,请继续保持" elif max(standard_proposal[...
[ "evaluate.getTimes" ]
[((1263, 1298), 'getdata.get_alldata', 'get_alldata', (['"""./jsonfile/standard/"""'], {}), "('./jsonfile/standard/')\n", (1274, 1298), False, 'from getdata import get_alldata\n'), ((1336, 1382), 'getdata.get_alldata', 'get_alldata', (['"""./jsonfile/now/"""'], {'isRunning': '(True)'}), "('./jsonfile/now/', isRunning=T...
import matplotlib.pyplot as plt import numpy as np import matplotlib as mpl import pandas as pd import sys sys.path.append("..") mpl.use('tkagg') # issues with Big Sur import matplotlib.pyplot as plt from strategy.bollinger_bands import bollinger_bands from backtest import Backtest from evaluate import PortfolioRetu...
[ "evaluate.CAGR", "evaluate.SharpeRatio", "evaluate.MaxDrawdown" ]
[((107, 128), 'sys.path.append', 'sys.path.append', (['""".."""'], {}), "('..')\n", (122, 128), False, 'import sys\n'), ((129, 145), 'matplotlib.use', 'mpl.use', (['"""tkagg"""'], {}), "('tkagg')\n", (136, 145), True, 'import matplotlib as mpl\n'), ((373, 481), 'pandas.read_csv', 'pd.read_csv', (['"""../../database/hke...
import os import re import sys sys.path.append('.') import cv2 import csv import math import time import string import random import scipy import argparse import matplotlib import numpy as np import pylab as plt import torch import torch.nn as nn import torch.nn.functional as F from torch.autograd import Variable from ...
[ "evaluate.coco_eval.get_outputs" ]
[((31, 51), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (46, 51), False, 'import sys\n'), ((964, 989), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (987, 989), False, 'import argparse\n'), ((7240, 7264), 'lib.config.update_config', 'update_config', (['cfg', 'args'], {}...
import os import matplotlib.pyplot as plt import matplotlib.ticker as ticker import torch import torch.nn as nn from torch.autograd import Variable from torch import optim import torch.nn.functional as F import logging from evaluate import test class Solver(object): default_adam_args = {"lr": 0.001, ...
[ "evaluate.test" ]
[((700, 719), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (717, 719), False, 'import logging\n'), ((1684, 1705), 'torch.nn.CrossEntropyLoss', 'nn.CrossEntropyLoss', ([], {}), '()\n', (1703, 1705), True, 'import torch.nn as nn\n'), ((1576, 1610), 'os.path.join', 'os.path.join', (['model_dir', '"""cnn.pkl...
import pathlib import itertools import logging import logging.config from dataclasses import dataclass import yaap import torch import torch.utils.tensorboard import torchmodels import utils import train import evaluate import models import models.dialog import datasets from datasets import Dialog from loopers import...
[ "evaluate.evaluate" ]
[((827, 838), 'yaap.Yaap', 'yaap.Yaap', ([], {}), '()\n', (836, 838), False, 'import yaap\n'), ((5537, 5564), 'pathlib.Path', 'pathlib.Path', (['args.save_dir'], {}), '(args.save_dir)\n', (5549, 5564), False, 'import pathlib\n'), ((5762, 5780), 'utils.ShellUtils', 'utils.ShellUtils', ([], {}), '()\n', (5778, 5780), Fal...
"""Train the model""" import argparse import logging import os from tqdm import tqdm import numpy as np import torch import torch.nn as nn import torch.optim as optim import torch.optim.lr_scheduler as lr_scheduler from torch.autograd import Variable import models.nets as nets import models.vgg as vgg import models....
[ "evaluate.evaluate" ]
[((402, 427), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (425, 427), False, 'import argparse\n'), ((1889, 1911), 'utils.RunningAverage', 'utils.RunningAverage', ([], {}), '()\n', (1909, 1911), False, 'import utils\n'), ((3892, 3942), 'logging.info', 'logging.info', (["('- Train metrics: ' +...
import train import evaluate if __name__ == "__main__": #trainer = train.TumorDetectionNet() #trainer.train(path_to_dataset="archive/", model_filename="Tumor_classifier_model.h5") evaluator = evaluate.Evaluator() evaluator.evaluate(model_path="Tumor_classifier_model_v2.h5", image_path="archive/validat...
[ "evaluate.Evaluator" ]
[((206, 226), 'evaluate.Evaluator', 'evaluate.Evaluator', ([], {}), '()\n', (224, 226), False, 'import evaluate\n')]
from evaluate import evaluate __author__ = "<NAME>" __email__ = "<EMAIL>" data = '[{"state": {"cities": ["Mumbai", "Pune", "Nagpur", "Bhusaval", "Jalgaon"], "name": "Maharashtra"}}, {"state": {"cities": ["Bangalore", "Hubli"], "name": "Karnataka"}}, {"state": {"states": ["Raipur", "Durg"], "name": "Chhattisgarh"}}]' ...
[ "evaluate.evaluate" ]
[((324, 338), 'evaluate.evaluate', 'evaluate', (['data'], {}), '(data)\n', (332, 338), False, 'from evaluate import evaluate\n')]
from django.contrib.auth import login from evaluate.models import Result from django.shortcuts import render, redirect from django.contrib.auth.decorators import login_required from evaluate.forms import TestCreateForm def findEXT(a): if a[-4] == '.': return a[-3:] if a[-5] == '.': return a[-4:...
[ "evaluate.models.Result.objects.get", "evaluate.forms.TestCreateForm" ]
[((1178, 1230), 'django.shortcuts.render', 'render', (['request', '"""evaluate/createTest.html"""', 'context'], {}), "(request, 'evaluate/createTest.html', context)\n", (1184, 1230), False, 'from django.shortcuts import render, redirect\n'), ((1289, 1314), 'evaluate.models.Result.objects.get', 'Result.objects.get', ([]...
import os import cv2 import torch from torch.utils.data import DataLoader import torch.optim as optim from torch.nn import CTCLoss from dataset import Synth90kDataset, synth90k_collate_fn from model import CRNN from evaluate import evaluate from config import train_config as config def train_batch(crnn, data, optim...
[ "evaluate.evaluate" ]
[((473, 519), 'torch.nn.functional.log_softmax', 'torch.nn.functional.log_softmax', (['logits'], {'dim': '(2)'}), '(logits, dim=2)\n', (504, 519), False, 'import torch\n'), ((642, 671), 'torch.flatten', 'torch.flatten', (['target_lengths'], {}), '(target_lengths)\n', (655, 671), False, 'import torch\n'), ((1508, 1604),...
from tqdm.auto import tqdm from crf import CRF import sys sys.path.append("../../../") from utils import * from evaluate import evaluation import os import pickle import pandas as pd import collections import csv import os import glob import pickle from enum import Enum from dataclasses import dataclass from functool...
[ "evaluate.evaluation.get_evaluation" ]
[((60, 88), 'sys.path.append', 'sys.path.append', (['"""../../../"""'], {}), "('../../../')\n", (75, 88), False, 'import sys\n'), ((645, 665), 'torch.manual_seed', 'torch.manual_seed', (['(0)'], {}), '(0)\n', (662, 665), False, 'import torch\n'), ((5274, 5321), 'os.path.join', 'os.path.join', (['exp_name', 'run_name', ...
#!/usr/bin/env python import sys import os import config import evaluate import msd_parse from pyspark.mllib.recommendation import MatrixFactorizationModel, ALS def main(): import configspark sc = configspark.SPARK_CONTEXT # user/song string ID to int ID mappings full_text = sc.textFile(config.MSD_...
[ "evaluate.evaluate_model" ]
[((399, 437), 'msd_parse.get_user_song_maps', 'msd_parse.get_user_song_maps', (['full_raw'], {}), '(full_raw)\n', (427, 437), False, 'import msd_parse\n'), ((599, 651), 'msd_parse.replace_raw_ids', 'msd_parse.replace_raw_ids', (['test_parsed', 'users', 'songs'], {}), '(test_parsed, users, songs)\n', (624, 651), False, ...
import argparse import itertools import os.path import os import time import logging import datetime import torch import torch.optim.lr_scheduler import numpy as np import os import evaluate import trees import vocabulary import nkutil from tqdm import tqdm import SAPar_model import random tokens = SAPar_model from ...
[ "evaluate.evalb" ]
[((950, 2051), 'nkutil.HParams', 'nkutil.HParams', ([], {'max_len_train': '(0)', 'max_len_dev': '(0)', 'sentence_max_len': '(300)', 'learning_rate': '(0.0008)', 'learning_rate_warmup_steps': '(160)', 'clip_grad_norm': '(0.0)', 'step_decay': '(True)', 'step_decay_factor': '(0.5)', 'step_decay_patience': '(5)', 'max_cons...
import numpy as np import sys import evaluate as EV matrix_list = ["all", "1_syllable", "2_syllable", "3_syllable", "4_syllable", "2-_syllable"] def average_cost(costlist, querylist): querylist_uniq = [] cost_dict = {} for i in range(len(querylist)): query_id = querylist[i].strip() keyword...
[ "evaluate.evaluate", "evaluate.build_occurance_dict", "evaluate.build_syllable_num_dict" ]
[((1009, 1045), 'evaluate.build_occurance_dict', 'EV.build_occurance_dict', (['sys.argv[4]'], {}), '(sys.argv[4])\n', (1032, 1045), True, 'import evaluate as EV\n'), ((1070, 1109), 'evaluate.build_syllable_num_dict', 'EV.build_syllable_num_dict', (['sys.argv[5]'], {}), '(sys.argv[5])\n', (1096, 1109), True, 'import eva...
import os import json import matplotlib matplotlib.use('TkAgg') import matplotlib.pyplot as plt from evaluate import evaluate if __name__ == '__main__': training_commands, predict_commands = [], [] choices = {"dan": range(1, 4+1), "gru": range(1, 4+1)} probing_accuracies = {"dan": [], "gru": []} fo...
[ "evaluate.evaluate" ]
[((40, 63), 'matplotlib.use', 'matplotlib.use', (['"""TkAgg"""'], {}), "('TkAgg')\n", (54, 63), False, 'import matplotlib\n'), ((4146, 4180), 'matplotlib.pyplot.style.use', 'plt.style.use', (['"""seaborn-whitegrid"""'], {}), "('seaborn-whitegrid')\n", (4159, 4180), True, 'import matplotlib.pyplot as plt\n'), ((447, 517...
import os import pickle import argparse import torch from torch.utils.data import Dataset, DataLoader from data_processing import prep_dataset, build_second_order_wikidata_graphs from stockdataset import StockDataset from model import MANSF from train import train_model from evaluate import evaluate_model STOCKNET_REP...
[ "evaluate.evaluate_model" ]
[((1095, 1337), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Preprocess data, train or evaluate the MAN-SF model. If executed with --train or --evaluate, --model is required. If executed with --preprocess_in, --preprocess_out is required."""'}), "(description=\n 'Pre...
import pandas as pd from evaluate.calculator import ( RecallCalculator, PrecisionCalculator, EmptyReportError, ) import pytest from unittest.mock import patch, Mock from evaluate.report import ( Report, PrecisionReport, RecallReport ) from tests.common import create_precision_report_row from io ...
[ "evaluate.calculator.RecallCalculator", "evaluate.calculator.RecallCalculator._calculate_info_wrt_truth_probes", "evaluate.calculator.RecallCalculator._calculate_info_wrt_variants", "evaluate.report.PrecisionReport", "evaluate.calculator.PrecisionCalculator" ]
[((4740, 5000), 'unittest.mock.patch.object', 'patch.object', (['Report', 'Report.get_classifications_as_list.__name__'], {'return_value': "['unmapped', 'partially_mapped', 'primary_correct', 'primary_incorrect',\n 'secondary_correct', 'secondary_incorrect', 'supplementary_correct',\n 'supplementary_incorrect']"}...
from data import build_corpus from evaluate import Metrics from models.hmm import HMMModel from models.crf import CRFModel from models.lstm import BiLSTM from models.BiLSTM_CRF import BiLSTM_CRF from utils import * def main(): print('读取数据...') train_word_lists, train_tag_lists, word2id, tag2id = ...
[ "evaluate.Metrics" ]
[((320, 341), 'data.build_corpus', 'build_corpus', (['"""train"""'], {}), "('train')\n", (332, 341), False, 'from data import build_corpus\n'), ((379, 416), 'data.build_corpus', 'build_corpus', (['"""dev"""'], {'maek_vocab': '(False)'}), "('dev', maek_vocab=False)\n", (391, 416), False, 'from data import build_corpus\n...