prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
"""Technical analysis on a trading Pandas DataFrame"""
from numpy import floor
from re import compile
from numpy import maximum, mean, minimum, nan, ndarray, round
from numpy import sum as np_sum
from numpy import where
from pandas import DataFrame, Series
from statsmodels.tsa.statespace.sarimax import SARIMAX
clas... | DataFrame() | pandas.DataFrame |
import numpy as np
import pandas as pd
from scipy.stats import mode
from sklearn.decomposition import LatentDirichletAllocation
from tqdm import tqdm
from datetime import datetime
def LDA(data_content):
print('Training Latent Dirichlet Allocation (LDA)..', flush=True)
lda = LatentDirichletAllocation(n_compo... | pd.merge(df, data_content.bikers_df, on='biker_id', how='left') | pandas.merge |
#Author: <NAME>. Email: <EMAIL>
#Packaged by: <NAME>. Email: <EMAIL>
#This work is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
#To view a copy of this license, visit http://creativecommons.org/licenses/by-nc-sa/4.0/ or send a letter to Creative Commons,
#PO Box... | pd.DataFrame({'Probability':P_dict,'Number of occurences':num_instances,'Sequences':motifs}) | pandas.DataFrame |
import pandas as pd
import numpy as np
from newsapi import NewsApiClient
from requests import get
from requests.exceptions import RequestException
from contextlib import closing
from bs4 import BeautifulSoup
from os import path
from .. import app, db
from .news_formatting import NewsFormatting
"""
This file contain... | pd.DataFrame.from_dict(all_articles) | pandas.DataFrame.from_dict |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
# This file is part of CbM (https://github.com/ec-jrc/cbm).
# Author : <NAME>
# Credits : GTCAP Team
# Copyright : 2021 European Commission, Joint Research Centre
# License : 3-Clause BSD
import json
import numpy as np
import pandas as pd
import matplotlib.pyplot ... | pd.read_csv(file_ts, index_col=0) | pandas.read_csv |
#!/usr/bin/python3
# -*- coding: utf-8 -*-
# *****************************************************************************/
# * Authors: <NAME>
# *****************************************************************************/
"""transformCSV.py
This module contains the basic functions for creating the content of... | pandas.StringDtype() | pandas.StringDtype |
import json
import math
import operator
import warnings
import numpy as np
import pandas as pd
warnings.simplefilter(action='ignore', category=FutureWarning)
warnings.simplefilter(action='ignore', category=np.RankWarning)
np.seterr(divide='ignore', invalid='ignore')
def ema(series, period):
values = np.zeros(le... | pd.Timedelta(period) | pandas.Timedelta |
#
# Copyright 2018 Quantopian, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in wr... | pd.Timestamp(self.minutes_nanos[0], tz=UTC) | pandas.Timestamp |
import warnings
import geopandas as gpd
import numpy as np
import pandas as pd
from shapely.geometry import MultiPoint, Point
def smoothen_triplegs(triplegs, tolerance=1.0, preserve_topology=True):
"""
Reduce number of points while retaining structure of tripleg.
A wrapper function using shapely.simplif... | pd.concat((trips, sp_tpls_only_act, gaps, user_change), axis=0, ignore_index=True) | pandas.concat |
""" test the scalar Timestamp """
import pytz
import pytest
import dateutil
import calendar
import locale
import numpy as np
from dateutil.tz import tzutc
from pytz import timezone, utc
from datetime import datetime, timedelta
import pandas.util.testing as tm
import pandas.util._test_decorators as td
from pandas.ts... | tm.get_locales() | pandas.util.testing.get_locales |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
from ddf_library.context import COMPSsContext
from ddf_library.utils import generate_info
from pycompss.api.task import task
import pandas as pd
import numpy as np
import time
from pandas.testing import assert_frame_equal
@task(returns=2)
def _generate_partition(size,... | pd.DataFrame([['b', 3], ['a', 1]], columns=cols) | pandas.DataFrame |
import pkg_resources
from unittest.mock import sentinel
import pandas as pd
import pytest
import osmo_jupyter.dataset.combine as module
@pytest.fixture
def test_picolog_file_path():
return pkg_resources.resource_filename(
"osmo_jupyter", "test_fixtures/test_picolog.csv"
)
@pytest.fixture
def test_... | pd.to_datetime("2022") | pandas.to_datetime |
#!/usr/bin/env python
# inst: university of bristol
# auth: <NAME>
# mail: <EMAIL> / <EMAIL>
import os
import shutil
from glob import glob
import zipfile
import numpy as np
import pandas as pd
import gdalutils
from osgeo import osr
def _secs_to_time(df, date1):
df = df.copy()
conversion = 86400 # 86400s =... | pd.concat([bdy, df[0]], axis=1) | pandas.concat |
# -*- coding: utf-8 -*-
"""
Created on Fri Oct 5 10:04:19 2018
@author: <NAME> and <NAME>
--Collector Probe W Analysis and Mapping--
"""
# set imports
import os
import pandas as pd
import numpy as np
#import matplotlib.pyplot as plt
# define file directory and file name
os.chdir(r'C:\Users\jduran2\Documents\Pyth... | pd.DataFrame() | pandas.DataFrame |
#!/usr/bin/env python
# -*- coding: utf_8 -*
#
# Script to reformat xlsx tables from Akamai State of the Internet into a single SQL table
#
# <NAME> 2014
# <EMAIL>
# requires pandas, xlrd
import pandas as pd
import sqlite3
def importdata(infile,table,connection):
# Get list of sheets or count of sheets
xfile... | pd.ExcelFile(infile) | pandas.ExcelFile |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
#
# Copyright (c) 2021 snaketao. All Rights Reserved
#
# @Version : 1.0
# @Author : snaketao
# @Time : 2021-10-21 12:21
# @FileName: insert_mongo.py
# @Desc : insert data to mongodb
import appbk_mongo
import pandas as pd
#数据处理,构造一个movies对应多个tagid的字典,并插入 mongodb 的mo... | pd.merge(grouped, file3, how='inner', on='tagId',left_index=False, right_index=False, sort=False,suffixes=('_x', '_y'), copy=True) | pandas.merge |
# -*- coding: utf-8 -*-
""" Simple multi-area model for Nordic electricity market
Created on Wed Jan 16 11:31:07 2019
@author: elisn
Notes:
1 - For conversion between dates (YYYYMMDD:HH) and weeks (YYYY:WW) weeks are counted as starting during the first hour
in a year and lasting 7 days, except for the last week wh... | pd.DataFrame(0.0,index=self.timerange_p1,columns=self.solar_areas) | pandas.DataFrame |
# -*- coding: utf-8 -*-
from copy import deepcopy
import warnings
from itertools import chain, combinations
from collections import Counter
from typing import Dict, Iterable, Iterator, List, Optional, Tuple, Union
import numpy as np
import pandas as pd
from scipy.stats import (pearsonr as pearsonR,
... | pd.concat([preserved, active, inactive]) | pandas.concat |
# -*- coding: utf-8 -*-
"""Tracker REITs (Real Estate Investment Trust) investments."""
import datetime
import os
from concurrent.futures import ProcessPoolExecutor
import numpy as np
import pandas as pd
import requests
import requests_cache
from .portutils import NUM_PROCESS, Singleton, UnitsTransactions
# Cac... | pd.json_normalize(df_tmp.dividends) | pandas.json_normalize |
from abc import abstractmethod, ABC
from pandas import DataFrame
from database.db_reader import read_table_header
class DataSource(ABC):
@staticmethod
def convert_header(table_name, dataframe):
columns = read_table_header(table_name)
dataframe.columns = columns
return dataframe
@... | DataFrame() | pandas.DataFrame |
from PyDSS.pyContrReader import pySubscriptionReader as pySR
from PyDSS.pyLogger import getLoggerTag
from PyDSS.unitDefinations import type_info as Types
from PyDSS.unitDefinations import unit_info as Units
from PyDSS.pyContrReader import pyExportReader as pyER
from PyDSS import unitDefinations
from PyDSS.exceptions im... | pd.MultiIndex.from_tuples(tuples, names=['timestamp', 'frequency', 'Simulation mode']) | pandas.MultiIndex.from_tuples |
"""ops.syncretism.io model"""
__docformat__ = "numpy"
import configparser
import logging
from typing import Tuple
import pandas as pd
import requests
import yfinance as yf
from gamestonk_terminal.decorators import log_start_end
from gamestonk_terminal.rich_config import console
from gamestonk_terminal.stocks.options... | pd.to_datetime(entry["timestamp"], unit="s") | pandas.to_datetime |
import os
import sys
import numpy as np
import pandas as pd
from pycompss.api.api import compss_wait_on
from pycompss.api.task import task
from data_managers.fundamentals_extraction import FundamentalsCollector
from data_managers.price_extraction import PriceExtractor
from data_managers.sic import load_sic
from model... | pd.cut(r.y, bins) | pandas.cut |
__all__ = [
'PrettyPachydermClient'
]
import logging
import re
from typing import Dict, List, Iterable, Union, Optional
from datetime import datetime
from dateutil.relativedelta import relativedelta
import pandas.io.formats.style as style
import pandas as pd
import numpy as np
import yaml
from IPython.core.displa... | pd.isna(x) | pandas.isna |
import numpy as np
import pandas as pd
from typing import List
from astropy.time import Time
from .orbits import Orbits
from .utils import assignPatchesSquare
__all__ = [
"findAverageOrbits",
"findTestOrbitsPatch",
"selectTestOrbits",
]
def findAverageOrbits(
ephemeris: pd.DataFrame,
orbi... | pd.DataFrame() | pandas.DataFrame |
# -*- coding:utf-8 -*-
# /usr/bin/env python
"""
Date: 2021/7/8 22:08
Desc: 金十数据中心-经济指标-美国
https://datacenter.jin10.com/economic
"""
import json
import time
import pandas as pd
import demjson
import requests
from akshare.economic.cons import (
JS_USA_NON_FARM_URL,
JS_USA_UNEMPLOYMENT_RATE_URL,
JS_USA_EIA_... | pd.to_datetime(temp_se.iloc[:, 0]) | pandas.to_datetime |
import nltk
import numpy as np
import pandas as pd
import bokeh as bk
from math import pi
from collections import Counter
from bokeh.transform import cumsum
from bokeh.palettes import Category20c
from bokeh.models.glyphs import VBar
from bokeh.models import ColumnDataSource, DataRange1d, Plot, LinearAxis, Grid
from bok... | pd.concat([analysis_df, temp], sort=True) | pandas.concat |
import pandas as pd
# import copy
from pathlib import Path
import pickle
pd.set_option('display.max_colwidth', -1)
pd.options.display.max_rows = 999
pd.options.mode.chained_assignment = None
import numpy as np
import math
import seaborn as sns
import matplotlib.pyplot as plt
import matplotlib.patches as mpatches
from s... | pd.concat(ldf, keys=keys_ldf) | pandas.concat |
from __future__ import division
'''
NeuroLearn Statistics Tools
===========================
Tools to help with statistical analyses.
'''
__all__ = ['pearson',
'zscore',
'fdr',
'holm_bonf',
'threshold',
'multi_threshold',
'winsorize',
'trim... | pd.Series(index=cutoff['std'], data=std) | pandas.Series |
import common, cost, fetchresults
import logging, sys, datetime
import numpy as np
import pandas as pd
import matplotlib
import matplotlib.pyplot as plt
matplotlib.style.use('ggplot')
import pickle
logging.basicConfig(level=logging.INFO)
#logging.basicConfig(level=logging.DEBUG)
if 1:
results = fetchresults.get_a... | pd.DataFrame(compute_hours_all, index=date) | pandas.DataFrame |
# -*- coding: utf-8 -*-
import pandas.util.testing as tm
from pandas.tseries import offsets
from pandas._libs.tslibs.frequencies import (get_rule_month,
_period_str_to_code,
INVALID_FREQ_ERR_MSG,
... | offsets.Minute() | pandas.tseries.offsets.Minute |
import numpy as np
import pandas as pd
import requests
from bs4 import BeautifulSoup
from Helpers import find_between
class CJH_Archives:
def __init__(self, repo, url=False):
self.repo = repo
self.url = url
def get_meta_data(self, object_type, page_to_start_at, maximum_pages_to_scrape):
... | pd.DataFrame.from_records(tupleList, columns=['Names', 'Link', 'Location']) | pandas.DataFrame.from_records |
import pandas as pd
import numpy as np
import math
import os
from scipy.interpolate import interp1d
import time
from sklearn.ensemble import RandomForestRegressor
import xgboost as xgb
from lightgbm import LGBMRegressor
from catboost import CatBoostRegressor
from information_measures import *
from joblib import Para... | pd.DataFrame(all_time_ids_byStock,columns=['time_id']) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
Created on Wed Oct 27 01:31:54 2021
@author: yoonseok
"""
import os
import pandas as pd
from tqdm import tqdm
from scipy.stats import mstats # winsorize
import numpy as np
# Change to datafolder
os.chdir(r"C:\data\car\\")
# 기본 테이블 입력
df = pd.read_csv("dataframe_h1.txt")
del df["Unnamed:... | pd.merge(result, asset[["key", "asset"]], how="inner", on=["key"]) | pandas.merge |
import pytest
from mapping import mappings
from pandas.util.testing import assert_frame_equal, assert_series_equal
import pandas as pd
from pandas import Timestamp as TS
import numpy as np
from pandas.tseries.offsets import BDay
@pytest.fixture
def dates():
return pd.Series(
[TS('2016-10-20'), TS('2016-11... | pd.DataFrame([[1.0, 0], [0, 1.0]], index=idx, columns=cols) | pandas.DataFrame |
import logging
l = logging.getLogger("abg")
import flask
from flask import Blueprint, flash, redirect, render_template, request, url_for
from flask_login import login_required, login_user, logout_user
from flask import Markup
from flask import send_file
from flask import abort
l.error("flask")
from abg_stats.extension... | pd.concat([player_winner, player_loser]) | pandas.concat |
# Copyright 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | testing.assert_frame_equal(output, expected) | pandas.testing.assert_frame_equal |
import pandas as pd
import numpy as np
file1 = '../data/STRIDE_PATIENT.xlsx'
x1 = pd.ExcelFile(file1)
stride_patient = x1.parse('Sheet1')
file2 = '../data//SURGERY.xlsx'
x2 = | pd.ExcelFile(file2) | pandas.ExcelFile |
# general
import math
import logging
import json
import os,sys
from pIMZ.regions import SpectraRegion
import random
from collections import defaultdict, Counter
import glob
import shutil, io, base64
# general package
from natsort import natsorted
import pandas as pd
import numpy as np
from numpy.ctypeslib import ndpo... | pd.DataFrame() | pandas.DataFrame |
# -*- coding: utf-8 -*-
# ---
# jupyter:
# jupytext:
# formats: ipynb,py:light
# text_representation:
# extension: .py
# format_name: light
# format_version: '1.5'
# jupytext_version: 1.4.2
# kernelspec:
# display_name: Python 3
# language: python
# name: python3
# ---
#... | pd.read_sql("SELECT * FROM processed", DB.conn) | pandas.read_sql |
import numpy as np
import pandas as pd
from sklearn.preprocessing import LabelEncoder
from sklearn.compose import ColumnTransformer
from sklearn.preprocessing import OneHotEncoder
from sklearn.impute import SimpleImputer
rand_dataset = | pd.read_csv("sample_submission.csv") | pandas.read_csv |
import os
import numpy as np
import pandas as pd
import tsplib95
import networkx as nx
from tqdm import tqdm
import sys
import re
def prepare_testset_FINDER(data_dir, scale_factor=0.000001):
graph_list = []
atoi = lambda text : int(text) if text.isdigit() else text
natural_keys = lambda text : [atoi(c) fo... | pd.read_csv(f'{result_dir}/{f}', index_col=0) | pandas.read_csv |
import re
import os
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
import statsmodels.api as sa
import statsmodels.formula.api as sfa
import scikit_posthocs as sp
import networkx as nx
from loguru import logger
from GEN_Utils import FileHandling
from ... | pd.merge(cluster_summary, inter_vs_intra, on='cluster_filter_type') | pandas.merge |
import re
import json
from itertools import filterfalse
from luigi import Task
from luigi.task import logger as luigi_logger
import pandas as pd
from lxml import objectify
from datapackage import Package
from ..utils import TargetOutput
from ..utils import SaltedOutput
from ..utils import SuffixPreservingLocalTarget... | pd.DataFrame(data=as_lists, columns=columns) | pandas.DataFrame |
# coding: utf-8
# In[ ]:
import re
import sys
import os
import argparse
import pandas as pd
from pprint import pprint
from collections import OrderedDict
from operator import itemgetter
import json
def main():
# Parse args
args = parse_args()
# read manifest + EFO mappings
FINNGEN_EFO=pd.read_csv... | pd.read_json(args.in_manifest, lines=True) | pandas.read_json |
"""
"""
import pandas as pd
import os, sys
import numpy as np
# tie-breaker for RQ1
TIE = 'avg' #'MAX'
def read_and_add_flag(filename):
"""
"""
df = pd.read_csv(filename)
indices = df.index.values
df['flag'] = df.true == df.pred
return df
def compute_acc(init_pred_df, aft_pred_df, classes):
"""
"""
from sk... | pd.DataFrame(data = {
'true':init_pred_df.true.values,
'pred':init_pred_df.pred.values,
'new_pred':aft_pred_df.pred.values,
'init_flag':init_pred_df.flag}) | pandas.DataFrame |
from flask import Flask, render_template, request, redirect, url_for, session
import pandas as pd
import pymysql
import os
import io
#from werkzeug.utils import secure_filename
from pulp import *
import numpy as np
import pymysql
import pymysql.cursors
from pandas.io import sql
#from sqlalchemy import create... | pd.DataFrame(result) | pandas.DataFrame |
import pandas as pd
import itertools
class Aprioripy:
def __init__(self,
table,
convert=True,
items=[],
excluded_items=[],
positive_label=1):
def converter(table):
item_list = list(set(sum([x.split(", ")... | pd.DataFrame(new_table) | pandas.DataFrame |
from cmath import nan
from tokenize import endpats
from matplotlib import pyplot as plt
from mpl_toolkits.mplot3d.art3d import Poly3DCollection, Line3DCollection
import chart_studio.plotly as py
import plotly.graph_objects as go
from plotly.offline import download_plotlyjs, init_notebook_mode, plot, iplot
import plotly... | pd.DataFrame(geodesic_points, columns = ['x1', 'x2','x3','x4']) | pandas.DataFrame |
import h5py
from pathlib import Path
from typing import Union, Tuple
import pickle
import json
import os
import gc
from tqdm import tqdm
import numpy as np
import pandas as pd
# TODO output check, verbose
def load_all_libsdata(path_to_folder: Union[str, Path]) -> Tuple[pd.DataFrame, list, pd.Series]:
"""
Functio... | pd.Series(samples) | pandas.Series |
#%%
# CARGO LOS DATASETS
import pandas as pd
import numpy as np
from shapely.geometry import Point
import shapely as shp
import geopandas as gpd
from geopandas.array import points_from_xy
path = "merged1_listas.pkl"
df_merge1 = pd.read_pickle(path)
df_merge1.reset_index(inplace=True)
#%%
#region PART 1... | pd.read_pickle('paso1.pkl') | pandas.read_pickle |
#!/usr/bin.env/python
# -*- coding: utf-8 -*-
"""
Gates are traditionally used to subset single cell data in one
or two dimensional space by hand-drawn polygons in a manual and laborious
process. cytopy attempts to emulate this using autonomous gates, driven
by unsupervised learning algorithms. The gate module contains... | pd.concat(data) | pandas.concat |
import time
import datetime
import numpy as np
import pandas as pd
import lightgbm as lgb
from dateutil.parser import parse
from sklearn.cross_validation import KFold
from sklearn.metrics import mean_squared_error
data_path = './'
train = pd.read_csv(data_path + 'f_train_20180204.csv', encoding='gb2312')
test = | pd.read_csv(data_path + 'f_test_a_20180204.csv', encoding='gb2312') | pandas.read_csv |
########################################################################
#
# Functions for loading financial data.
#
# Data from several different files are combined into a single
# Pandas DataFrame for each stock or stock-index.
#
# The price-data is read from CSV-files from Yahoo Finance.
# Other financial data (Sale... | pd.read_csv(path, sep=",", parse_dates=[3], index_col=3) | pandas.read_csv |
from itertools import groupby, zip_longest
from fractions import Fraction
from random import sample
import json
import pandas as pd
import numpy as np
import music21 as m21
from music21.meter import TimeSignatureException
m21.humdrum.spineParser.flavors['JRP'] = True
from collections import defaultdict
#song has no ... | pd.isna(ix) | pandas.isna |
import numpy as np
import pandas as pd
import tensorflow as tf
from shfl.data_base.data_base import DataBase
from shfl.data_distribution.data_distribution_iid import IidDataDistribution
class TestDataBase(DataBase):
def __init__(self):
super(TestDataBase, self).__init__()
def load_data(self):
... | pd.testing.assert_frame_equal(all_data, train_data.iloc[all_data.index.values]) | pandas.testing.assert_frame_equal |
from __future__ import print_function, absolute_import
import unittest, math
import pandas as pd
import numpy as np
from . import *
class T(base_pandas_extensions_tester.BasePandasExtensionsTester):
def test_concat(self):
df = pd.DataFrame({'c_1':['a', 'b', 'c'], 'c_2': ['d', 'e', 'f']})
df.en... | pd.DataFrame({'n_1': [1, 2, 3], 'n_2': [4, 5, 6]}) | pandas.DataFrame |
"Test suite of AirBnbModel.source.processing module"
import numpy as np
import pandas as pd
import pytest
from pandas._testing import assert_index_equal
from AirBnbModel.source.processing import intersect_index
class TestIntersectIndex(object):
"Test suite for intersect_index method"
def test_first_input_n... | pd.Series(data=[1, 2, 3, 4], index=["foo", "bar", "bar", np.nan]) | pandas.Series |
"""Run unit tests.
Use this to run tests and understand how tasks.py works.
Example:
Create directories::
mkdir -p test-data/input
mkdir -p test-data/output
Run tests::
pytest test_combine.py -s
Notes:
* this will create sample csv, xls and xlsx files
* test_combine_(... | pd.concat([df1, df2, df3], join='inner') | pandas.concat |
import numpy as np
import pandas as pd
import pandas.util.testing as tm
import pandas.tseries.period as period
from pandas import period_range, PeriodIndex, Index, date_range
def _permute(obj):
return obj.take(np.random.permutation(len(obj)))
class TestPeriodIndex(tm.TestCase):
def setUp(self):
pa... | pd.period_range('1/1/2000', freq='D', periods=5) | pandas.period_range |
import logging
import pandas as pd
import time
import wikifier
import typing
import hashlib
import json
import os
from .parser_base import ParserBase, PreParsedResult
from pandas.util import hash_pandas_object
from datamart_isi.materializers.general_materializer import GeneralMaterializer
from datamart_isi.materializ... | pd.read_csv(loaded_data, dtype=str) | pandas.read_csv |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Wed Jun 27 09:20:01 2018
@authors: <NAME>
Last modified: 2020-02-19
------------------------------------------
** Semantic Search Analysis: Start-up **
------------------------------------------
This script: Import search queries from Google Analytics, ... | pd.Series(foreignNo) | pandas.Series |
'''
Pulls data from xml and creates an array for each user consisting of PMID,
type, and annotation. Uses NLTK scoring metrics tools to determine
precision, recall, and f-score. By including PMID in the hash, this version
allows for examining user to user comparisons across multiple documents in the
... | pd.DataFrame(all_users_arr, columns=('user_id', 'pairings', 'total_f')) | pandas.DataFrame |
import pandas as pd
import numpy as np
import torch
import os.path
from glob import glob
from datetime import datetime
from base.torchvision_dataset import TorchvisionDataset
from torch.utils.data import TensorDataset
class HR_Dataset(TorchvisionDataset):
def __init__(self, root:str, normal_class):
supe... | pd.read_csv("/workspace/dim3_train.txt") | pandas.read_csv |
import pandas as pd
import numpy as np
import seaborn as sns
from scipy import stats
import matplotlib.pyplot as plt
import os
import re
from sklearn.model_selection import train_test_split
import random
import scorecardpy as sc
# split train into train data and test data
# os.chdir(r'D:\GWU\Aihan\DATS 6103 Data Mini... | pd.concat([X_train, y_train], axis=1) | pandas.concat |
import time
import numpy as np
import pandas as pd
import json
import matplotlib.pyplot as plt
class CFE():
def __init__(self, cfg_file=None):
super(CFE, self).__init__()
self.cfg_file = cfg_file
############################################################
# _______... | pd.Timestamp(year=1970, month=1, day=1, hour=0) | pandas.Timestamp |
import copy
import os
from functools import partial
import joblib
import numpy as np
import optuna
import pandas as pd
import lightgbm as lgbm
from .enums import ProblemType
from .logger import logger
from .metrics import Metrics
from .params import get_params
optuna.logging.set_verbosity(optuna.logging.INFO)
def... | pd.DataFrame.from_dict(final_valid_predictions, orient="index") | pandas.DataFrame.from_dict |
# Copyright 1999-2021 Alibaba Group Holding Ltd.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or a... | pd.testing.assert_series_equal(result, expected) | pandas.testing.assert_series_equal |
#coding:utf-8
import pandas as pd
import numpy as np
# 读取个人信息
train_agg = pd.read_csv('../data/train_agg.csv',sep='\t')
test_agg = pd.read_csv('../data/test_agg.csv',sep='\t')
agg = | pd.concat([train_agg,test_agg],copy=False) | pandas.concat |
# -*- coding: utf-8 -*-
from __future__ import print_function
from datetime import datetime, time
from numpy import nan
from numpy.random import randn
import numpy as np
from pandas import (DataFrame, Series, Index,
Timestamp, DatetimeIndex,
to_datetime, date_range)
import pa... | pd.Timedelta('00:01:00') | pandas.Timedelta |
# 从Binance币安在线api下载1分钟k线,进行回测
import requests
import backtrader as bt
import backtrader.analyzers as btanalyzers
import json
import pandas as pd
import datetime as dt
import matplotlib.pyplot as plt
def get_binance_bars(symbol, interval, startTime, endTime):
url = "https://api.binance.com/api/v3/klines"
... | pd.concat(df_list) | pandas.concat |
'''
@Author: mendeslbruno
Date: 2021-01-26
Descr: Performs some simple analyzes for several actions of the index SP500.
'''
import pandas as pd
import yfinance as yf
import streamlit as st
import datetime as dt
import plotly.graph_objects as go
from plotly.subplots import make_subplots
snp500 = pd.read_csv("datasets/... | pd.DataFrame(data=marketInfo, index=[0]) | pandas.DataFrame |
# ----------------------------------------------------------------------------
# Copyright (c) 2013--, scikit-bio development team.
#
# Distributed under the terms of the Modified BSD License.
#
# The full license is in the file COPYING.txt, distributed with this software.
# --------------------------------------------... | pd.DataFrame({'foo': [0], 'bar': [100]}) | pandas.DataFrame |
import pandas as pd
import numpy as np
from scipy import signal
import os
def get_timedeltas(login_timestamps, return_floats=True):
"""
Helper function that returns the time differences (delta t's) between consecutive logins for a user.
We just input the datetime stamps as an index, hence this method... | pd.Series(timedelta_sample) | pandas.Series |
import pandas as pd
import numpy as np
import torch
from scipy.io import arff
from abc import ABC, abstractmethod
from torch.utils.data import DataLoader, TensorDataset
class BaseADDataset(ABC):
"""Anomaly detection dataset base class."""
def __init__(self, root: str):
super().__init__()
self... | pd.DataFrame(loaded[0]) | pandas.DataFrame |
# Copyright (c) 2021 ING Wholesale Banking Advanced Analytics
#
# Permission is hereby granted, free of charge, to any person obtaining a copy of
# this software and associated documentation files (the "Software"), to deal in
# the Software without restriction, including without limitation the rights to
# use, copy, mo... | pd.Series(data=obj, index=df.index) | pandas.Series |
"""
Unit test of Inverse Transform
"""
import unittest
import pandas as pd
import numpy as np
import category_encoders as ce
import catboost as cb
import sklearn
import lightgbm
import xgboost
from shapash.utils.transform import inverse_transform, apply_preprocessing, get_col_mapping_ce
class TestInverseTransformCate... | pd.testing.assert_frame_equal(train, original) | pandas.testing.assert_frame_equal |
#%%
import os
from pyteomics import mzid, mzml
import pandas as pd
import numpy as np
import glob
"""
Files are downloaded and manually randomly divided into different folders
the following code is repeated but has the same effect, it is applied to various folders to
generate pandas data frames and to store all the d... | pd.DataFrame({'file':mzml_location,'id':ids,'mz':mz,'intensities':intensities}) | pandas.DataFrame |
import os
from io import StringIO
from typing import List
import matplotlib as mpl
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
import roadrunner
import seaborn as sns
import tellurium as te
from tellurium.roadrunner.extended_roadrunner import ExtendedRoadRunner
from sres import SRES
mpl.us... | pd.DataFrame(dataValues, columns=["time", "L", "E", "P", "R"]) | pandas.DataFrame |
# flake8: noqa
import os
from carla import log
os.environ["TF_CPP_MIN_LOG_LEVEL"] = "3"
import warnings
import pandas as pd
warnings.simplefilter(action="ignore", category=FutureWarning)
import argparse
from typing import Dict, Optional
import numpy as np
import yaml
from tensorflow import Graph, Session
from c... | pd.concat([results, df_benchmark], axis=0) | pandas.concat |
# -*- coding: utf-8 -*-
import os
import numpy as np
import pandas as pd
from sqlalchemy import create_engine
from tablizer.inputs import Inputs, Base
from tablizer.defaults import Units, Methods, Fields
from tablizer.tools import create_sqlite_database, check_inputs_table, insert, \
make_session, check_existing_r... | pd.to_datetime(date) | pandas.to_datetime |
import os
import copy
import pytest
import numpy as np
import pandas as pd
import pyarrow as pa
from pyarrow import feather as pf
from pyarrow import parquet as pq
from time_series_transform.io.base import io_base
from time_series_transform.io.numpy import (
from_numpy,
to_numpy
)
from time_series_transfor... | pd.DataFrame(expect_collection_expandTime['remove']) | pandas.DataFrame |
import threading
import time
import datetime
import pandas as pd
from functools import reduce, wraps
from datetime import datetime, timedelta
import numpy as np
from scipy.stats import zscore
import model.queries as qrs
from model.NodesMetaData import NodesMetaData
import utils.helpers as hp
from utils.helpers import... | pd.merge(result, grouped, on=['site', 'lat', 'lon'], how='outer') | pandas.merge |
# flu prediction
import os
import pandas as pd
import feather
from utils.fastai.structured import *
from utils.fastai.column_data import *
from sklearn import preprocessing
from sklearn.metrics import classification_report, confusion_matrix
import keras
from keras.layers import Input, Embedding, Dense, Dropout
from ke... | pd.read_feather(data_path + 'joined_df.feather') | pandas.read_feather |
# -*- coding: utf-8 -*-
import logging
import numpy as np
import pandas as pd
import pymc3 as pm
from scipy.stats import wilcoxon
from sklearn.metrics import log_loss
from sklearn.model_selection import StratifiedKFold
from sklearn.ensemble import GradientBoostingClassifier
from sklearn.base import BaseEstimator, Class... | pd.DataFrame(data=baseline_prediction) | pandas.DataFrame |
import os
import math
import copy
import random
import calendar
import csv
import pandas as pd
import numpy as np
import networkx as nx
import matplotlib
import matplotlib.pyplot as plt
import matplotlib.dates as mdates
import matplotlib.ticker as ticker
import sqlite3
import seaborn as sns
#from atnresilience import ... | pd.DataFrame() | pandas.DataFrame |
#code will get the proper values like emyield, marketcap, cacl, etc, and supply a string and value to put back into the dataframe.
import pandas as pd
import numpy as np
import logging
import inspect
from scipy import stats
from dateutil.relativedelta import relativedelta
from datetime import datetime
from scipy import... | pd.Series(ltcacls) | pandas.Series |
# -*- coding: utf-8 -*-
import numpy as np
import pandas as pd
import pytest
from toolz.dicttoolz import valmap
from kartothek.core.factory import DatasetFactory
from kartothek.io.eager import store_dataframes_as_dataset
def assert_index_dct_equal(dict1, dict2):
dict1 = valmap(sorted, dict1)
dict2 = valmap(s... | pd.DataFrame({"p": [1, 2]}) | pandas.DataFrame |
# Created by fw at 8/14/20
import torch
import numpy as np
import pandas as pd
import joblib
from torch.utils.data import Dataset as _Dataset
# from typing import Union,List
import lmdb
import io
import os
def get_dataset(cfg, city, dataset_type):
cfg = cfg.DATASET
assert city.upper() in ["BERLIN", "ISTANBU... | pd.to_datetime("2019-01-02") | pandas.to_datetime |
import pandas as pd
from collections import OrderedDict
import csv
print("Started.")
# define paths to your existing zooniverse classifications and metadata locations
# These are the files you get as output from the Jupyter Notebook convert_zooniverse.ipynb
classifs=pd.read_csv("../zooniverse_classifications/zoonive... | pd.read_csv("../metadata/filename_links.csv") | pandas.read_csv |
#!/usr/bin/env python
import numpy as np
import pandas as pa
import pb
from glccIndex import *
from collections import OrderedDict
def calc_area_from_LCCmatrix(veg5ini,list_LCCmatrix,area):
"""
Caculate the time series of area from initial vegetation
array (veg5ini) and a list of LCCmatrix.
P... | pa.DataFrame(glccReal_from_glccpftmtc,columns=['f','g','p','c'],index=['f','g','p','c']) | pandas.DataFrame |
import dash
import dash_core_components as dcc
import dash_html_components as html
from dash.dependencies import Input, Output
import plotly.express as px
import plotly.graph_objects as go
import pandas as pd
import geopandas as gpd
import numpy as np
# for debugging purposes
import json
external_stylesheets = ['style... | pd.merge(gdf, df, on="neighborhood code") | pandas.merge |
'''
@Author: your name
@Date: 2020-05-25 10:31:42
@LastEditTime: 2020-05-25 11:03:34
@LastEditors: Please set LastEditors
@Description: In User Settings Edit
@FilePath: /dataprocess/transform/colconcat-neu.py
'''
import pandas as pd
import argparse
import os
import shutil
import random
parser = argparse.ArgumentParse... | pd.read_csv(left_dataset) | pandas.read_csv |
#!/usr/bin/python3
import json
from datetime import date, timedelta
from operator import itemgetter
from pprint import pprint
from typing import Dict, Tuple
import jinja2
import pandas as pd
from matplotlib import pyplot as plt
from rpm import labelCompare
# start statistics at this date (used as starting point for ... | pd.DataFrame(avg_bl_durs, index=dates) | pandas.DataFrame |
import re
import numpy as np
import pandas.compat as compat
import pandas as pd
from pandas.compat import u
from pandas.core.base import FrozenList, FrozenNDArray
from pandas.util.testing import assertRaisesRegexp, assert_isinstance
from pandas import Series, Index, DatetimeIndex, PeriodIndex
from pandas import _np_ver... | FrozenList(self.lst + [1, 2, 3]) | pandas.core.base.FrozenList |
"""
lcopt.model
-----------
Module containing the LcoptModel class.
"""
from lcopt.io import *
#from lcopt.ipython_interactive import IFS
from lcopt.interact import FlaskSandbox
from lcopt.bw2_export import Bw2Exporter
from lcopt.analysis import Bw2Analysis
# This is a copy straight from bw2data.query, extracted so ... | pd.DataFrame(to_df) | pandas.DataFrame |
#! /usr/bin/env python
### Script containing varous plotting functions for splitting Measurements
import pandas as pd
import sys
import os
import shlex
from subprocess import call
import matplotlib.pyplot as plt
from matplotlib.gridspec import GridSpec
import matplotlib
import numpy as np
from stack import Stacker
from... | pd.read_csv(pdiff_l2,delim_whitespace=True,converters=date_time_convert) | pandas.read_csv |
import numpy as np
import pandas as pd
import os
DATA_FOLDER = 'data'
class Vehicle(object):
def __init__(self):
self.x = 0
self.y = 0
self.rides = []
def dist_to_ride(self, ride):
return np.abs(self.x - ride.start_x) + np.abs(self.y - ride.start_y)
def distance(coords1, coo... | pd.DataFrame(data=rides, columns=["start_x", "start_y", "end_x", "end_y", "min_start", "max_finish"]) | pandas.DataFrame |
from tkinter import *
import tkinter as tk
import os
import datetime as dt
from dateutil.relativedelta import relativedelta
import pytz
from openpyxl import load_workbook
import win32com.client
import win32api
import xlrd
from simple_salesforce import Salesforce
import pandas as pd
from openpyxl.styles import Font, Col... | pd.DataFrame(records3) | pandas.DataFrame |
import pdb
import unittest
import torch
import pandas as pd
import numpy as np
from agents.SACAgent import SACAgent
from cobs.model import Model
from test.test_config import state_name, sac_network_map, eplus_naming_dict, eplus_var_types, \
SatAction, BlindActionSingleZone, ThermActionSin... | pd.read_csv('test/agent_tests/saved_results/sac_blinds_multi_zone_single_stpt_obs.csv') | pandas.read_csv |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.