crossfile_context_retrievalwref
dict
prompt
stringlengths
252
32.6k
right_context
stringlengths
0
81.2k
metadata
dict
crossfile_context_retrieval
dict
groundtruth
stringlengths
5
208
{ "list": [ { "filename": "tests/test_prompt_tokenizers.py", "retrieved_chunk": " )\n def test_sharegpt_integration(self):\n with open(\n Path(__file__).parent / \"fixtures/conversation.json\", encoding=\"utf-8\"\n ) as fin:\n data = fin.read()\n ...
"""Module for testing dataset sequence packing""" import unittest from pathlib import Path from datasets import Dataset, load_dataset from transformers import AutoTokenizer from axolotl.datasets import ConstantLengthDataset, TokenizedPromptDataset from axolotl.prompt_tokenizers import AlpacaPromptTokenizingStrategy ...
constant_len_dataset = ConstantLengthDataset( self.tokenizer, [dataset], seq_length=2048, ) packed_dataset = Dataset.from_list(list(constant_len_dataset)) example = packed_dataset[0] next_bos_index = ( example["input_ids"][1:].ind...
{ "context_start_lineno": 0, "file": "tests/test_packed_dataset.py", "groundtruth_start_lineno": 41, "repository": "OpenAccess-AI-Collective-axolotl-2c37bf6", "right_context_start_lineno": 42, "task_id": "project_cc_python/2249" }
{ "list": [ { "filename": "tests/test_prompt_tokenizers.py", "retrieved_chunk": " ) as fin:\n data = fin.read()\n tokenized_conversation = json.loads(data)\n prompter = ShareGPTPrompter(\"chat\")\n strat = ShareGPTPromptTokenizingStrategy(\n prompt...
from_list(list(TokenizedPromptDataset(strat, dateset)))
{ "list": [ { "filename": "rdfreader/chem/mol.py", "retrieved_chunk": " A mol block string.\n Returns\n -------\n Molecule\n A Molecule object.\n \"\"\"\n mol = cls()\n mol._from_mol_block(mol_block, properties)\n return mol\n d...
from rdkit.Chem import Mol, MolFromMolBlock, MolToSmiles from rdfreader.chem.mol import Molecule def assert_molecule_from_mol_block(mol: Molecule, mol_block: str): """Assert that a molecule object created from a mol block string has the correct properties.""" assert mol.rd_mol is not None assert mol....
def test_molecule_to_rdkit_mol(sample_molecule, sample_mol_block): """Test the Molecule.rd_mol property.""" rd_mol: Mol = sample_molecule.rd_mol # can't directly compare Mol objects, so we'll just check that it is of # the right type and is not None assert rd_mol is not None assert isinstance...
{ "context_start_lineno": 0, "file": "test/test_molecule.py", "groundtruth_start_lineno": 30, "repository": "deepmatterltd-rdfreader-a811645", "right_context_start_lineno": 31, "task_id": "project_cc_python/2421" }
{ "list": [ { "filename": "rdfreader/chem/mol.py", "retrieved_chunk": " \"\"\"Returns True if the molecules are equal.\n Returns\n -------\n bool\n True if the molecules are equal.\n \"\"\"\n if not isinstance(__o, Molecule):\n return Fal...
mol_block is None
{ "list": [ { "filename": "GCoNet_plus/models/modules.py", "retrieved_chunk": " return x\nclass CoAttLayer(nn.Module):\n def __init__(self, channel_in=512):\n super(CoAttLayer, self).__init__()\n self.all_attention = eval(Config().relation_module + '(channel_in)')\n self...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F # import fvcore.nn.weight_init as weight_init from functools import partial from MCCL.config import Config config = Config() class ResBlk(nn.Module): def __init__(self, channel_in=64, channel_out=64, groups=0): super(...
channel_per_class = x.shape[0] // config.loadN x_per_class_corr_list = [] for idx in range(0, x.shape[0], channel_per_class): x_per_class = x[idx:idx+channel_per_class] x_new_per_class = self.all_attention(x_per_class) ...
{ "context_start_lineno": 0, "file": "MCCL/models/modules.py", "groundtruth_start_lineno": 102, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 103, "task_id": "project_cc_python/2404" }
{ "list": [ { "filename": "GCoNet_plus/models/modules.py", "retrieved_chunk": " def forward(self, x5):\n if self.training:\n f_begin = 0\n f_end = int(x5.shape[0] / 2)\n s_begin = f_end\n s_end = int(x5.shape[0])\n x5_1 = x5[f_begin: f_e...
loadN > 1:
{ "list": [ { "filename": "MCCL/models/modules.py", "retrieved_chunk": " def forward(self, x):\n if self.training:\n if config.loadN > 1:\n channel_per_class = x.shape[0] // config.loadN\n x_per_class_corr_list = []\n for idx in range(0...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F # import fvcore.nn.weight_init as weight_init from GCoNet_plus.config import Config config = Config() class ResBlk(nn.Module): def __init__(self, channel_in=64, channel_out=64): super(ResBlk, self).__init__() ...
else: a = torch.exp(-self.k * (x - y)) if torch.isinf(a).any(): a = torch.exp(-50 * (x - y)) return torch.reciprocal(1 + a) class RefUnet(nn.Module): # Refinement def __init__(self, in_ch, inc_ch): super(RefUnet, self).__init__() self.conv0 = nn...
{ "context_start_lineno": 0, "file": "GCoNet_plus/models/modules.py", "groundtruth_start_lineno": 398, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 399, "task_id": "project_cc_python/2383" }
{ "list": [ { "filename": "CoSOD_CoADNet/code/ops.py", "retrieved_chunk": "class DilConv3(nn.Module):\n # Dilated Convolution with 3*3 kernel size\n def __init__(self, ic, oc, is_bn, na, dr):\n super(DilConv3, self).__init__()\n # ic: input channels\n # oc: output channels\...
k_alpha) * mask_neg_inv))
{ "list": [ { "filename": "tests/destinationinstanceservice_test.py", "retrieved_chunk": "import unittest\nfrom unittest.mock import patch, MagicMock, ANY\nclass DestinationInstanceServiceTest(unittest.TestCase):\n ADHOC_INPUT = {\"flow1\": [\"dataset1\"], \"flow2\": [\"dataset2\", \"dataset3\"]}\n...
# Copyright 2023 Adobe. All rights reserved. # This file is licensed to you under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. You may obtain a copy # of the License at http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
assert (stats_result is not None) assert (stats_result == instance_conn.getData.return_value.get("results")) instance_conn.getData.assert_called_once()
{ "context_start_lineno": 0, "file": "tests/schema_test.py", "groundtruth_start_lineno": 57, "repository": "adobe-aepp-0e23c55", "right_context_start_lineno": 58, "task_id": "project_cc_python/2419" }
{ "list": [ { "filename": "tests/destinationinstanceservice_test.py", "retrieved_chunk": " result = destination_instance_service_obj.createAdHocDatasetExport(self.ADHOC_INPUT)\n assert(result is not None)\n instance_conn.postData.assert_called_once()\n instance_conn.postDat...
getBehaviors()
{ "list": [ { "filename": "CADC/Modules/Modules.py", "retrieved_chunk": " f_div_C = F.softmax(attention, dim=-1)\n g_x = self.g((feature.contiguous().view(-1, c)))\n y = torch.matmul(f_div_C, g_x)\n # (Nx46)xc/2\n W_y = self.W(y).contiguous().view(N, num, c)\n ...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F # import fvcore.nn.weight_init as weight_init from GCoNet_plus.config import Config config = Config() class ResBlk(nn.Module): def __init__(self, channel_in=64, channel_out=64): super(ResBlk, self).__init__() ...
super().__init__() self.k = k self.binarize = nn.Sequential( nn.Conv2d(channel_in, channel_in, 3, 1, 1), *[nn.BatchNorm2d(channel_in), nn.ReLU(inplace=True)] if config.use_bn else nn.ReLU(inplace=True), nn.Conv2d(channel_in, channel_in, 3, 1, 1), ...
{ "context_start_lineno": 0, "file": "GCoNet_plus/models/modules.py", "groundtruth_start_lineno": 367, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 368, "task_id": "project_cc_python/2381" }
{ "list": [ { "filename": "CADC/Modules/Modules.py", "retrieved_chunk": " self.maxpool1 = torch.nn.AdaptiveMaxPool2d((1,1))\n self.maxpool2 = torch.nn.AdaptiveMaxPool2d((3,3))\n self.maxpool3 = torch.nn.AdaptiveMaxPool2d((6,6))\n def forward(self, feature):\n batch_size,...
db_k):
{ "list": [ { "filename": "aepp/destination.py", "retrieved_chunk": " self.connector = connector.AdobeRequest(\n config=config,\n header=header,\n loggingEnabled=self.loggingEnabled,\n logger=self.logger,\n )\n self.header = self.connect...
# Copyright 2023 Adobe. All rights reserved. # This file is licensed to you under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. You may obtain a copy # of the License at http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
self.token = self.connector.token self.header['Authorization'] = 'bearer '+self.token def getConfigObject(self)->dict: """ Return the config object expected. """ return self.__configObject__ def getConfigHeader(self)->dict: """ Return th...
{ "context_start_lineno": 0, "file": "aepp/configs.py", "groundtruth_start_lineno": 343, "repository": "adobe-aepp-0e23c55", "right_context_start_lineno": 344, "task_id": "project_cc_python/2409" }
{ "list": [ { "filename": "aepp/connector.py", "retrieved_chunk": " }\n response = requests.post(\n config[\"oauthTokenEndpointV2\"], data=oauth_payload, verify=False\n )\n return self._token_postprocess(response=response, verbose=verbose, save=sa...
AdobeRequest(self.__configObject__,self.header)
{ "list": [ { "filename": "MCCL/config.py", "retrieved_chunk": "import os\nclass Config():\n def __init__(self) -> None:\n # Backbone\n self.bb = ['cnn-vgg16', 'cnn-vgg16bn', 'cnn-resnet50', 'trans-pvt'][3]\n self.pvt_weights = ['../bb_weights/pvt_v2_b2.pth', ''][0]\n # ...
from collections import OrderedDict import torch from torch.functional import norm import torch.nn as nn import torch.nn.functional as F from torchvision.models import vgg16, vgg16_bn from torchvision.models import resnet50 from MCCL.models.modules import ResBlk, CoAttLayer from MCCL.models.pvt import pvt_v2_b2 from M...
self.co_x4 = CoAttLayer(channel_in=lateral_channels_in[bb][0]) elif self.config.consensus == 'SGS': self.co_x4 = SGS(channel_in=lateral_channels_in[bb][0]) elif self.config.consensus == 'GWM': self.co_x4 = GWM(channel_in=lateral_channels_in[bb][0]) if self.c...
{ "context_start_lineno": 0, "file": "MCCL/models/GCoNet.py", "groundtruth_start_lineno": 60, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 61, "task_id": "project_cc_python/2406" }
{ "list": [ { "filename": "MCCL/config.py", "retrieved_chunk": " # Components\n self.consensus = ['', 'GCAM', 'GWM', 'SGS'][1]\n self.dec_blk = ['ResBlk'][0]\n self.GCAM_metric = ['online', 'offline', ''][0] if self.consensus else ''\n # Training\n self.batch_...
consensus == 'GCAM':
{ "list": [ { "filename": "tests/schema_test.py", "retrieved_chunk": "import unittest\nfrom unittest.mock import patch, MagicMock\nclass SchemaTest(unittest.TestCase):\n @patch(\"aepp.connector.AdobeRequest\")\n def test_schema_get_resource(self, mock_connector):\n instance_conn = mock_co...
# Copyright 2023 Adobe. All rights reserved. # This file is licensed to you under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. You may obtain a copy # of the License at http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
assert(result is not None) instance_conn.postData.assert_called_once() instance_conn.postData.assert_called_with(ANY, data=self.ADHOC_EXPECTED_PAYLOAD) @patch("aepp.connector.AdobeRequest") def test_create_adhoc_dataset_export_invalid_input(self, mock_connector): destination_in...
{ "context_start_lineno": 0, "file": "tests/destinationinstanceservice_test.py", "groundtruth_start_lineno": 24, "repository": "adobe-aepp-0e23c55", "right_context_start_lineno": 25, "task_id": "project_cc_python/2413" }
{ "list": [ { "filename": "tests/schema_test.py", "retrieved_chunk": " instance_conn.getData.assert_called_once()\n @patch(\"aepp.connector.AdobeRequest\")\n def test_schema_update_sandbox(self, mock_connector):\n schema_obj = Schema()\n test_sandbox = \"prod\"\n sche...
createAdHocDatasetExport(self.ADHOC_INPUT)
{ "list": [ { "filename": "tests/schema_test.py", "retrieved_chunk": "import unittest\nfrom unittest.mock import patch, MagicMock\nclass SchemaTest(unittest.TestCase):\n @patch(\"aepp.connector.AdobeRequest\")\n def test_schema_get_resource(self, mock_connector):\n instance_conn = mock_co...
# Copyright 2023 Adobe. All rights reserved. # This file is licensed to you under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. You may obtain a copy # of the License at http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
@patch('aepp.destinationinstanceservice.DestinationInstanceService.createAdHocDatasetExport', MagicMock(return_value = adhoc_non_retry_error)) @patch("aepp.connector.AdobeRequest", MagicMock()) def test_no_retry_error(self): export_obj = ExportDatasetToDataLandingZone(config= self.config, header= ...
{ "context_start_lineno": 0, "file": "tests/exportDatasetToDatalandingZone_test.py", "groundtruth_start_lineno": 145, "repository": "adobe-aepp-0e23c55", "right_context_start_lineno": 146, "task_id": "project_cc_python/2412" }
{ "list": [ { "filename": "tests/schema_test.py", "retrieved_chunk": " instance_conn.getData.assert_called_once()\n @patch(\"aepp.connector.AdobeRequest\")\n def test_schema_update_sandbox(self, mock_connector):\n schema_obj = Schema()\n test_sandbox = \"prod\"\n sche...
retryOnNotReadyException("test", "test", 1, 1) == self.adhoc_success_response)
{ "list": [ { "filename": "MCCL/models/modules.py", "retrieved_chunk": " x5 = F.interpolate(x5, size=x4.size()[2:], mode='bilinear', align_corners=True)\n x = torch.cat((x1, x2, x3, x4, x5), dim=1)\n x = self.conv1(x)\n x = self.bn1(x)\n x = self.relu(x)\n ret...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F # import fvcore.nn.weight_init as weight_init from GCoNet_plus.config import Config config = Config() class ResBlk(nn.Module): def __init__(self, channel_in=64, channel_out=64): super(ResBlk, self).__init__() ...
self.conv_output = nn.Conv2d(channel_in, channel_in, kernel_size=1, stride=1, padding=0) self.conv_transform = nn.Conv2d(channel_in, channel_in, kernel_size=1, stride=1, padding=0) self.fc_transform = nn.Linear(channel_in, channel_in) # for layer in [self.conv_output, self.conv_trans...
{ "context_start_lineno": 0, "file": "GCoNet_plus/models/modules.py", "groundtruth_start_lineno": 92, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 93, "task_id": "project_cc_python/2380" }
{ "list": [ { "filename": "GCoNet/models/GCoNet.py", "retrieved_chunk": " nn.ReLU(inplace=True),\n #nn.Conv2d(int(in_channel/2), int(in_channel/4), kernel_size=3, stride=1, padding=1),\n #nn.ReLU(inplace=True),\n )\n self.predlayer = nn.Sequential(\n ...
relation_module + '(channel_in)')
{ "list": [ { "filename": "MCCL/models/GCoNet.py", "retrieved_chunk": " self.co_x4 = GWM(channel_in=lateral_channels_in[bb][0])\n if self.config.dec_blk == 'ResBlk':\n DecBlk = ResBlk\n self.top_layer = DecBlk(lateral_channels_in[bb][0], lateral_channels_in[bb][1])\...
from collections import OrderedDict import torch from torch.functional import norm import torch.nn as nn import torch.nn.functional as F from torchvision.models import vgg16, vgg16_bn # import fvcore.nn.weight_init as weight_init from torchvision.models import resnet50 from GCoNet_plus.models.modules import ResBlk, DS...
ch_decoder //= 2 self.enlayer4 = ResBlk(ch_decoder*2, ch_decoder) if self.config.conv_after_itp: self.dslayer4 = DSLayer(ch_decoder, ch_decoder) self.latlayer4 = ResBlk(lateral_channels_in[2], ch_decoder) if self.config.complex_lateral_connection else nn.Conv2d(lateral_chan...
{ "context_start_lineno": 0, "file": "GCoNet_plus/models/GCoNet.py", "groundtruth_start_lineno": 60, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 61, "task_id": "project_cc_python/2387" }
{ "list": [ { "filename": "MCCL/models/GCoNet.py", "retrieved_chunk": " self.dec_layer1 = DecBlk(lateral_channels_in[bb][3], lateral_channels_in[bb][3]//2)\n self.conv_out1 = nn.Sequential(nn.Conv2d(lateral_channels_in[bb][3]//2, 1, 1, 1, 0))\n def forward(self, x):\n #########...
complex_lateral_connection else nn.Conv2d(lateral_channels_in[1], ch_decoder, 1, 1, 0)
{ "list": [ { "filename": "GCoNet_plus/models/modules.py", "retrieved_chunk": " self.upscore2 = nn.Upsample(scale_factor=2, mode='bilinear', align_corners=True)\n if config.db_output_refiner:\n self.db_output_refiner = DBHead(64)\n def forward(self, x):\n hx = x\n ...
from collections import OrderedDict import torch from torch.functional import norm import torch.nn as nn import torch.nn.functional as F from torchvision.models import vgg16, vgg16_bn # import fvcore.nn.weight_init as weight_init from torchvision.models import resnet50 from GCoNet_plus.models.modules import ResBlk, DS...
self.conv_cat_mask = nn.Conv2d(4, 3, 1, 1, 0) def forward(self, x): ########## Encoder ########## [N, _, H, W] = x.size() x1 = self.bb.conv1(x) x2 = self.bb.conv2(x1) x3 = self.bb.conv3(x2) x4 = self.bb.conv4(x3) x5 = self.bb.conv5(x4) ...
{ "context_start_lineno": 0, "file": "GCoNet_plus/models/GCoNet.py", "groundtruth_start_lineno": 110, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 111, "task_id": "project_cc_python/2395" }
{ "list": [ { "filename": "GCoNet_plus/models/modules.py", "retrieved_chunk": " x = self.relu_in(x)\n x = self.conv_out(x)\n if config.use_bn:\n x = self.bn_out(x)\n return x\nclass DSLayer(nn.Module):\n def __init__(self, channel_in=64, channel_out=1, activat...
cls_mask_operation == 'c':
{ "list": [ { "filename": "tests/destinationinstanceservice_test.py", "retrieved_chunk": "import unittest\nfrom unittest.mock import patch, MagicMock, ANY\nclass DestinationInstanceServiceTest(unittest.TestCase):\n ADHOC_INPUT = {\"flow1\": [\"dataset1\"], \"flow2\": [\"dataset2\", \"dataset3\"]}\n...
# Copyright 2023 Adobe. All rights reserved. # This file is licensed to you under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. You may obtain a copy # of the License at http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
assert(result is not None) instance_conn.getData.assert_called_once() @patch("aepp.connector.AdobeRequest") def test_schema_update_sandbox(self, mock_connector): schema_obj = Schema() test_sandbox = "prod" schema_obj.updateSandbox(test_sandbox) assert(schema_obj...
{ "context_start_lineno": 0, "file": "tests/schema_test.py", "groundtruth_start_lineno": 22, "repository": "adobe-aepp-0e23c55", "right_context_start_lineno": 23, "task_id": "project_cc_python/2414" }
{ "list": [ { "filename": "tests/destinationinstanceservice_test.py", "retrieved_chunk": " result = destination_instance_service_obj.createAdHocDatasetExport(self.ADHOC_INPUT)\n assert(result is not None)\n instance_conn.postData.assert_called_once()\n instance_conn.postDat...
getResource(MagicMock(), MagicMock(), MagicMock(), MagicMock())
{ "list": [ { "filename": "MCCL/models/GCoNet.py", "retrieved_chunk": " self.co_x4 = GWM(channel_in=lateral_channels_in[bb][0])\n if self.config.dec_blk == 'ResBlk':\n DecBlk = ResBlk\n self.top_layer = DecBlk(lateral_channels_in[bb][0], lateral_channels_in[bb][1])\...
from collections import OrderedDict import torch from torch.functional import norm import torch.nn as nn import torch.nn.functional as F from torchvision.models import vgg16, vgg16_bn # import fvcore.nn.weight_init as weight_init from torchvision.models import resnet50 from GCoNet_plus.models.modules import ResBlk, DS...
self.conv_out4 = nn.Sequential(nn.Conv2d(ch_decoder, 32, 1, 1, 0), nn.ReLU(inplace=True), nn.Conv2d(32, 1, 1, 1, 0)) ch_decoder //= 2 self.enlayer3 = ResBlk(ch_decoder*2, ch_decoder) if self.config.conv_after_itp: self.dslayer3 = DSLayer(ch_decoder, ch_decoder) ...
{ "context_start_lineno": 0, "file": "GCoNet_plus/models/GCoNet.py", "groundtruth_start_lineno": 67, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 68, "task_id": "project_cc_python/2388" }
{ "list": [ { "filename": "MCCL/models/GCoNet.py", "retrieved_chunk": " self.dec_layer1 = DecBlk(lateral_channels_in[bb][3], lateral_channels_in[bb][3]//2)\n self.conv_out1 = nn.Sequential(nn.Conv2d(lateral_channels_in[bb][3]//2, 1, 1, 1, 0))\n def forward(self, x):\n #########...
output_number >= 4:
{ "list": [ { "filename": "CoSOD_CoADNet/code/ops.py", "retrieved_chunk": " self.activation = nn.Sigmoid()\n def forward(self, x):\n # x: [B, ic]\n # y: [B, oc]\n y = self.linear(x)\n if self.is_bn:\n y = self.batch_normalization(y)\n if self...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F # import fvcore.nn.weight_init as weight_init from GCoNet_plus.config import Config config = Config() class ResBlk(nn.Module): def __init__(self, channel_in=64, channel_out=64): super(ResBlk, self).__init__() ...
z = x - y mask_neg_inv = 1 - 2 * (z < 0) a = torch.exp(-self.k * (torch.pow(z * mask_neg_inv + 1e-16, 1/config.k_alpha) * mask_neg_inv)) else: a = torch.exp(-self.k * (x - y)) if torch.isinf(a).any(): a = torch.exp(-50 * (x - y)) retur...
{ "context_start_lineno": 0, "file": "GCoNet_plus/models/modules.py", "groundtruth_start_lineno": 395, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 396, "task_id": "project_cc_python/2382" }
{ "list": [ { "filename": "MCCL/models/modules.py", "retrieved_chunk": " x = self.conv_in(x)\n if config.use_bn:\n x = self.bn_in(x)\n x = self.relu_in(x)\n if config.dec_att:\n x = self.dec_att(x)\n x = self.conv_out(x)\n if config.use_b...
db_k_alpha != 1:
{ "list": [ { "filename": "MCCL/models/GCoNet.py", "retrieved_chunk": " p1_out = self.conv_out1(p1)\n scaled_preds.append(p1_out)\n if self.training:\n return_values = [scaled_preds, x4]\n return return_values\n else:\n return scaled_pre...
from collections import OrderedDict import torch from torch.functional import norm import torch.nn as nn import torch.nn.functional as F from torchvision.models import vgg16, vgg16_bn # import fvcore.nn.weight_init as weight_init from torchvision.models import resnet50 from GCoNet_plus.models.modules import ResBlk, DS...
norm_features = [] if '_x5' in self.config.triplet: norm_features.append(_x5) if 'mask' in self.config.triplet: norm_features.append(norm_features_mask[0]) return_values.append(norm_features) return retu...
{ "context_start_lineno": 0, "file": "GCoNet_plus/models/GCoNet.py", "groundtruth_start_lineno": 238, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 239, "task_id": "project_cc_python/2397" }
{ "list": [ { "filename": "MCCL/models/GCoNet.py", "retrieved_chunk": " p1_out = self.conv_out1(p1)\n scaled_preds.append(p1_out)\n if self.training:\n return_values = [scaled_preds, x4]\n return return_values\n else:\n return scaled_pre...
lambdas_sal_last['triplet']:
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " Please check if the file is organized properly.\")\n return sequences\ndef readRAD(filename):\n if os.path.exists(filename):\n return np.load(filename)\n else:\n return N...
"""raddet dataset.""" import numpy as np import tensorflow as tf import tensorflow_datasets.public_api as tfds from utils import loader, helper # TODO(raddet): Markdown description that will appear on the catalog page. _DESCRIPTION = """ Description is **formatted** as markdown. It should also contain any processi...
gt_instances = loader.readRadarInstances(gt_filename) if gt_instances is None: raise ValueError("gt file not found, please double check the path") # Get RD spectrum RD_data = helper.getSumDim(RAD_data, target_axis=1) # Get RD bboxes ...
{ "context_start_lineno": 0, "file": "datasets/raddet_builder/raddet.py", "groundtruth_start_lineno": 93, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 94, "task_id": "project_cc_python/2464" }
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " gt_file = os.path.join(prefix, \"gt\") + RAD_file_spec.replace(\"npy\", \"pickle\")\n return gt_file\ndef imgfileFromRADfile(RAD_file, prefix):\n \"\"\" Transfer RAD filename to gt filename \"\"\"\n R...
gtfileFromRADfile(RAD_filename, path)
{ "list": [ { "filename": "MCCL/config.py", "retrieved_chunk": " # Components\n self.consensus = ['', 'GCAM', 'GWM', 'SGS'][1]\n self.dec_blk = ['ResBlk'][0]\n self.GCAM_metric = ['online', 'offline', ''][0] if self.consensus else ''\n # Training\n self.batch_...
from collections import OrderedDict import torch from torch.functional import norm import torch.nn as nn import torch.nn.functional as F from torchvision.models import vgg16, vgg16_bn from torchvision.models import resnet50 from MCCL.models.modules import ResBlk, CoAttLayer from MCCL.models.pvt import pvt_v2_b2 from M...
DecBlk = ResBlk self.top_layer = DecBlk(lateral_channels_in[bb][0], lateral_channels_in[bb][1]) self.dec_layer4 = DecBlk(lateral_channels_in[bb][1], lateral_channels_in[bb][1]) self.lat_layer4 = nn.Conv2d(lateral_channels_in[bb][1], lateral_channels_in[bb][1], 1, 1, 0) se...
{ "context_start_lineno": 0, "file": "MCCL/models/GCoNet.py", "groundtruth_start_lineno": 67, "repository": "ZhengPeng7-CoSOD_fps_collection-bee3764", "right_context_start_lineno": 68, "task_id": "project_cc_python/2407" }
{ "list": [ { "filename": "MCCL/config.py", "retrieved_chunk": " self.lr = 1e-4\n self.freeze = True\n self.lr_decay_epochs = [-20] # Set to negative N to decay the lr in the last N-th epoch.\n self.forward_per_dataset = True\n # Adv\n self.lambda_adv_g = 1...
dec_blk == 'ResBlk':
{ "list": [ { "filename": "eval.py", "retrieved_chunk": " # Prepare dataset\n if dataset == \"carrada\":\n batched_test_dataset, dataset_info = data_utils.prepare_dataset(split=\"test\", config=config, seed=seed)\n elif dataset == \"raddet\":\n batched_test_dataset, dataset_info...
import time import tensorflow as tf from darod.models import model from darod.trainers.trainer import Trainer from darod.utils import data_utils, bbox_utils, io_utils seed = 42 # Set memory growth to avoid GPU to crash physical_devices = tf.config.experimental.list_physical_devices('GPU') tf.config.experimental.set...
# batched_test_dataset, _ = data_utils.prepare_dataset(split="test", config=config, seed=seed) batched_val_dataset, _ = data_utils.prepare_dataset(split="val", config=config, seed=seed) else: batched_train_dataset, dataset_info = data_utils.prepare_dataset(split="train[:90%]", config=config, seed=seed)...
{ "context_start_lineno": 0, "file": "train.py", "groundtruth_start_lineno": 23, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 24, "task_id": "project_cc_python/2453" }
{ "list": [ { "filename": "vizualize_testdata.py", "retrieved_chunk": " faster_rcnn_model = model.R2D2(config, anchors)\n if layout == \"2D\":\n faster_rcnn_model.build(input_shape=(None, 256, 64, 1))\n else:\n fake_input = tf.zeros(shape=(config[\"training\"][\"batch_size\"], c...
get_total_item_size(dataset_info, "train")
{ "list": [ { "filename": "darod/layers/frcnn_layers.py", "retrieved_chunk": " variances = self.config[\"fastrcnn\"][\"variances_boxes\"]\n adaptive_ratio = self.config[\"fastrcnn\"][\"adaptive_ratio\"]\n positive_th = self.config[\"fastrcnn\"][\"positive_th\"]\n batch_size...
import time import numpy as np import tensorflow as tf import tensorflow_addons as tfa from ..utils import bbox_utils def compute_eta(total_duration, epoch_duration, epoch, total_epochs): """Compute training ETA""" total_duration += epoch_duration eta = (total_epochs - epoch) * total_duration / (epoch) ...
# Get max index value for each row max_indices_each_row = tf.argmax(iou_map, axis=2, output_type=tf.int32) # Get max index value for each column max_indices_each_column = tf.argmax(iou_map, axis=1, output_type=tf.int32) # IoU map has iou values for every gt boxes and we merge these values column wi...
{ "context_start_lineno": 0, "file": "darod/utils/train_utils.py", "groundtruth_start_lineno": 53, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 54, "task_id": "project_cc_python/2458" }
{ "list": [ { "filename": "darod/layers/frcnn_layers.py", "retrieved_chunk": " #\n pos_mask = tf.greater(merged_iou_map, positive_th)\n pos_mask = train_utils.randomly_select_xyz_mask(pos_mask, tf.constant([total_pos_bboxes], dtype=tf.int32),\n ...
generate_iou_map(anchors, gt_boxes)
{ "list": [ { "filename": "datasets/carrada_builder/carrada.py", "retrieved_chunk": " 'image/filename': image_fn,\n 'spectrum/id': s_id,\n 'objects': objects\n }\n s_id += 1\n...
"""raddet dataset.""" import numpy as np import tensorflow as tf import tensorflow_datasets.public_api as tfds from utils import loader, helper # TODO(raddet): Markdown description that will appear on the catalog page. _DESCRIPTION = """ Description is **formatted** as markdown. It should also contain any processi...
if RAD_complex is None: raise ValueError("RAD file not found, please double check the path.") RAD_data = helper.complexTo2channels(RAD_complex) # Normalize data RAD_data = (RAD_data - global_mean_log) / global_variance_log # RAD_data = (RAD_da...
{ "context_start_lineno": 0, "file": "datasets/raddet_builder/raddet.py", "groundtruth_start_lineno": 85, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 86, "task_id": "project_cc_python/2462" }
{ "list": [ { "filename": "datasets/carrada_builder/carrada.py", "retrieved_chunk": " y.append(y_)\n y1, x1, y2, x2 = min(y), min(x), max(y), max(x)\n bbox = [y1, x1, y2, x2]\n area = self._compute_area(bbox)\n return tfds.features.BBox(\n ymin=y1 / h,...
readRAD(RAD_filename)
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " Please check if the file is organized properly.\")\n return sequences\ndef readRAD(filename):\n if os.path.exists(filename):\n return np.load(filename)\n else:\n return N...
"""raddet dataset.""" import numpy as np import tensorflow as tf import tensorflow_datasets.public_api as tfds from utils import loader, helper # TODO(raddet): Markdown description that will appear on the catalog page. _DESCRIPTION = """ Description is **formatted** as markdown. It should also contain any processi...
if gt_instances is None: raise ValueError("gt file not found, please double check the path") # Get RD spectrum RD_data = helper.getSumDim(RAD_data, target_axis=1) # Get RD bboxes bboxes, classes = helper.readAndEncodeGtRD(gt_instances, RD_data...
{ "context_start_lineno": 0, "file": "datasets/raddet_builder/raddet.py", "groundtruth_start_lineno": 94, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 95, "task_id": "project_cc_python/2465" }
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " gt_file = os.path.join(prefix, \"gt\") + RAD_file_spec.replace(\"npy\", \"pickle\")\n return gt_file\ndef imgfileFromRADfile(RAD_file, prefix):\n \"\"\" Transfer RAD filename to gt filename \"\"\"\n R...
readRadarInstances(gt_filename)
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " Please check if the file is organized properly.\")\n return sequences\ndef readRAD(filename):\n if os.path.exists(filename):\n return np.load(filename)\n else:\n return N...
"""raddet dataset.""" import numpy as np import tensorflow as tf import tensorflow_datasets.public_api as tfds from utils import loader, helper # TODO(raddet): Markdown description that will appear on the catalog page. _DESCRIPTION = """ Description is **formatted** as markdown. It should also contain any processi...
# Normalize data RAD_data = (RAD_data - global_mean_log) / global_variance_log # RAD_data = (RAD_data - global_min_log) / (global_max_log - global_min_log) # Load GT instances gt_filename = loader.gtfileFromRADfile(RAD_filename, path) gt_instances...
{ "context_start_lineno": 0, "file": "datasets/raddet_builder/raddet.py", "groundtruth_start_lineno": 88, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 89, "task_id": "project_cc_python/2463" }
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " gt_file = os.path.join(prefix, \"gt\") + RAD_file_spec.replace(\"npy\", \"pickle\")\n return gt_file\ndef imgfileFromRADfile(RAD_file, prefix):\n \"\"\" Transfer RAD filename to gt filename \"\"\"\n R...
complexTo2channels(RAD_complex)
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " Please check if the file is organized properly.\")\n return sequences\ndef readRAD(filename):\n if os.path.exists(filename):\n return np.load(filename)\n else:\n return N...
"""raddet dataset.""" import numpy as np import tensorflow as tf import tensorflow_datasets.public_api as tfds from utils import loader, helper # TODO(raddet): Markdown description that will appear on the catalog page. _DESCRIPTION = """ Description is **formatted** as markdown. It should also contain any processi...
seq_id = RAD_filename.split('/')[-2].split('_')[-1] for (box, class_) in zip(bboxes, classes): bbox, area = helper.buildTfdsBoxes(box) objects.append({ 'bbox': bbox, 'label': classes_list.index(class_), ...
{ "context_start_lineno": 0, "file": "datasets/raddet_builder/raddet.py", "groundtruth_start_lineno": 100, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 101, "task_id": "project_cc_python/2467" }
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " gt_file = os.path.join(prefix, \"gt\") + RAD_file_spec.replace(\"npy\", \"pickle\")\n return gt_file\ndef imgfileFromRADfile(RAD_file, prefix):\n \"\"\" Transfer RAD filename to gt filename \"\"\"\n R...
readAndEncodeGtRD(gt_instances, RD_data.shape)
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " Please check if the file is organized properly.\")\n return sequences\ndef readRAD(filename):\n if os.path.exists(filename):\n return np.load(filename)\n else:\n return N...
"""raddet dataset.""" import numpy as np import tensorflow as tf import tensorflow_datasets.public_api as tfds from utils import loader, helper # TODO(raddet): Markdown description that will appear on the catalog page. _DESCRIPTION = """ Description is **formatted** as markdown. It should also contain any processi...
# Get RD bboxes bboxes, classes = helper.readAndEncodeGtRD(gt_instances, RD_data.shape) seq_id = RAD_filename.split('/')[-2].split('_')[-1] for (box, class_) in zip(bboxes, classes): bbox, area = helper.buildTfdsBoxes(box) objects.append({...
{ "context_start_lineno": 0, "file": "datasets/raddet_builder/raddet.py", "groundtruth_start_lineno": 98, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 99, "task_id": "project_cc_python/2466" }
{ "list": [ { "filename": "datasets/raddet_builder/utils/loader.py", "retrieved_chunk": " gt_file = os.path.join(prefix, \"gt\") + RAD_file_spec.replace(\"npy\", \"pickle\")\n return gt_file\ndef imgfileFromRADfile(RAD_file, prefix):\n \"\"\" Transfer RAD filename to gt filename \"\"\"\n R...
getSumDim(RAD_data, target_axis=1)
{ "list": [ { "filename": "datasets/carrada_builder/carrada.py", "retrieved_chunk": " 'label': label - 1,\n 'area': area,\n 'id': a_id\n })\n a_id += 1\n ...
"""raddet dataset.""" import numpy as np import tensorflow as tf import tensorflow_datasets.public_api as tfds from utils import loader, helper # TODO(raddet): Markdown description that will appear on the catalog page. _DESCRIPTION = """ Description is **formatted** as markdown. It should also contain any processi...
example = { 'spectrum': RD_data.astype(np.float32), 'spectrum/filename': RAD_filename, 'sequence/id': int(seq_id), 'image': image_filename, 'spectrum/id': count, 'objects': objects } coun...
{ "context_start_lineno": 0, "file": "datasets/raddet_builder/raddet.py", "groundtruth_start_lineno": 111, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 112, "task_id": "project_cc_python/2469" }
{ "list": [ { "filename": "datasets/carrada_builder/carrada.py", "retrieved_chunk": " 'image/filename': image_fn,\n 'spectrum/id': s_id,\n 'objects': objects\n }\n s_id += 1\n...
imgfileFromRADfile(RAD_filename, path)
{ "list": [ { "filename": "vizualize_testdata.py", "retrieved_chunk": " target_id = args.seq_id\n eval_best = args.eval_best if args.eval_best is not None else False\n # Prepare dataset\n if dataset == \"carrada\":\n batched_test_dataset, dataset_info = data_utils.prepare_dataset(sp...
import time import tensorflow as tf from darod.models import model from darod.trainers.trainer import Trainer from darod.utils import data_utils, bbox_utils, io_utils seed = 42 # Set memory growth to avoid GPU to crash physical_devices = tf.config.experimental.list_physical_devices('GPU') tf.config.experimental.set...
num_train_example = data_utils.get_total_item_size(dataset_info, "train") # batched_test_dataset, _ = data_utils.prepare_dataset(split="test", config=config, seed=seed) batched_val_dataset, _ = data_utils.prepare_dataset(split="val", config=config, seed=seed) else: batched_train_dataset, dataset_in...
{ "context_start_lineno": 0, "file": "train.py", "groundtruth_start_lineno": 22, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 23, "task_id": "project_cc_python/2452" }
{ "list": [ { "filename": "eval.py", "retrieved_chunk": "def main():\n args = io_utils.handle_args_eval()\n # summary_writer = tf.summary.create_file_writer(args.path)\n config_pth = os.path.join(args.path, \"config.json\")\n with open(config_pth, 'r') as file:\n config = json.load(...
prepare_dataset(split="train", config=config, seed=seed)
{ "list": [ { "filename": "vizualize_testdata.py", "retrieved_chunk": " faster_rcnn_model = model.R2D2(config, anchors)\n if layout == \"2D\":\n faster_rcnn_model.build(input_shape=(None, 256, 64, 1))\n else:\n fake_input = tf.zeros(shape=(config[\"training\"][\"batch_size\"], c...
import time import tensorflow as tf from darod.models import model from darod.trainers.trainer import Trainer from darod.utils import data_utils, bbox_utils, io_utils seed = 42 # Set memory growth to avoid GPU to crash physical_devices = tf.config.experimental.list_physical_devices('GPU') tf.config.experimental.set...
{ "context_start_lineno": 0, "file": "train.py", "groundtruth_start_lineno": 52, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 53, "task_id": "project_cc_python/2457" }
{ "list": [ { "filename": "vizualize_testdata.py", "retrieved_chunk": " use_scheduler = config[\"training\"][\"scheduler\"]\n momentum = config[\"training\"][\"momentum\"]\n if optimizer == \"SGD\":\n optimizer = tf.optimizers.SGD(learning_rate=lr, momentum=momentum)\n ...
train(anchors, batched_train_dataset, batched_val_dataset)
{ "list": [ { "filename": "darod/layers/frcnn_layers.py", "retrieved_chunk": " neg_mask = train_utils.randomly_select_xyz_mask(neg_mask, total_neg_bboxes,\n seed=self.config[\"training\"][\"seed\"])\n else:\n neg_m...
import time import numpy as np import tensorflow as tf import tensorflow_addons as tfa from ..utils import bbox_utils def compute_eta(total_duration, epoch_duration, epoch, total_epochs): """Compute training ETA""" total_duration += epoch_duration eta = (total_epochs - epoch) * total_duration / (epoch) ...
# bbox_labels = tf.reshape(bbox_labels, (batch_size, output_height, output_width, anchor_count)) # return bbox_deltas, bbox_labels def frcnn_cls_loss(*args): """ Calculating faster rcnn class loss value. :param args: could be (y_true, y_pred) or ((y_true, y_pred), ) :return: CE loss ...
{ "context_start_lineno": 0, "file": "darod/utils/train_utils.py", "groundtruth_start_lineno": 92, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 93, "task_id": "project_cc_python/2459" }
{ "list": [ { "filename": "darod/layers/frcnn_layers.py", "retrieved_chunk": " pos_gt_labels = tf.where(pos_mask, gt_labels_map, tf.constant(-1, dtype=tf.int32))\n neg_gt_labels = tf.cast(neg_mask, dtype=tf.int32)\n expanded_gt_labels = pos_gt_labels + neg_gt_labels\n #\n ...
get_deltas_from_bboxes(anchors, expanded_gt_boxes) / variances
{ "list": [ { "filename": "darod/trainers/trainer.py", "retrieved_chunk": " \"\"\"\n Update val metrics\n :param loss: total loss\n :param rpn_reg_loss: rpn regression loss\n :param rpn_cls_loss: rpn classification loss\n :param frcnn_reg_loss: fast rcnn regre...
import time import numpy as np import tensorflow as tf import tensorflow_addons as tfa from ..utils import bbox_utils def compute_eta(total_duration, epoch_duration, epoch, total_epochs): """Compute training ETA""" total_duration += epoch_duration eta = (total_epochs - epoch) * total_duration / (epoch) ...
# loss_fn = tfa.losses.GIoULoss(reduction=tf.losses.Reduction.NONE) loss_for_all = loss_fn(y_true, y_pred) # loss_for_all = tf.reduce_sum(loss_for_all, axis=-1) # pos_cond = tf.reduce_any(tf.not_equal(y_true, tf.constant(0.0)), axis=-1) pos_mask = tf.cast(pos_cond, dtype=tf.float32) # ...
{ "context_start_lineno": 0, "file": "darod/utils/train_utils.py", "groundtruth_start_lineno": 163, "repository": "colindecourt-darod-ab4878e", "right_context_start_lineno": 164, "task_id": "project_cc_python/2460" }
{ "list": [ { "filename": "darod/trainers/trainer.py", "retrieved_chunk": " self.val_rpn_reg_loss(rpn_reg_loss)\n self.val_rpn_cls_loss(rpn_cls_loss)\n self.val_frcnn_reg_loss(frcnn_reg_loss)\n self.val_frcnn_cls_loss(frcnn_cls_loss)\n def train_step(self, input_data, an...
get_bboxes_from_deltas(roi_bboxes, y_pred)
{ "list": [ { "filename": "src/transformations/array.py", "retrieved_chunk": "from typing import List, Optional\nfrom codegen.data import ListyType, PolymorphicModel, PolymorphicType, UnionType\nfrom swagger import Schema\nfrom transformations.data import MutableContext\ndef is_schema_array(schema: Sc...
from typing import List, Optional from codegen.data import ContainerModel, GenericType, ModelType, OptionType, PolymorphicModel, PolymorphicType, PrimitiveType, PrimitiveTypeEnum, UnionType from swagger import Schema, SwaggerDataType from transformations.data import MutableContext from transformations.util import attac...
non_null_types = [type for type in types if type != SwaggerDataType.Null] # avoid a union for a singular type type = types_to_union(non_null_types) if len(non_null_types) > 1 else union_mapping[non_null_types[0]] if is_nullable: return OptionType(type).to_polymorphic() else: ret...
{ "context_start_lineno": 0, "file": "src/transformations/union.py", "groundtruth_start_lineno": 23, "repository": "UMEssen-matrix-scala-bd97a56", "right_context_start_lineno": 24, "task_id": "project_cc_python/2475" }
{ "list": [ { "filename": "src/transformations/generic.py", "retrieved_chunk": "from codegen.data import GenericType, PolymorphicType\nfrom swagger import Schema, SwaggerDataType\nfrom transformations.data import MutableContext\ndef is_schema_generic(schema: Schema) -> bool:\n no_additional_propert...
Null in types
{ "list": [ { "filename": "src/transformations/model.py", "retrieved_chunk": " return (schema.properties is not None) or (schema.allOf is not None)\n@dataclass\nclass FieldTransformation:\n arg: Argument\ndef transform_model_field(ctx: MutableContext, field_name: str, field_schema: Union[Schema,...
from typing import TypeVar import re from codegen.data import OptionType, PolymorphicModel, PolymorphicType T = TypeVar("T") def PLACEHOLDER(x: T) -> T: return x def flatten_2d(outer: list[list[T]]) -> list[T]: return [item for sublist in outer for item in sublist] # "Some cool name" => "SomeCoolName" de...
return wrapped
{ "context_start_lineno": 0, "file": "src/transformations/util.py", "groundtruth_start_lineno": 35, "repository": "UMEssen-matrix-scala-bd97a56", "right_context_start_lineno": 36, "task_id": "project_cc_python/2488" }
{ "list": [ { "filename": "src/codegen/data.py", "retrieved_chunk": " is_data_model: bool = False\n is_string_enum: bool = False\n is_responses_trait: bool = False\n is_response_box: bool = False\n @staticmethod\n def from_basic(data: BaseModel) -> PolymorphicModel:\n if isins...
to_polymorphic() if not is_required else t
{ "list": [ { "filename": "xlmr/src/generate2.py", "retrieved_chunk": " target_str = tgt_dict.string(\n target_tokens,\n cfg.common_eval.post_process,\n escape_unk=True,\n extra_symbols_to_ig...
#!/usr/bin/env python3 -u # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """ Translate pre-processed data with a trained model. """ import ast import logging import math import os import sy...
continue if not cfg.common_eval.quiet: if src_dict is not None: print("S-{}\t{}".format(sample_id, src_str), file=output_file) # if has_target: # print("T-{}\t{}".format(sample_id, target_str), file=output_file) ...
{ "context_start_lineno": 0, "file": "xlmr/src/generate.py", "groundtruth_start_lineno": 253, "repository": "dropreg-ChatMLM-273477a", "right_context_start_lineno": 254, "task_id": "project_cc_python/2500" }
{ "list": [ { "filename": "xlmr/src/generate2.py", "retrieved_chunk": " target_str = decode_fn(target_str)\n if not cfg.common_eval.quiet:\n if src_dict is not None:\n print(\"S-{}\\t{}\".format(sample_id, src_str), file=output_file)\n ...
get_model_parallel_rank())
{ "list": [ { "filename": "src/transformations/types.py", "retrieved_chunk": " return transform_schema_as_file(ctx, schema)\n if is_schema_primitive(schema):\n return transform_schema_as_primitive(ctx, schema)\n if is_schema_map(schema):\n return transform_schema_as_map(ctx,...
from typing import List, Optional from codegen.data import ListyType, PolymorphicModel, PolymorphicType, UnionType from swagger import Schema from transformations.data import MutableContext def is_schema_array(schema: Schema) -> bool: return schema.items is not None def transform_item_types(ctx: MutableContext...
array_type = ListyType( inner_type=inner_type, ) return array_type.to_polymorphic()
{ "context_start_lineno": 0, "file": "src/transformations/array.py", "groundtruth_start_lineno": 27, "repository": "UMEssen-matrix-scala-bd97a56", "right_context_start_lineno": 28, "task_id": "project_cc_python/2485" }
{ "list": [ { "filename": "src/transformations/types.py", "retrieved_chunk": " return transform_schema_as_generic(ctx, schema)\n raise Exception(\"failed to transform schema\")", "score": 60.36888826522735 }, { "filename": "src/transformations/model.py", "retrieved_...
to_polymorphic() if len(item_types) > 1 else item_types[0]
{ "list": [ { "filename": "sdk/python/src/cakework/task_server.py", "retrieved_chunk": " def __init__(self, user_task, local=False):\n self.user_task = user_task\n self.server = grpc.server(futures.ThreadPoolExecutor(max_workers=1)) # what should the default be?\n self.local = ...
import subprocess import os import shutil import sys import inspect import json from concurrent import futures from cakework import cakework_pb2 from cakework import cakework_pb2_grpc from .task_server import TaskServer import importlib import logging logging.basicConfig(level=logging.INFO) class Cakework: def __ini...
server.add_insecure_port('[::]:' + port) server.start() logging.info("Server started, listening on " + port) server.wait_for_termination()
{ "context_start_lineno": 0, "file": "sdk/python/src/cakework/cakework.py", "groundtruth_start_lineno": 29, "repository": "usecakework-async-backend-c288d0b", "right_context_start_lineno": 30, "task_id": "project_cc_python/2247" }
{ "list": [ { "filename": "sdk/python/src/cakework/task_server.py", "retrieved_chunk": " self.server.wait_for_termination()", "score": 60.81456587313835 }, { "filename": "sdk/python/src/cakework/task_server.py", "retrieved_chunk": " def __init__(self, user_task, loc...
add_CakeworkServicer_to_server(cakework_pb2_grpc.Cakework(), server)
{ "list": [ { "filename": "sdk/python/src/cakework/cakework.py", "retrieved_chunk": "\t\tactivity_server.start()\ndef serve():\n port = '50051'\n server = grpc.server(futures.ThreadPoolExecutor(max_workers=1)) # what should the default be?\n cakework_pb2_grpc.add_CakeworkServicer_to_server(ca...
from concurrent import futures import logging import grpc from cakework import cakework_pb2 from cakework import cakework_pb2_grpc import json import threading import requests import os import logging logging.basicConfig(level=logging.INFO) def get_token(context): metadict = dict(context.invocation_metadata()) ...
server.add_insecure_port('[::]:' + port) server.start() logging.info("Server started, listening on " + port) server.wait_for_termination() if __name__ == '__main__': logging.basicConfig() serve() class TaskServer: def __init__(self, user_task, local=False): self.user_task = user_t...
{ "context_start_lineno": 0, "file": "sdk/python/src/cakework/task_server.py", "groundtruth_start_lineno": 81, "repository": "usecakework-async-backend-c288d0b", "right_context_start_lineno": 82, "task_id": "project_cc_python/2245" }
{ "list": [ { "filename": "sdk/python/src/cakework/cakework.py", "retrieved_chunk": "\t\tactivity_server.start()\ndef serve():\n port = '50051'\n server = grpc.server(futures.ThreadPoolExecutor(max_workers=1)) # what should the default be?\n cakework_pb2_grpc.add_CakeworkServicer_to_server(ca...
add_CakeworkServicer_to_server(Cakework(), server)
{ "list": [ { "filename": "ros2_benchmark/ros2_benchmark/basic_performance_calculator.py", "retrieved_chunk": " perf_data[BasicPerformanceMetrics.MAX_JITTER] = float(numpy.max(jitters))\n perf_data[BasicPerformanceMetrics.MIN_JITTER] = float(numpy.min(jitters))\n perf_...
# SPDX-FileCopyrightText: NVIDIA CORPORATION & AFFILIATES # Copyright (c) 2021-2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # h...
return profile_data def reset(self): """Reset the profiler state.""" self._profile_data_list.clear() return def conclude_results(self) -> dict: """Conclude final profiling outcome based on all previous results.""" if len(self._profile_data_list) == 0: ...
{ "context_start_lineno": 0, "file": "ros2_benchmark/ros2_benchmark/utils/cpu_profiler.py", "groundtruth_start_lineno": 105, "repository": "NVIDIA-ISAAC-ROS-ros2_benchmark-d7725da", "right_context_start_lineno": 106, "task_id": "project_cc_python/2526" }
{ "list": [ { "filename": "ros2_benchmark/ros2_benchmark/basic_performance_calculator.py", "retrieved_chunk": " self._perf_data_list.append(perf_data)\n if self._report_prefix != '':\n return {self._report_prefix: perf_data}\n return perf_data\n def conclude_performa...
_profile_data_list.append(profile_data)
{ "list": [ { "filename": "ros2_benchmark/ros2_benchmark/ros2_benchmark_test.py", "retrieved_chunk": " ----------\n test_func\n The benchmark function to be tested\n Returns\n -------\n float\n The maximum sustainable test pulbisher framerate\n ...
# SPDX-FileCopyrightText: NVIDIA CORPORATION & AFFILIATES # Copyright (c) 2021-2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # h...
while self._is_running: logfile.write( f'{psutil.cpu_percent(interval=interval, percpu=True)}\n') self.psutil_thread = Thread(target=psutil_log) self.psutil_thread.start() return self._log_file_path def stop_profiling(self): ...
{ "context_start_lineno": 0, "file": "ros2_benchmark/ros2_benchmark/utils/cpu_profiler.py", "groundtruth_start_lineno": 61, "repository": "NVIDIA-ISAAC-ROS-ros2_benchmark-d7725da", "right_context_start_lineno": 62, "task_id": "project_cc_python/2523" }
{ "list": [ { "filename": "ros2_benchmark/ros2_benchmark/utils/profiler.py", "retrieved_chunk": " self._is_running = True\n # Create log file folders if they don't exist already\n os.makedirs(log_dir, exist_ok=True)\n self._log_file_path = os.path.join(\n log_dir...
_log_file_path, 'w+') as logfile:
{ "list": [ { "filename": "src/auditnlg/explain.py", "retrieved_chunk": " data = example_format_checker(data)\n gen_kwargs = {\n \"max_tokens\": 128\n }\n explanations = []\n for i, sample in enumerate(tqdm(data)):\n input_ = explain_template.format(\n instructi...
''' * Copyright (c) 2023, Salesforce, Inc. * All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause ''' import re import numpy as np from ..llm_wrapper import OpenAILLM, LocalLLM from .openAIscorers...
ppl = np.exp(avg_loss) scores.append(1 / ppl) meta_data.append(f"PPL:{ppl}") return scores, meta_data def summac_metric(data, global_knowledge, use_cuda, gpu_device): from .summac_utils.evaluator import SummacEvaluator device = "cpu" if use_cuda: device = f"cuda:{gpu_d...
{ "context_start_lineno": 0, "file": "src/auditnlg/factualness/classifier.py", "groundtruth_start_lineno": 123, "repository": "salesforce-AuditNLG-c473044", "right_context_start_lineno": 124, "task_id": "project_cc_python/2511" }
{ "list": [ { "filename": "src/auditnlg/explain.py", "retrieved_chunk": " output = sample[\"output\"])\n if \"openai\" in method:\n result = model.generate(input_, **gen_kwargs)\n else:\n result = model.generate(input_)\n explanations.append(result...
score(instruction, target, prompt)
{ "list": [ { "filename": "src/auditnlg/factualness/summac_utils/evaluator.py", "retrieved_chunk": " sample['knowledge'],\n retrieve_global_knowledge(sample['output'], global_knowledge)])\n sources.append(source)\n gen...
''' * Copyright (c) 2023, Salesforce, Inc. * All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause ''' import os import shutil import gdown import wget from .overall_model import QAFactEval from ....
scores = [x[0]['qa-eval']['lerc_quip'] if x[1][0] else 0.5 for x in results] meta = [] for x in results: answer_list = [y for y in x[1][0] if y["prediction"]["f1"] <= 0.60] meta.append(answer_list) return scores, meta
{ "context_start_lineno": 0, "file": "src/auditnlg/factualness/qafacteval_utils/evaluator.py", "groundtruth_start_lineno": 56, "repository": "salesforce-AuditNLG-c473044", "right_context_start_lineno": 57, "task_id": "project_cc_python/2512" }
{ "list": [ { "filename": "src/auditnlg/factualness/summac_utils/evaluator.py", "retrieved_chunk": " sample['knowledge'],\n retrieve_global_knowledge(sample['output'], global_knowledge)])\n sources.append(source)\n gen...
score_batch_qafacteval(sources, generateds, return_qa_pairs=True)
{ "list": [ { "filename": "src/auditnlg/regeneration/prompt_expansion/prompt_with_guidelines.py", "retrieved_chunk": " model_input=model_input, model_output=model_output\n )\n gen_param = {\n \"max_tokens\": 128\n }\n guideline = model.generate(prompt, **gen_param)\n promp...
''' * Copyright (c) 2023, Salesforce, Inc. * All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause ''' import re import numpy as np from ..llm_wrapper import OpenAILLM, LocalLLM from .openAIscorers...
i = result['logprobs']['text_offset'].index(len(input_) - 1) if i == 0: i = 1 loss = -sum(result['logprobs']["token_logprobs"][i:-1]) avg_loss = loss / (len(result['logprobs']['text_offset']) - i - 1) # 1 is the last '.' ppl = np.exp(avg_loss) scores.append(...
{ "context_start_lineno": 0, "file": "src/auditnlg/factualness/classifier.py", "groundtruth_start_lineno": 37, "repository": "salesforce-AuditNLG-c473044", "right_context_start_lineno": 38, "task_id": "project_cc_python/2509" }
{ "list": [ { "filename": "src/auditnlg/explain.py", "retrieved_chunk": " output = sample[\"output\"])\n if \"openai\" in method:\n result = model.generate(input_, **gen_kwargs)\n else:\n result = model.generate(input_)\n explanations.append(result...
responses[-1]['choices'][0]
{ "list": [ { "filename": "src/auditnlg/regeneration/prompt_expansion/prompt_with_guidelines.py", "retrieved_chunk": " model_input=model_input, model_output=model_output\n )\n gen_param = {\n \"max_tokens\": 128\n }\n guideline = model.generate(prompt, **gen_param)\n promp...
''' * Copyright (c) 2023, Salesforce, Inc. * All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause ''' import re import numpy as np from ..llm_wrapper import OpenAILLM, LocalLLM from .openAIscorers...
result = gpt3model.responses[-1]['choices'][0] i = result['logprobs']['text_offset'].index(len(input_) - 1) if i == 0: i = 1 loss = -sum(result['logprobs']["token_logprobs"][i:-1]) avg_loss = loss / (len(result['logprobs']['text_offset']) - i - 1) # 1 is the last '....
{ "context_start_lineno": 0, "file": "src/auditnlg/factualness/classifier.py", "groundtruth_start_lineno": 36, "repository": "salesforce-AuditNLG-c473044", "right_context_start_lineno": 37, "task_id": "project_cc_python/2508" }
{ "list": [ { "filename": "src/auditnlg/factualness/qafacteval_utils/evaluator.py", "retrieved_chunk": " meta.append(answer_list)\n return scores, meta", "score": 19.289572137272607 }, { "filename": "src/auditnlg/factualness/utils.py", "retrieved_chunk": " ...
generate(input_ + target, **gen_param)
{ "list": [ { "filename": "src/auditnlg/safety/safety_generator.py", "retrieved_chunk": " inputs = []\n for sample in data:\n if \"prompt_all\" in sample.keys() and sample[\"prompt_all\"] != \"\":\n input_sample = \"<Text> {} <Context> {}\".format(sample[\"output\"], sample[\"p...
''' * Copyright (c) 2023, Salesforce, Inc. * All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause ''' from .prompts import constraint_identification_prompt, constraint_checking_prompt from ..llm_w...
if 'No Constraints.' in constraints_found: constraint_scores.append(1.0) score_reasoning.append(constraints_found) continue # if specific constraints found prompt_checking = constraint_checking_prompt.format(llm_output=llm_output, constraints=constraints_fo...
{ "context_start_lineno": 0, "file": "src/auditnlg/constraint/constraint_checker.py", "groundtruth_start_lineno": 50, "repository": "salesforce-AuditNLG-c473044", "right_context_start_lineno": 51, "task_id": "project_cc_python/2516" }
{ "list": [ { "filename": "src/auditnlg/safety/safety_generator.py", "retrieved_chunk": " return inputs\ndef batching(inputs, batch_size, tokenizer, max_source_length):\n for i in range(0, len(inputs), batch_size):\n batch = tokenizer(inputs[i:i+batch_size], max_length=max_source_length, ...
generate(prompt=prompt_identification, messages="")
{ "list": [ { "filename": "src/auditnlg/constraint/prompts.py", "retrieved_chunk": "1. Constraint 2: Yes.\n2. Constraint 11: Yes\nInput Text: {llm_output}\n{constraints}\nOutput: \n\"\"\"\nconstraint_checking_prompt = PromptTemplate(\n input_variables=[\"llm_output\", \"constraints\"],\n templat...
''' * Copyright (c) 2023, Salesforce, Inc. * All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause ''' from .prompts import constraint_identification_prompt, constraint_checking_prompt from ..llm_w...
constraints_checked = model.generate(prompt=prompt_checking, messages="") satisfied = [] for constraint in constraints_checked.split("\n"): if constraint and 'Constraint' in constraint: # count the num of yes-es if "Yes" in constraint:...
{ "context_start_lineno": 0, "file": "src/auditnlg/constraint/constraint_checker.py", "groundtruth_start_lineno": 58, "repository": "salesforce-AuditNLG-c473044", "right_context_start_lineno": 59, "task_id": "project_cc_python/2518" }
{ "list": [ { "filename": "src/auditnlg/llm_wrapper.py", "retrieved_chunk": " response = openai.Completion.create(\n model=self.model_name,\n prompt=prompt,\n **gen_kwargs\n )\n message = response[\"choices\"][0][\"text\"]\n...
format(llm_output=llm_output, constraints=constraints_found)
{ "list": [ { "filename": "src/auditnlg/explain.py", "retrieved_chunk": "def llm_explanation(\n method = \"openai/gpt-3.5-turbo\", \n data = [], \n global_knowledge = \"\"\n ):\n if \"openai\" in method:\n model = OpenAILLM(model_name = method.split(\"/\")[-1])\n e...
''' * Copyright (c) 2023, Salesforce, Inc. * All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause ''' from .prompts import constraint_identification_prompt, constraint_checking_prompt from ..llm_w...
else: prompt_identification = constraint_identification_prompt.format(instructions=task) # run the prompt constraints_found = model.generate(prompt=prompt_identification, messages="") if 'No Constraints.' in constraints_found: constraint_scores.append(1.0) ...
{ "context_start_lineno": 0, "file": "src/auditnlg/constraint/constraint_checker.py", "groundtruth_start_lineno": 45, "repository": "salesforce-AuditNLG-c473044", "right_context_start_lineno": 46, "task_id": "project_cc_python/2515" }
{ "list": [ { "filename": "src/auditnlg/explain.py", "retrieved_chunk": " data = example_format_checker(data)\n gen_kwargs = {\n \"max_tokens\": 128\n }\n explanations = []\n for i, sample in enumerate(tqdm(data)):\n input_ = explain_template.format(\n instructi...
format(instructions=prompt_all)
{ "list": [ { "filename": "src/auditnlg/factualness/qafacteval_utils/evaluator.py", "retrieved_chunk": " source = '\\n'.join([get_instruction(sample),\n sample['knowledge'],\n retrieve_global_knowledge(sample['output'], global_kn...
''' * Copyright (c) 2023, Salesforce, Inc. * All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause ''' from .summac.summac.model_summac import SummaCConv from ..utils import retrieve_global_knowled...
return scores
{ "context_start_lineno": 0, "file": "src/auditnlg/factualness/summac_utils/evaluator.py", "groundtruth_start_lineno": 28, "repository": "salesforce-AuditNLG-c473044", "right_context_start_lineno": 29, "task_id": "project_cc_python/2514" }
{ "list": [ { "filename": "src/auditnlg/factualness/qafacteval_utils/evaluator.py", "retrieved_chunk": " meta.append(answer_list)\n return scores, meta", "score": 85.20489949915985 }, { "filename": "src/auditnlg/factualness/unieval_utils/evaluator.py", "retr...
score(sources, generateds)["scores"]
{ "list": [ { "filename": "tests/test_route_users.py", "retrieved_chunk": " assert 'detail' in response.json()\n assert response.json()['detail'] == messages.MSC404_USER_NOT_FOUND \n response = client.put('api/users/me/1/', headers=headers, json=admin)\n assert response.status_code == stat...
from unittest.mock import MagicMock import pytest from fastapi.testclient import TestClient from sqlalchemy import create_engine, select from sqlalchemy.orm import sessionmaker from main import app from src.database.models import Base, Role, User from src.database.db import get_db SQLALCHEMY_DATABASE_URL = 'sqlite:...
current_user.confirmed = True session.commit() response = client.post( '/api/auth/login', data={'username': admin.get('email'), 'password': admin.get('password')}, ) return response.json()['access_token'] @pytest.fixtur...
{ "context_start_lineno": 0, "file": "tests/conftest.py", "groundtruth_start_lineno": 131, "repository": "theneonwhale-frt-photo-share-6659afb", "right_context_start_lineno": 132, "task_id": "project_cc_python/2562" }
{ "list": [ { "filename": "tests/test_route_auth.py", "retrieved_chunk": " assert response1.status_code == status.HTTP_401_UNAUTHORIZED\n assert response1.json()['detail'] == m.MSC401_EMAIL\n response2 = client.post('api/auth/login', data={'username': user.get('email'), 'password': user.get('...
email == admin['email']))
{ "list": [ { "filename": "Live_Addon/ui/n_panel.py", "retrieved_chunk": " else:\n if os.path.exists(bpy.context.window_manager.my_addon_props.file_path):\n last_save_time = datetime.datetime.fromtimestamp(os.path.getmtime(bpy.context.window_manager.my_addon_props.file...
import os import bpy import platform import datetime import re import tempfile from .. import props from . import common def get_default_path(): if platform.system() == "Windows": appdata_folder = os.path.join(os.environ["APPDATA"]) else: appdata_folder = os.path.join(os.environ["HOME"], ".con...
return image_file_path def make_image_file_path(image_name): image_dir = make_image_dir() currrent_time = datetime.datetime.now().strftime("%Y.%m.%d_%H-%M-%S") image_file_path = os.path.join(image_dir, image_name + currrent_time + common.file_extension_format()) return image_f...
{ "context_start_lineno": 0, "file": "Live_Addon/utils/file_path.py", "groundtruth_start_lineno": 78, "repository": "pro470-Live-Save-for-Blender-3b5c1c2", "right_context_start_lineno": 79, "task_id": "project_cc_python/2617" }
{ "list": [ { "filename": "Live_Addon/ui/n_panel.py", "retrieved_chunk": " else:\n if os.path.exists(bpy.context.window_manager.my_addon_props.file_path):\n last_save_time = datetime.datetime.fromtimestamp(os.path.getmtime(bpy.context.window_manager.my_addon_props.file...
file_extension_format())
{ "list": [ { "filename": "src/database/db.py", "retrieved_chunk": "from src.conf.messages import MSC500_DATABASE_CONFIG, MSC500_DATABASE_CONNECT\nfrom src.services.asyncdevlogging import async_logging_to_file\nURI = settings.sqlalchemy_database_url\nengine = create_engine(URI, echo=True)\nDBSession =...
from unittest.mock import MagicMock import pytest from fastapi.testclient import TestClient from sqlalchemy import create_engine, select from sqlalchemy.orm import sessionmaker from main import app from src.database.models import Base, Role, User from src.database.db import get_db SQLALCHEMY_DATABASE_URL = 'sqlite:...
Base.metadata.create_all(bind=engine) db = TestingSessionLocal() try: yield db finally: db.close() @pytest.fixture(scope='module') def client(session): def override_get_db(): try: yield session finally: session.close() app.dependency_o...
{ "context_start_lineno": 0, "file": "tests/conftest.py", "groundtruth_start_lineno": 26, "repository": "theneonwhale-frt-photo-share-6659afb", "right_context_start_lineno": 27, "task_id": "project_cc_python/2558" }
{ "list": [ { "filename": "src/database/db.py", "retrieved_chunk": " except SQLAlchemyError as err:\n db.rollback()\n raise HTTPException(status_code=status.HTTP_400_BAD_REQUEST, detail=str(err))\n finally:\n db.close()\ndef get_redis(is_async: bool = True):\n module = {\...
metadata.drop_all(bind=engine)
{ "list": [ { "filename": "docs_src/queries/queries.py", "retrieved_chunk": "values = [\n {\"name\": \"databasez\", \"address\": \"London, United Kingdom\"},\n {\"name\": \"another name\", \"address\": \"The Hague, Netherlands\"},\n]\nawait database.execute_many(query=query, values=values)\n# Fe...
from databasez import Database database = Database("postgresql+asyncpg://localhost/example") # Establish the connection pool await database.connect() # Execute query = "INSERT INTO users(name, address) VALUES (:name, :address)" values = {"text": "databasez", "address": "London, United Kingdom"} await database.execu...
# Fetch single row query = "SELECT * FROM users WHERE id = :id" result = await database.fetch_one(query=query, values={"id": 1})
{ "context_start_lineno": 0, "file": "docs_src/queries/raw_queries.py", "groundtruth_start_lineno": 23, "repository": "tarsil-databasez-01b23b0", "right_context_start_lineno": 24, "task_id": "project_cc_python/2625" }
{ "list": [ { "filename": "docs_src/queries/queries.py", "retrieved_chunk": "row = await database.fetch_one(query=query)\n# Fetch single value, defaults to `column=0`.\nquery = users.select()\nvalue = await database.fetch_val(query=query)\n# Fetch multiple rows without loading them all into memory at ...
fetch_all(query=query, values={"address": "London, United Kingdom"})
{ "list": [ { "filename": "src/somesy/cli/init.py", "retrieved_chunk": " options[\"debug\"] = typer.confirm(\"Do you want to show debug logs?\")\n set_log_level(\n SomesyLogLevel.from_flags(\n debug=options[\"debug\"],\n verbose=options[\"verbose\"],\n inf...
"""CLI command to initialize somesy configuration file.""" import logging from pathlib import Path import tomlkit from somesy.core.core import get_input_content from somesy.core.models import SomesyInput logger = logging.getLogger("somesy") def init_config(input_path: Path, options: dict) -> None: """Initializ...
input_file_type = "somesy" if is_somesy else "pyproject" msg = f"Found input file with {input_file_type} format." logger.verbose(msg) logger.debug(f"Input file content: {options}") if "input_file" in options: del options["input_file"] if is_somesy: content["config"] = options ...
{ "context_start_lineno": 0, "file": "src/somesy/commands/init_config.py", "groundtruth_start_lineno": 23, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 24, "task_id": "project_cc_python/2570" }
{ "list": [ { "filename": "src/somesy/cli/init.py", "retrieved_chunk": " logger.info(\n f\"[bold green]Input file is updated/created at {input_file}[/bold green]\"\n )", "score": 48.03040138104996 }, { "filename": "tests/commands/test_init_config.py", "retrieved_...
is_somesy_file_path(input_path)
{ "list": [ { "filename": "src/somesy/core/writer.py", "retrieved_chunk": " self._set_property(self._get_key(\"description\"), description)\n @property\n def authors(self):\n \"\"\"Return the authors of the project.\"\"\"\n return self._get_property(self._get_key(\"authors\"...
"""package.json parser and saver.""" import json import logging from collections import OrderedDict from pathlib import Path from typing import List, Optional from rich.pretty import pretty_repr from somesy.core.models import Person, ProjectMetadata from somesy.core.writer import ProjectMetadataWriter from somesy.pac...
@authors.setter def authors(self, authors: List[Person]) -> None: """Set the authors of the project.""" authors = self._from_person(authors[0]) self._set_property(self._get_key("authors"), authors) @property def contributors(self): """Return the contributors of the pac...
{ "context_start_lineno": 0, "file": "src/somesy/package_json/writer.py", "groundtruth_start_lineno": 35, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 36, "task_id": "project_cc_python/2577" }
{ "list": [ { "filename": "src/somesy/cff/writer.py", "retrieved_chunk": " def _init_new_file(self):\n \"\"\"Initialize new CFF file.\"\"\"\n self._data = {\n \"cff-version\": \"1.2.0\",\n \"message\": \"If you use this software, please cite it using these metada...
_get_property(self._get_key("authors"))]
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " \"family-names\": \"Doe\",\n \"orcid\": \"https://orcid.org/0123-4567-8910\",\n }\n ret = Person(**p)\n ret.set_key_order(list(p.keys())) # custom order!\n return ret\ndef test_from_to_person(pers...
from pathlib import Path import pytest from somesy.core.models import Person, SomesyInput from somesy.pyproject.writer import Poetry, Pyproject, SetupTools @pytest.fixture def poetry_path(): return Path("tests/pyproject/data/pyproject.full.toml") @pytest.fixture def pyproject_poetry(poetry_path): return P...
assert p.full_name == person.full_name assert p.email == person.email
{ "context_start_lineno": 0, "file": "tests/pyproject/test_pyproject_misc.py", "groundtruth_start_lineno": 104, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 105, "task_id": "project_cc_python/2592" }
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " assert p.email == person.email\n assert p.orcid == person.orcid\ndef test_person_merge(tmp_path, person):\n def to_cff_keys(lst):\n return list(map(lambda s: s.replace(\"_\", \"-\"), lst))\n cff_path = tm...
_to_person(SetupTools._from_person(person))
{ "list": [ { "filename": "tests/cli/test_command_init_config.py", "retrieved_chunk": "from pathlib import Path\nfrom typer.testing import CliRunner\nfrom somesy.main import app\nrunner = CliRunner()\ndef test_cli_config_init(tmp_path, create_poetry_file, create_package_json):\n input_file = tmp_pa...
from pathlib import Path import pytest from somesy.core.core import discover_input from somesy.core.models import ProjectMetadata, SomesyConfig, SomesyInput @pytest.fixture def somesy_metadata_only(): return Path("tests/core/data/.somesy.toml") @pytest.fixture def somesy_with_config(): return Path("tests/...
assert isinstance(metadata, ProjectMetadata) assert metadata.name == "somesy" assert metadata.version == "0.1.0" def test_with_extract_cli_config(somesy_with_config, somesy_metadata_only): # test with config exists config = SomesyInput.from_input_file(somesy_with_config).config assert isinsta...
{ "context_start_lineno": 0, "file": "tests/core/test_core_core.py", "groundtruth_start_lineno": 37, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 38, "task_id": "project_cc_python/2601" }
{ "list": [ { "filename": "tests/cli/test_command_sync.py", "retrieved_chunk": " )\n assert result.exit_code == 1\n assert \"No somesy input file found.\" in result.stdout", "score": 57.628101471031385 }, { "filename": "tests/cli/test_command_init_config.py", "retrie...
from_input_file(somesy_metadata_only).project
{ "list": [ { "filename": "tests/test_connection_options.py", "retrieved_chunk": "from databasez.core import DatabaseURL\ndef test_postgres_pool_size():\n backend = PostgresBackend(\"postgres://localhost/database?min_size=1&max_size=20\")\n kwargs = backend._get_connection_kwargs()\n assert k...
from urllib.parse import quote import pytest from databasez import DatabaseURL def test_database_url_repr(): u = DatabaseURL("postgresql://localhost/name") assert repr(u) == "DatabaseURL('postgresql://localhost/name')" u = DatabaseURL("postgresql://username@localhost/name") assert repr(u) == "Datab...
assert u.username == "username" assert u.password == "password" assert u.hostname == "localhost" assert u.port == 123 assert u.database == "mydatabase" u = DatabaseURL( "postgresql://username:password@/mydatabase?host=/var/run/postgresql/.s.PGSQL.5432" ) assert u.dialect == "po...
{ "context_start_lineno": 0, "file": "tests/test_database_url.py", "groundtruth_start_lineno": 24, "repository": "tarsil-databasez-01b23b0", "right_context_start_lineno": 25, "task_id": "project_cc_python/2657" }
{ "list": [ { "filename": "tests/test_connection_options.py", "retrieved_chunk": " backend = PostgresBackend(url + \"?min_size=1&max_size=20\")\n await backend.connect()\n await backend.disconnect()\ndef test_postgres_explicit_pool_size():\n backend = PostgresBackend(\"postgres...
driver == "asyncpg"
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " assert file_path.is_file()\n reject_path = Path(\"tests/cff/data/reject.cff\")\n with pytest.raises(ValueError):\n CFF(reject_path)\n@pytest.fixture\ndef cff():\n return CFF(Path(\"tests/cff/data/CITATION...
from pathlib import Path import pytest from somesy.core.log import SomesyLogLevel, set_log_level from somesy.core.models import SomesyInput from somesy.package_json.writer import PackageJSON @pytest.fixture(scope="session", autouse=True) def init_somesy_logger(): set_log_level(SomesyLogLevel.DEBUG) @pytest.fi...
@pytest.fixture def package_json() -> PackageJSON: return PackageJSON(Path("tests/data/package.json"))
{ "context_start_lineno": 0, "file": "tests/conftest.py", "groundtruth_start_lineno": 88, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 89, "task_id": "project_cc_python/2585" }
{ "list": [ { "filename": "src/somesy/commands/init_config.py", "retrieved_chunk": " with open(input_path, \"w\") as f:\n tomlkit.dump(content, f)\n logger.info(f\"Input file ({input_path}) updated.\")\n logger.debug(f\"Input file content: {content}\")", "score": 44.1930529972136...
from_input_file(Path("tests/data/somesy.toml"))
{ "list": [ { "filename": "src/somesy/core/writer.py", "retrieved_chunk": " self._set_property(self._get_key(\"description\"), description)\n @property\n def authors(self):\n \"\"\"Return the authors of the project.\"\"\"\n return self._get_property(self._get_key(\"authors\"...
"""package.json parser and saver.""" import json import logging from collections import OrderedDict from pathlib import Path from typing import List, Optional from rich.pretty import pretty_repr from somesy.core.models import Person, ProjectMetadata from somesy.core.writer import ProjectMetadataWriter from somesy.pac...
@authors.setter def authors(self, authors: List[Person]) -> None: """Set the authors of the project.""" authors = self._from_person(authors[0]) self._set_property(self._get_key("authors"), authors) @property def contributors(self): """Return the contributors of the pac...
{ "context_start_lineno": 0, "file": "src/somesy/package_json/writer.py", "groundtruth_start_lineno": 35, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 36, "task_id": "project_cc_python/2578" }
{ "list": [ { "filename": "src/somesy/cff/writer.py", "retrieved_chunk": " mappings = {\n \"name\": [\"title\"],\n \"description\": [\"abstract\"],\n \"homepage\": [\"url\"],\n \"repository\": [\"repository-code\"],\n \"maintainers\": [\"co...
_get_key("authors"))]
{ "list": [ { "filename": "src/somesy/core/writer.py", "retrieved_chunk": " This method is existing for the publication_author special case\n when synchronizing to CITATION.cff.\n \"\"\"\n self.authors = self._sync_person_list(self.authors, metadata.authors())\n def sync...
"""package.json parser and saver.""" import json import logging from collections import OrderedDict from pathlib import Path from typing import List, Optional from rich.pretty import pretty_repr from somesy.core.models import Person, ProjectMetadata from somesy.core.writer import ProjectMetadataWriter from somesy.pac...
if metadata.repository: self.repository = {"type": "git", "url": metadata.repository}
{ "context_start_lineno": 0, "file": "src/somesy/package_json/writer.py", "groundtruth_start_lineno": 110, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 111, "task_id": "project_cc_python/2583" }
{ "list": [ { "filename": "src/somesy/pyproject/writer.py", "retrieved_chunk": " )\n # NOTE: for our purposes, does not matter what are given or family names,\n # we only compare on full_name anyway.\n return Person(\n **{\n \"given-names\": \" \"....
_sync_person_list(self.contributors, metadata.people)
{ "list": [ { "filename": "src/somesy/core/writer.py", "retrieved_chunk": " self._set_property(self._get_key(\"description\"), description)\n @property\n def authors(self):\n \"\"\"Return the authors of the project.\"\"\"\n return self._get_property(self._get_key(\"authors\"...
"""package.json parser and saver.""" import json import logging from collections import OrderedDict from pathlib import Path from typing import List, Optional from rich.pretty import pretty_repr from somesy.core.models import Person, ProjectMetadata from somesy.core.writer import ProjectMetadataWriter from somesy.pac...
self._data = json.load(f, object_pairs_hook=OrderedDict) def _validate(self) -> None: """Validate package.json content using pydantic class.""" config = dict(self._get_property([])) logger.debug( f"Validating config using {PackageJsonConfig.__name__}: {pretty_repr(c...
{ "context_start_lineno": 0, "file": "src/somesy/package_json/writer.py", "groundtruth_start_lineno": 56, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 57, "task_id": "project_cc_python/2580" }
{ "list": [ { "filename": "src/somesy/core/writer.py", "retrieved_chunk": " @property\n def maintainers(self):\n \"\"\"Return the maintainers of the project.\"\"\"\n return self._get_property(self._get_key(\"maintainers\"))\n @maintainers.setter\n def maintainers(self, mainta...
path.open() as f:
{ "list": [ { "filename": "databasez/core.py", "retrieved_chunk": " username = kwargs.pop(\"username\", self.components.username)\n password = kwargs.pop(\"password\", self.components.password)\n netloc = hostname\n if port is not None:\n netl...
from urllib.parse import quote import pytest from databasez import DatabaseURL def test_database_url_repr(): u = DatabaseURL("postgresql://localhost/name") assert repr(u) == "DatabaseURL('postgresql://localhost/name')" u = DatabaseURL("postgresql://username@localhost/name") assert repr(u) == "Datab...
assert u.password == "password" assert u.hostname == "localhost" assert u.port == 123 assert u.database == "mydatabase" u = DatabaseURL( "postgresql://username:password@/mydatabase?host=/var/run/postgresql/.s.PGSQL.5432" ) assert u.dialect == "postgresql" assert u.username == "...
{ "context_start_lineno": 0, "file": "tests/test_database_url.py", "groundtruth_start_lineno": 25, "repository": "tarsil-databasez-01b23b0", "right_context_start_lineno": 26, "task_id": "project_cc_python/2658" }
{ "list": [ { "filename": "tests/test_connection_options.py", "retrieved_chunk": " backend = PostgresBackend(url + \"?min_size=1&max_size=20\")\n await backend.connect()\n await backend.disconnect()\ndef test_postgres_explicit_pool_size():\n backend = PostgresBackend(\"postgres...
username == "username"
{ "list": [ { "filename": "src/somesy/pyproject/writer.py", "retrieved_chunk": " }\n super().__init__(\n path, section=section, direct_mappings=mappings, model_cls=SetuptoolsConfig\n )\n @staticmethod\n def _from_person(person: Person):\n \"\"\"Convert proj...
"""package.json parser and saver.""" import json import logging from collections import OrderedDict from pathlib import Path from typing import List, Optional from rich.pretty import pretty_repr from somesy.core.models import Person, ProjectMetadata from somesy.core.writer import ProjectMetadataWriter from somesy.pac...
names = list(map(lambda s: s.strip(), person["name"].split())) person_obj = { "given-names": " ".join(names[:-1]), "family-names": names[-1], } if "email" in person: person_obj["email"] = person["email"].strip() if "url" in person: ...
{ "context_start_lineno": 0, "file": "src/somesy/package_json/writer.py", "groundtruth_start_lineno": 91, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 92, "task_id": "project_cc_python/2581" }
{ "list": [ { "filename": "src/somesy/pyproject/writer.py", "retrieved_chunk": " \"\"\"Parse setuptools person string to a Person.\"\"\"\n # NOTE: for our purposes, does not matter what are given or family names,\n # we only compare on full_name anyway.\n names = list(map(l...
convert_author(person).dict(exclude_none=True)
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " \"family-names\": \"Doe\",\n \"orcid\": \"https://orcid.org/0123-4567-8910\",\n }\n ret = Person(**p)\n ret.set_key_order(list(p.keys())) # custom order!\n return ret\ndef test_from_to_person(pers...
import json from pathlib import Path import pytest from somesy.core.models import Person, ProjectMetadata, SomesyInput p1 = { "given-names": "Jane", "family-names": "Doe", "email": "j.doe@example.com", "orcid": "https://orcid.org/0123-4567-8910", } p2 = {"given-names": "Foo", "family-names": "Bar", "...
{ "context_start_lineno": 0, "file": "tests/core/test_core_models.py", "groundtruth_start_lineno": 107, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 108, "task_id": "project_cc_python/2600" }
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " assert p.email == person.email\n assert p.orcid == person.orcid\ndef test_person_merge(tmp_path, person):\n def to_cff_keys(lst):\n return list(map(lambda s: s.replace(\"_\", \"-\"), lst))\n cff_path = tm...
copy()._key_order == p._key_order
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " \"family-names\": \"Doe\",\n \"orcid\": \"https://orcid.org/0123-4567-8910\",\n }\n ret = Person(**p)\n ret.set_key_order(list(p.keys())) # custom order!\n return ret\ndef test_from_to_person(pers...
from pathlib import Path import pytest from somesy.core.models import Person, SomesyInput from somesy.pyproject.writer import Poetry, Pyproject, SetupTools @pytest.fixture def poetry_path(): return Path("tests/pyproject/data/pyproject.full.toml") @pytest.fixture def pyproject_poetry(poetry_path): return P...
assert p.full_name == person.full_name assert p.email == person.email def test_setuptools_from_to_person(person): p = SetupTools._to_person(SetupTools._from_person(person)) assert p.full_name == person.full_name assert p.email == person.email
{ "context_start_lineno": 0, "file": "tests/pyproject/test_pyproject_misc.py", "groundtruth_start_lineno": 98, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 99, "task_id": "project_cc_python/2591" }
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " assert p.email == person.email\n assert p.orcid == person.orcid\ndef test_person_merge(tmp_path, person):\n def to_cff_keys(lst):\n return list(map(lambda s: s.replace(\"_\", \"-\"), lst))\n cff_path = tm...
_to_person(Poetry._from_person(person))
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " \"given-names\": \"AA\",\n \"email\": \"test@test.teset\",\n \"orcid\": \"https://orcid.org/0000-0001-2345-6789\",\n }\n return Person(**people)\n@pytest.fixture\ndef project_metadata():\n retu...
from pathlib import Path import pytest from somesy.core.models import Person, SomesyInput from somesy.pyproject.writer import Poetry, Pyproject, SetupTools @pytest.fixture def poetry_path(): return Path("tests/pyproject/data/pyproject.full.toml") @pytest.fixture def pyproject_poetry(poetry_path): return P...
def test_pyproject_init_path(pyproject_poetry, poetry_path): # test if pyproject object is wrapped with Poetry object assert pyproject_poetry.path == poetry_path def test_pyproject_init(tmp_path): path = tmp_path / "pyproject.toml" # test if it gives error when pyproject.toml file is not found ...
{ "context_start_lineno": 0, "file": "tests/pyproject/test_pyproject_misc.py", "groundtruth_start_lineno": 30, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 31, "task_id": "project_cc_python/2588" }
{ "list": [ { "filename": "tests/conftest.py", "retrieved_chunk": " f2.write(content)\n yield _create_cff_file\n@pytest.fixture\ndef somesy() -> dict:\n return SomesyInput.from_input_file(Path(\"tests/data/somesy.toml\"))\n@pytest.fixture\ndef package_json() -> PackageJSON:\n r...
from_input_file(poetry_path).project
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " \"family-names\": \"Doe\",\n \"orcid\": \"https://orcid.org/0123-4567-8910\",\n }\n ret = Person(**p)\n ret.set_key_order(list(p.keys())) # custom order!\n return ret\ndef test_from_to_person(pers...
import json from pathlib import Path import pytest from somesy.core.models import Person, ProjectMetadata, SomesyInput p1 = { "given-names": "Jane", "family-names": "Doe", "email": "j.doe@example.com", "orcid": "https://orcid.org/0123-4567-8910", } p2 = {"given-names": "Foo", "family-names": "Bar", "...
assert list(json.loads(p.json(exclude_none=True)).keys()) == expected_order # added field appears in right spot p.orcid = "https://orcid.org/1234-5678-9101" assert list(p.dict(exclude_none=True).keys()) == p._key_order assert list(json.loads(p.json(exclude_none=True)).keys()) == p._key_order ...
{ "context_start_lineno": 0, "file": "tests/core/test_core_models.py", "groundtruth_start_lineno": 92, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 93, "task_id": "project_cc_python/2597" }
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " \"family-names\": \"Doe\",\n \"orcid\": \"https://orcid.org/0123-4567-8910\",\n }\n ret = Person(**p)\n ret.set_key_order(list(p.keys())) # custom order!\n return ret\ndef test_from_to_person(pers...
dict(exclude_none=True).keys()) == expected_order
{ "list": [ { "filename": "src/somesy/core/models.py", "retrieved_chunk": " if self.name_particle:\n names.append(self.name_particle)\n if self.family_names:\n names.append(self.family_names)\n if self.name_suffix:\n names.append(self.name_suffix)\...
import json from pathlib import Path import pytest from somesy.core.models import Person, ProjectMetadata, SomesyInput p1 = { "given-names": "Jane", "family-names": "Doe", "email": "j.doe@example.com", "orcid": "https://orcid.org/0123-4567-8910", } p2 = {"given-names": "Foo", "family-names": "Bar", "...
meta = metadata.copy() meta.people.append(p1) ProjectMetadata(**meta.dict()) # P1 ~= P3 meta.people.append(p3) with pytest.raises(ValueError): ProjectMetadata(**meta.dict()) # P1 ~= P4 meta.people.pop() meta.people.append(p4) with pytest.raises(ValueError): Pr...
{ "context_start_lineno": 0, "file": "tests/core/test_core_models.py", "groundtruth_start_lineno": 40, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 41, "task_id": "project_cc_python/2595" }
{ "list": [ { "filename": "src/somesy/core/models.py", "retrieved_chunk": " \"\"\"\n if self.orcid is not None and other.orcid is not None:\n # having orcids is the best case, a real identifier\n return self.orcid == other.orcid\n # otherwise, try to match ac...
from_input_file(Path("tests/core/data/.somesy.toml")).project
{ "list": [ { "filename": "src/somesy/core/log.py", "retrieved_chunk": "_log_level: Optional[SomesyLogLevel] = None\ndef get_log_level() -> Optional[SomesyLogLevel]:\n \"\"\"Return current user-defined log level.\"\"\"\n return _log_level\ndef set_log_level(log_level: SomesyLogLevel) -> None:\n ...
"""Utility functions for CLI commands.""" import logging import traceback from typing import Optional import typer import wrapt from rich.pretty import pretty_repr from somesy.core.core import discover_input from somesy.core.log import SomesyLogLevel, get_log_level, set_log_level from somesy.core.models import Somesy...
somesy_input: SomesyInput = somesy_conf.get_input() if cli_log_level is None: # no cli log level -> set it according to the loaded configuration set_log_level(somesy_input.config.log_level()) logger.debug( f"Combined config (Defaults + File + CLI):\n{pretty_repr(somesy_input.conf...
{ "context_start_lineno": 0, "file": "src/somesy/cli/util.py", "groundtruth_start_lineno": 45, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 46, "task_id": "project_cc_python/2568" }
{ "list": [ { "filename": "src/somesy/core/log.py", "retrieved_chunk": " init_log()\n # set the current logging log level\n logger.setLevel(SomesyLogLevel.to_logging(log_level))\ndef init_log():\n \"\"\"Initialize logging (add VERBOSE log level and Rich formatter).\"\"\"\n _add_verbose_...
update_log_level(cli_log_level)
{ "list": [ { "filename": "src/somesy/pyproject/writer.py", "retrieved_chunk": " )\n # NOTE: for our purposes, does not matter what are given or family names,\n # we only compare on full_name anyway.\n return Person(\n **{\n \"given-names\": \" \"....
import json from pathlib import Path import pytest from somesy.core.models import Person, ProjectMetadata, SomesyInput p1 = { "given-names": "Jane", "family-names": "Doe", "email": "j.doe@example.com", "orcid": "https://orcid.org/0123-4567-8910", } p2 = {"given-names": "Foo", "family-names": "Bar", "...
# correct subsequence of order expected_order = ["given_names", "family_names", "email"] assert list(p.dict(exclude_none=True).keys()) == expected_order assert list(json.loads(p.json(exclude_none=True)).keys()) == expected_order # added field appears in right spot p.orcid = "https://orcid.org...
{ "context_start_lineno": 0, "file": "tests/core/test_core_models.py", "groundtruth_start_lineno": 88, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 89, "task_id": "project_cc_python/2596" }
{ "list": [ { "filename": "src/somesy/pyproject/writer.py", "retrieved_chunk": "class SetupTools(PyprojectCommon):\n \"\"\"Setuptools config file handler parsed from setup.cfg.\"\"\"\n def __init__(self, path: Path):\n \"\"\"Setuptools config file handler parsed from pyproject.toml.\n ...
set_key_order(key_order)
{ "list": [ { "filename": "src/somesy/core/models.py", "retrieved_chunk": " extra = Extra.ignore\n @validator(\"people\")\n def ensure_distinct_people(cls, people):\n \"\"\"Make sure that no person is listed twice in the same person list.\"\"\"\n for i in range(len(people)):...
import json from pathlib import Path import pytest from somesy.core.models import Person, ProjectMetadata, SomesyInput p1 = { "given-names": "Jane", "family-names": "Doe", "email": "j.doe@example.com", "orcid": "https://orcid.org/0123-4567-8910", } p2 = {"given-names": "Foo", "family-names": "Bar", "...
# missing orcid, different mail, different name -> not same assert not Person(**p1).same_person(Person(**p2)) # missing orcid, different mail, same name -> same assert Person(**p1).same_person(Person(**p3)) # missing orcid, same mail -> same assert Person(**p1).same_person(Person(**p4)) # d...
{ "context_start_lineno": 0, "file": "tests/core/test_core_models.py", "groundtruth_start_lineno": 26, "repository": "Materials-Data-Science-and-Informatics-somesy-a47ca93", "right_context_start_lineno": 27, "task_id": "project_cc_python/2594" }
{ "list": [ { "filename": "tests/cff/test_cff_writer.py", "retrieved_chunk": " \"given-names\": \"AA\",\n \"email\": \"test@test.teset\",\n \"orcid\": \"https://orcid.org/0000-0001-2345-6789\",\n }\n return Person(**people)\n@pytest.fixture\ndef project_metadata():\n retu...
same_person(Person(**p1))
{ "list": [ { "filename": "src/contour_flow_net.py", "retrieved_chunk": " # split segmentations\n segmentation1 = segmentations[:, 0, :, :, :]\n segmentation2 = segmentations[:, 1, :, :, :]\n # Compute flow in frame of image1.\n # noinspection PyCallingNonCallable\n flow = self._flow...
# coding=utf-8 # Copyright 2023 Junbong Jang. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agree...
warped2 = uflow_utils.resample(features2, warp_up) # --------------- Compute cost volume by comparing features1 and warped features2. features1_normalized, warped2_normalized = normalize_features( [features1, warped2], normalize=self._normalize_before_cost_volume, cen...
{ "context_start_lineno": 0, "file": "src/contour_flow_model.py", "groundtruth_start_lineno": 219, "repository": "JunbongJang-contour-tracking-1219b66", "right_context_start_lineno": 220, "task_id": "project_cc_python/2681" }
{ "list": [ { "filename": "src/contour_flow_net.py", "retrieved_chunk": " # originally, the shape is [1, 160, 160, 1] before the resize\n warps = uflow_utils.resize(\n warps, orig_height, orig_width, is_flow=False)\n valid_warp_masks = uflow_utils.resize(\n valid_warp_...
flow_to_warp(flow_up)
{ "list": [ { "filename": "src/uflow_augmentation.py", "retrieved_chunk": " new_height = tf.cast(\n tf.math.ceil(tf.cast(orig_height, tf.float32) * scale), tf.int32)\n new_width = tf.cast(\n tf.math.ceil(tf.cast(orig_width, tf.float32) * scale), tf.int32)\n # rescale the images (and flow)...
# coding=utf-8 # Copyright 2023 Junbong Jang. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agree...
if self._num_context_up_channels: context_up = self._context_up_layers[level](context) # Append results to list. flows.insert(0, flow) # Refine flow at level 1. refinement = self._refine_model([context, flow]) if (training and self._drop_out_rate): refinement *= tf.cast( ...
{ "context_start_lineno": 0, "file": "src/contour_flow_model.py", "groundtruth_start_lineno": 285, "repository": "JunbongJang-contour-tracking-1219b66", "right_context_start_lineno": 286, "task_id": "project_cc_python/2683" }
{ "list": [ { "filename": "src/uflow_augmentation.py", "retrieved_chunk": "def random_scale_second(\n images, flow=None, mask=None, min_scale=1.0, max_scale=1.0):\n \"\"\"Performs a random scaling on the second image in the given range.\"\"\"\n # choose a random scale factor and compute new resol...
upsample(flow, is_flow=True)
{ "list": [ { "filename": "src/contour_flow_net.py", "retrieved_chunk": " # split segmentations\n segmentation1 = segmentations[:, 0, :, :, :]\n segmentation2 = segmentations[:, 1, :, :, :]\n # Compute flow in frame of image1.\n # noinspection PyCallingNonCallable\n flow = self._flow...
# coding=utf-8 # Copyright 2023 Junbong Jang. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agree...
# --------------- Compute cost volume by comparing features1 and warped features2. features1_normalized, warped2_normalized = normalize_features( [features1, warped2], normalize=self._normalize_before_cost_volume, center=self._normalize_before_cost_volume, moments_a...
{ "context_start_lineno": 0, "file": "src/contour_flow_model.py", "groundtruth_start_lineno": 220, "repository": "JunbongJang-contour-tracking-1219b66", "right_context_start_lineno": 221, "task_id": "project_cc_python/2682" }
{ "list": [ { "filename": "src/data/data_utils.py", "retrieved_chunk": " output.append(flow_uv)\n # create valid mask\n flow_valid = tf.ones_like(flow_uv[Ellipsis, :1], dtype=tf.float32)\n output.append(flow_valid)\n if include_occlusion:\n occlusion_mask = deserialize(context_parsed['...
resample(features2, warp_up)
{ "list": [ { "filename": "src/data/data_utils.py", "retrieved_chunk": " output = [images]\n if include_flow:\n flow_uv = deserialize(context_parsed['flow_uv'], tf.float32, 2)\n flow_uv = flow_uv[Ellipsis, ::-1]\n if height is not None and width is not None and resize_gt_flow:\n flow_u...
# coding=utf-8 # Copyright 2023 Junbong Jang. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agree...
def save_result_to_npy(a_result, save_path): if index == 1: list_of_result = np.expand_dims(a_result, axis=0) else: list_of_result = np.load(save_path, allow_pickle=True) list_of_result = np.append([a_result], list_of_result, axis=0) np.save(save_path, list_of_result, allow...
{ "context_start_lineno": 0, "file": "src/uflow_plotting.py", "groundtruth_start_lineno": 386, "repository": "JunbongJang-contour-tracking-1219b66", "right_context_start_lineno": 387, "task_id": "project_cc_python/2684" }
{ "list": [ { "filename": "src/data/data_utils.py", "retrieved_chunk": " output.append(flow_uv)\n # create valid mask\n flow_valid = tf.ones_like(flow_uv[Ellipsis, :1], dtype=tf.float32)\n output.append(flow_valid)\n if include_occlusion:\n occlusion_mask = deserialize(context_parsed['...
flow_to_warp_np(flow_uv)
{ "list": [ { "filename": "databasez/core.py", "retrieved_chunk": " username = kwargs.pop(\"username\", self.components.username)\n password = kwargs.pop(\"password\", self.components.password)\n netloc = hostname\n if port is not None:\n netl...
from urllib.parse import quote import pytest from databasez import DatabaseURL def test_database_url_repr(): u = DatabaseURL("postgresql://localhost/name") assert repr(u) == "DatabaseURL('postgresql://localhost/name')" u = DatabaseURL("postgresql://username@localhost/name") assert repr(u) == "Datab...
assert u.port == 123 assert u.database == "mydatabase" u = DatabaseURL( "postgresql://username:password@/mydatabase?host=/var/run/postgresql/.s.PGSQL.5432" ) assert u.dialect == "postgresql" assert u.username == "username" assert u.password == "password" assert u.hostname == "/...
{ "context_start_lineno": 0, "file": "tests/test_database_url.py", "groundtruth_start_lineno": 27, "repository": "tarsil-databasez-01b23b0", "right_context_start_lineno": 28, "task_id": "project_cc_python/2660" }
{ "list": [ { "filename": "databasez/core.py", "retrieved_chunk": " kwargs[\"netloc\"] = netloc\n if \"database\" in kwargs:\n kwargs[\"path\"] = \"/\" + kwargs.pop(\"database\")\n if \"dialect\" in kwargs or \"driver\" in kwargs:\n dialect = kwargs.pop(\...
hostname == "localhost"
{ "list": [ { "filename": "src/preprocessing/main_process_tracking_points.py", "retrieved_chunk": " print('dataset', dataset_folder)\n # --------------------------- Data preprocessing --------------------------------------\n copy_paste_images_to_generated_path(root_assets_path, ro...
# coding=utf-8 # Copyright 2023 Junbong Jang. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agree...
save_fig(plot_dir, index, frame_skip, name='predicted_warped_image', cv2_imwrite_data=warped_image1.numpy()[0]) # Warp contour # tf.unique_with_counts(tf.reshape(warped_contour1, -1)) warped_contour1 = uflow_utils.resample_np(segmentation2, a_warp) # uint8 [ 474, 392, 1] -> uint8 [474, 392, 1] warped_cont...
{ "context_start_lineno": 0, "file": "src/uflow_plotting.py", "groundtruth_start_lineno": 401, "repository": "JunbongJang-contour-tracking-1219b66", "right_context_start_lineno": 402, "task_id": "project_cc_python/2685" }
{ "list": [ { "filename": "src/preprocessing/main_process_tracking_points.py", "retrieved_chunk": " # contour_points_path_list = glob(f\"{root_generated_path}{dataset_folder}/contour_points/*.txt\")\n # img_path_list = glob(f\"{root_assets_path}/{dataset_folder}/{image_folder}/*{image_fo...
resample_np(image2, a_warp)
{ "list": [ { "filename": "databasez/core.py", "retrieved_chunk": " username = kwargs.pop(\"username\", self.components.username)\n password = kwargs.pop(\"password\", self.components.password)\n netloc = hostname\n if port is not None:\n netl...
from urllib.parse import quote import pytest from databasez import DatabaseURL def test_database_url_repr(): u = DatabaseURL("postgresql://localhost/name") assert repr(u) == "DatabaseURL('postgresql://localhost/name')" u = DatabaseURL("postgresql://username@localhost/name") assert repr(u) == "Datab...
u2 = DatabaseURL(u) assert u2.password == "[password" u3 = DatabaseURL(str(u)) assert u3.password == "[password" def test_database_url_constructor(): with pytest.raises(TypeError): DatabaseURL(("postgresql", "username", "password", "localhost", "mydatabase")) u = DatabaseURL("postg...
{ "context_start_lineno": 0, "file": "tests/test_database_url.py", "groundtruth_start_lineno": 50, "repository": "tarsil-databasez-01b23b0", "right_context_start_lineno": 51, "task_id": "project_cc_python/2663" }
{ "list": [ { "filename": "databasez/core.py", "retrieved_chunk": " kwargs[\"netloc\"] = netloc\n if \"database\" in kwargs:\n kwargs[\"path\"] = \"/\" + kwargs.pop(\"database\")\n if \"dialect\" in kwargs or \"driver\" in kwargs:\n dialect = kwargs.pop(\...
userinfo == f"username:{quote('[password')}".encode("utf-8")
{ "list": [ { "filename": "src/tracking_utils.py", "retrieved_chunk": " orig_height = tf.cast(orig_height, dtype='float32')\n else:\n orig_height = tf.constant(orig_height, dtype='float32')\n if tf.is_tensor(orig_width):\n orig_width = tf.cast(orig_width, dtype='float32')\n ...
# coding=utf-8 # Copyright 2021 The Google Research Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicab...
if flow is not None: flow, mask = uflow_utils.resize( flow, new_height, new_width, is_flow=True, mask=mask) return images, flow, mask def random_scale_second( images, flow=None, mask=None, min_scale=1.0, max_scale=1.0): """Performs a random scaling on the second image in the given range.""" #...
{ "context_start_lineno": 0, "file": "src/uflow_augmentation.py", "groundtruth_start_lineno": 331, "repository": "JunbongJang-contour-tracking-1219b66", "right_context_start_lineno": 332, "task_id": "project_cc_python/2686" }
{ "list": [ { "filename": "src/tracking_utils.py", "retrieved_chunk": " new_width = tf.constant(new_width, dtype='float32')\n Ry = new_height / orig_height\n Rx = new_width / orig_width\n tracking_points = tf.cast(tracking_points, dtype='float32')\n x_points = tracking_points[:, :, ...
resize(images, new_height, new_width, is_flow=False)
{ "list": [ { "filename": "tests/test_connection_options.py", "retrieved_chunk": "def test_aiopg_pool_size():\n backend = AiopgBackend(\"postgresql+aiopg://localhost/database?min_size=1&max_size=20\")\n kwargs = backend._get_connection_kwargs()\n assert kwargs == {\"minsize\": 1, \"maxsize\":...
from urllib.parse import quote import pytest from databasez import DatabaseURL def test_database_url_repr(): u = DatabaseURL("postgresql://localhost/name") assert repr(u) == "DatabaseURL('postgresql://localhost/name')" u = DatabaseURL("postgresql://username@localhost/name") assert repr(u) == "Datab...
u = DatabaseURL("mysql+asyncmy://username/testsuite?unix_socket=/tmp/mysqld/mysqld.sock") assert u.options == {"unix_socket": "/tmp/mysqld/mysqld.sock"} def test_replace_database_url_components(): u = DatabaseURL("postgresql://localhost/mydatabase") assert u.database == "mydatabase" new = u.rep...
{ "context_start_lineno": 0, "file": "tests/test_database_url.py", "groundtruth_start_lineno": 69, "repository": "tarsil-databasez-01b23b0", "right_context_start_lineno": 70, "task_id": "project_cc_python/2664" }
{ "list": [ { "filename": "tests/test_connection_options.py", "retrieved_chunk": " backend = PostgresBackend(url + \"?min_size=1&max_size=20\")\n await backend.connect()\n await backend.disconnect()\ndef test_postgres_explicit_pool_size():\n backend = PostgresBackend(\"postgres...
options == {"pool_size": "20", "ssl": "true"}
{ "list": [ { "filename": "src/preprocessing/MATLAB_tracking_points.py", "retrieved_chunk": " cm = pylab.get_cmap('gist_rainbow')\n for a_folder in dataset_folders:\n print(a_folder)\n image_format = '.png'\n image_path_list = glob(f'{root_path}{a_folder}/{image_folder}/*{im...
''' Author: Junbong Jang Date: 2/7/2022 It loads PC, HACKS, and Jellyfish videos, GT labels, pseudo-labels from Mechanical model, predictions from various contour tracking algorithms. Then, it draws manuscript figures and evaluate models' performance by spatial and contour accuracy. ''' import os import cv2 from glo...
# a_image_name = a_image_name[-3:] assert a_image_name == get_image_name(img_path, image_format ) a_img = plt.imread(img_path) a_mask = plt.imread(mask_path).astype('uint8') # sample all points along the contour with order ordered_contour_points = get_ordered_contour_p...
{ "context_start_lineno": 0, "file": "src/preprocessing/main_process_tracking_points.py", "groundtruth_start_lineno": 139, "repository": "JunbongJang-contour-tracking-1219b66", "right_context_start_lineno": 140, "task_id": "project_cc_python/2687" }
{ "list": [ { "filename": "src/preprocessing/contour_tracking_manuscript_figures.py", "retrieved_chunk": " initial_frame = 0\n final_frame = 4\n total_frames = final_frame - initial_frame\n plt.rcParams[\"font.family\"] = \"Times New Roman\"\n a_image = plt.imread(img_path_list[initial_...
replace('refined_', '') # to make the name of mask the same as the name of image
{ "list": [ { "filename": "tests/test_connection_options.py", "retrieved_chunk": "from databasez.core import DatabaseURL\ndef test_postgres_pool_size():\n backend = PostgresBackend(\"postgres://localhost/database?min_size=1&max_size=20\")\n kwargs = backend._get_connection_kwargs()\n assert k...
from urllib.parse import quote import pytest from databasez import DatabaseURL def test_database_url_repr(): u = DatabaseURL("postgresql://localhost/name") assert repr(u) == "DatabaseURL('postgresql://localhost/name')" u = DatabaseURL("postgresql://username@localhost/name") assert repr(u) == "Datab...
assert u.driver == "asyncpg" assert u.username == "username" assert u.password == "password" assert u.hostname == "localhost" assert u.port == 123 assert u.database == "mydatabase" u = DatabaseURL( "postgresql://username:password@/mydatabase?host=/var/run/postgresql/.s.PGSQL.5432" ...
{ "context_start_lineno": 0, "file": "tests/test_database_url.py", "groundtruth_start_lineno": 23, "repository": "tarsil-databasez-01b23b0", "right_context_start_lineno": 24, "task_id": "project_cc_python/2656" }
{ "list": [ { "filename": "tests/test_connection_options.py", "retrieved_chunk": " backend = PostgresBackend(url + \"?min_size=1&max_size=20\")\n await backend.connect()\n await backend.disconnect()\ndef test_postgres_explicit_pool_size():\n backend = PostgresBackend(\"postgres...
dialect == "postgresql"
{ "list": [ { "filename": "databasez/backends/mysql.py", "retrieved_chunk": " return kwargs\n async def connect(self) -> None:\n assert self._pool is None, \"DatabaseBackend is already running\"\n kwargs = self._get_connection_kwargs()\n self._pool = await asyncmy.create...
import getpass import logging import typing import uuid import aioodbc import pyodbc as ext_pyodbc from sqlalchemy.dialects.mssql import pyodbc from sqlalchemy.engine.cursor import CursorResultMetaData from sqlalchemy.engine.interfaces import Dialect, ExecutionContext from sqlalchemy.sql import ClauseElement from sqla...
user = self._database_url.username or getpass.getuser() password = self._database_url.password timeout = kwargs.pop("timeout") if port: dsn = f"Driver={driver};Database={database};Server={hostname},{port};UID={user};PWD={password};Connection+Timeout={timeout}" else:...
{ "context_start_lineno": 0, "file": "databasez/backends/mssql.py", "groundtruth_start_lineno": 82, "repository": "tarsil-databasez-01b23b0", "right_context_start_lineno": 83, "task_id": "project_cc_python/2670" }
{ "list": [ { "filename": "databasez/backends/mysql.py", "retrieved_chunk": " autocommit=True,\n **kwargs,\n )\n async def disconnect(self) -> None:\n assert self._pool is not None, \"DatabaseBackend is not running\"\n self._pool.close()\n pool, sel...
port or 1433
{ "list": [ { "filename": "google/cloud/alloydb/connector/connector.py", "retrieved_chunk": "from google.auth import default\nfrom google.auth.credentials import with_scopes_if_required\nfrom google.cloud.alloydb.connector.client import AlloyDBClient\nfrom google.cloud.alloydb.connector.instance impor...
# Copyright 2023 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, s...
assert connector._client is None assert connector._credentials == credentials connector.close() def test_Connector_context_manager(credentials: FakeCredentials) -> None: """ Test to check whether the __init__ method of Connector properly sets defaults as context manager. """ with Conn...
{ "context_start_lineno": 0, "file": "tests/unit/test_connector.py", "groundtruth_start_lineno": 31, "repository": "GoogleCloudPlatform-alloydb-python-connector-a397ea7", "right_context_start_lineno": 32, "task_id": "project_cc_python/2695" }
{ "list": [ { "filename": "google/cloud/alloydb/connector/connector.py", "retrieved_chunk": " credentials (google.auth.credentials.Credentials):\n A credentials object created from the google-auth Python library.\n If not specified, Application Default Credentials are used...
_alloydb_api_endpoint == "https://alloydb.googleapis.com"
{ "list": [ { "filename": "tests/unit/test_client.py", "retrieved_chunk": " assert client_cert == \"This is the client cert\"\n assert cert_chain[0] == \"This is the intermediate cert\"\n assert cert_chain[1] == \"This is the root cert\"\n@pytest.mark.asyncio\nasync def test_AlloyDBClient_ini...
# Copyright 2023 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, s...
connector.close() def test_Connector_context_manager(credentials: FakeCredentials) -> None: """ Test to check whether the __init__ method of Connector properly sets defaults as context manager. """ with Connector(credentials) as connector: assert connector._quota_project is None ...
{ "context_start_lineno": 0, "file": "tests/unit/test_connector.py", "groundtruth_start_lineno": 33, "repository": "GoogleCloudPlatform-alloydb-python-connector-a397ea7", "right_context_start_lineno": 34, "task_id": "project_cc_python/2697" }
{ "list": [ { "filename": "tests/unit/test_client.py", "retrieved_chunk": " # verify base endpoint is set\n assert client._alloydb_api_endpoint == \"www.test-endpoint.com\"\n # verify proper headers are set\n assert client._client.headers[\"User-Agent\"] == f\"alloydb-python-connector/{ver...
_credentials == credentials
{ "list": [ { "filename": "google/cloud/alloydb/connector/connector.py", "retrieved_chunk": "from google.auth import default\nfrom google.auth.credentials import with_scopes_if_required\nfrom google.cloud.alloydb.connector.client import AlloyDBClient\nfrom google.cloud.alloydb.connector.instance impor...
# Copyright 2023 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, s...
assert connector._alloydb_api_endpoint == "https://alloydb.googleapis.com" assert connector._client is None assert connector._credentials == credentials connector.close() def test_Connector_context_manager(credentials: FakeCredentials) -> None: """ Test to check whether the __init__ method of...
{ "context_start_lineno": 0, "file": "tests/unit/test_connector.py", "groundtruth_start_lineno": 30, "repository": "GoogleCloudPlatform-alloydb-python-connector-a397ea7", "right_context_start_lineno": 31, "task_id": "project_cc_python/2694" }
{ "list": [ { "filename": "google/cloud/alloydb/connector/connector.py", "retrieved_chunk": " credentials (google.auth.credentials.Credentials):\n A credentials object created from the google-auth Python library.\n If not specified, Application Default Credentials are used...
_quota_project is None
{ "list": [ { "filename": "tests/unit/mocks.py", "retrieved_chunk": " \"\"\"Helper method to get all certs in pem string format.\"\"\"\n pem_root = self.root_cert.public_bytes(\n encoding=serialization.Encoding.PEM\n ).decode(\"UTF-8\")\n pem_intermediate = self....
# Copyright 2023 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, s...
data = { "pemCsr": csr_str, "certDuration": "3600s", } resp = await self._client.post( url, headers=headers, json=data, raise_for_status=True ) resp_dict = await resp.json() return (resp_dict["pemCertificate"], resp_dict["pemCertifi...
{ "context_start_lineno": 0, "file": "google/cloud/alloydb/connector/client.py", "groundtruth_start_lineno": 154, "repository": "GoogleCloudPlatform-alloydb-python-connector-a397ea7", "right_context_start_lineno": 155, "task_id": "project_cc_python/2693" }
{ "list": [ { "filename": "tests/unit/mocks.py", "retrieved_chunk": " \"\"\"Checks if the credentials are expired.\n Note that credentials can be invalid but not expired because\n Credentials with expiry set to None are considered to never\n expire.\n \"\"\"\n ...
public_bytes(encoding=serialization.Encoding.PEM).decode("utf-8")
{ "list": [ { "filename": "tests/test_connection_options.py", "retrieved_chunk": " backend = AsyncMyBackend(\"mysql+asyncmy://localhost/database?min_size=1&max_size=20\")\n kwargs = backend._get_connection_kwargs()\n assert kwargs == {\"minsize\": 1, \"maxsize\": 20}\n@pytest.mark.skipif(sys....
from urllib.parse import quote import pytest from databasez import DatabaseURL def test_database_url_repr(): u = DatabaseURL("postgresql://localhost/name") assert repr(u) == "DatabaseURL('postgresql://localhost/name')" u = DatabaseURL("postgresql://username@localhost/name") assert repr(u) == "Datab...
assert new.database == "test_mydatabase" assert str(new) == "postgresql://localhost/test_mydatabase" assert u.driver == "" new = u.replace(driver="asyncpg") assert new.driver == "asyncpg" assert str(new) == "postgresql+asyncpg://localhost/mydatabase" assert u.port is None new = u.repl...
{ "context_start_lineno": 0, "file": "tests/test_database_url.py", "groundtruth_start_lineno": 79, "repository": "tarsil-databasez-01b23b0", "right_context_start_lineno": 80, "task_id": "project_cc_python/2665" }
{ "list": [ { "filename": "tests/test_connection_options.py", "retrieved_chunk": "@pytest.mark.skipif(sys.version_info < (3, 7), reason=\"requires python3.7 or higher\")\ndef test_asyncmy_explicit_pool_size():\n backend = AsyncMyBackend(\"mysql://localhost/database\", min_size=1, max_size=20)\n ...
replace(database="test_" + u.database)
{ "list": [ { "filename": "tests/unittests/test_dimreduction.py", "retrieved_chunk": " try:\n self.assertTrue(\n (res.__round__(4) != res_labels.__round__(4)).nnz == 0,\n \"Results are different !\")\n except AssertionError as ...
import unittest from cnlp import other_methods from tests import Configs import numpy as np from scipy import sparse class TestDimensionalityReductionMethods(unittest.TestCase): def __perform_test(self, fun, params: dict = {}, debug: bool = False): g = Configs.load_normal_dataset() g_labels = Con...
def test_pathentropy(self): self.__perform_test(other_methods.information_theory.path_entropy) if __name__ == '__main__': unittest.main()
{ "context_start_lineno": 0, "file": "tests/unittests/test_othermethods.py", "groundtruth_start_lineno": 51, "repository": "Typing-Monkeys-complex-network-link-prediction-cf69190", "right_context_start_lineno": 52, "task_id": "project_cc_python/2692" }
{ "list": [ { "filename": "tests/unittests/test_probabilisticmethods.py", "retrieved_chunk": " self.__perform_test(probabilistic_methods.stochastic_block_model, {\n 'n': 1,\n })", "score": 90.53148653783545 }, { "filename": "tests/unittests/test_dimreductio...
information_theory.MI)
{ "list": [ { "filename": "DecAF/Knowledge/linearize.py", "retrieved_chunk": " return dict_name\ndef load_nameid_dict(file_dir, lower):\n print(\"Loading name2id and id2name dict...\")\n name2id_dict = defaultdict(list)\n id2name_dict = {}\n for file in tqdm(os.listdir(file_dir)):\n ...
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved. # SPDX-License-Identifier: CC-BY-NC-4.0 import os import jsonlines import csv from tqdm import tqdm from collections import defaultdict from multiprocessing import Pool from functools import partial import argparse from DecAF.Knowledge.linearize im...
continue if triple.obj.startswith("m") and triple.obj not in id2name_dict: continue subj = triple.subj if triple.subj not in id2name_dict: triple.subj = "" grouped_entity_triples[subj].append(con...
{ "context_start_lineno": 0, "file": "DecAF/Knowledge/process_freebase.py", "groundtruth_start_lineno": 94, "repository": "awslabs-decode-answer-logical-form-08d5790", "right_context_start_lineno": 95, "task_id": "project_cc_python/2701" }
{ "list": [ { "filename": "DecAF/Reading/process_fid.py", "retrieved_chunk": " new_data_sp = []\n for data_i in tqdm(data):\n if args.mode != \"QA\":\n if \"LF_processed\" in data_i:\n new_data_i = {\n \"id\": str(data_i[\"QuestionId\"]) + \":S...
should_ignore(id2name_dict):
{ "list": [ { "filename": "DecAF/Knowledge/linearize.py", "retrieved_chunk": " return dict_name\ndef load_nameid_dict(file_dir, lower):\n print(\"Loading name2id and id2name dict...\")\n name2id_dict = defaultdict(list)\n id2name_dict = {}\n for file in tqdm(os.listdir(file_dir)):\n ...
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved. # SPDX-License-Identifier: CC-BY-NC-4.0 import os import jsonlines import csv from tqdm import tqdm from collections import defaultdict from multiprocessing import Pool from functools import partial import argparse from DecAF.Knowledge.linearize im...
continue subj = triple.subj if triple.subj not in id2name_dict: triple.subj = "" grouped_entity_triples[subj].append(convert_relation_to_text(triple, id2name_dict)) with jsonlines.open(os.path.join(save_dir...
{ "context_start_lineno": 0, "file": "DecAF/Knowledge/process_freebase.py", "groundtruth_start_lineno": 96, "repository": "awslabs-decode-answer-logical-form-08d5790", "right_context_start_lineno": 97, "task_id": "project_cc_python/2702" }
{ "list": [ { "filename": "DecAF/Knowledge/linearize.py", "retrieved_chunk": " procesed_name = row[2].lower()\n else:\n procesed_name = row[2]\n name2id_dict[procesed_name].append(row[0])\n id2name_dict[row[0]] = proces...
obj.startswith("m") and triple.obj not in id2name_dict:
{ "list": [ { "filename": "e2e_tests/src/e2e_conduit/fixtures/importers/algod.py", "retrieved_chunk": " return \"algod\"\n @property\n def lastblock(self):\n if self.last is None:\n raise RuntimeError(\"algod importer has no blockfiles configured\")\n return self....
import glob import json import logging import os import boto3 from botocore.config import Config from botocore import UNSIGNED from e2e_common.util import hassuffix, xrun, firstFromS3Prefix, countblocks, atexitrun from e2e_conduit.fixtures.importers.importer_plugin import ImporterPlugin logger = logging.getLogger(__n...
with open(os.path.join(self.algoddir, "algod.net"), "r") as algod_net: self.config_input["netaddr"] = "http://" + algod_net.read().strip() with open(os.path.join(self.algoddir, "algod.token"), "r") as algod_token: self.config_input["token"] = algod_token.read().strip() def ...
{ "context_start_lineno": 0, "file": "e2e_tests/src/e2e_conduit/fixtures/importers/follower_algod.py", "groundtruth_start_lineno": 32, "repository": "algorand-conduit-bcb54fd", "right_context_start_lineno": 33, "task_id": "project_cc_python/2678" }
{ "list": [ { "filename": "e2e_tests/src/e2e_conduit/fixtures/importers/algod.py", "retrieved_chunk": " self.config_input[\"token\"] = algod_token.read().strip()\n def resolve_config_output(self):\n self.config_output[\"algod_token\"] = self.config_input[\"token\"]\n self.c...
config_input["mode"] = "follower"
{ "list": [ { "filename": "autoresearcher/workflows/literature_review/literature_review.py", "retrieved_chunk": " research_question (str): The research question to generate a literature review for.\n output_file (str, optional): The file path to save the literature review to.\n Returns:\n...
from autoresearcher.llms.openai import openai_call from autoresearcher.utils.prompts import keyword_combination_prompt # Generate keyword combinations for a given research question def generate_keyword_combinations(research_question): """ Generates keyword combinations for a given research question. Args:...
return [ combination.split(": ")[1] for combination in combinations if ": " in combination ]
{ "context_start_lineno": 0, "file": "autoresearcher/utils/generate_keyword_combinations.py", "groundtruth_start_lineno": 18, "repository": "eimenhmdt-autoresearcher-5235ee7", "right_context_start_lineno": 19, "task_id": "project_cc_python/2713" }
{ "list": [ { "filename": "autoresearcher/workflows/literature_review/literature_review.py", "retrieved_chunk": " Fetching top 20 papers...\n Top 20 papers fetched!\n Extracting research findings from papers...\n Research findings extracted!\n Synthesizing answers...\n Li...
split("\n")
{ "list": [ { "filename": "autoresearcher/workflows/literature_review/literature_review.py", "retrieved_chunk": " research_question (str): The research question to generate a literature review for.\n output_file (str, optional): The file path to save the literature review to.\n Returns:\n...
from autoresearcher.llms.openai import openai_call from autoresearcher.utils.prompts import keyword_combination_prompt # Generate keyword combinations for a given research question def generate_keyword_combinations(research_question): """ Generates keyword combinations for a given research question. Args:...
response = openai_call(prompt, use_gpt4=False, temperature=0, max_tokens=200) combinations = response.split("\n") return [ combination.split(": ")[1] for combination in combinations if ": " in combination ]
{ "context_start_lineno": 0, "file": "autoresearcher/utils/generate_keyword_combinations.py", "groundtruth_start_lineno": 16, "repository": "eimenhmdt-autoresearcher-5235ee7", "right_context_start_lineno": 17, "task_id": "project_cc_python/2712" }
{ "list": [ { "filename": "autoresearcher/workflows/literature_review/literature_review.py", "retrieved_chunk": " Fetching top 20 papers...\n Top 20 papers fetched!\n Extracting research findings from papers...\n Research findings extracted!\n Synthesizing answers...\n Li...
format(research_question=research_question)
{ "list": [ { "filename": "training_pipeline/modules/ViT.py", "retrieved_chunk": " label2id={c: str(i) for i, c in enumerate(LABELS)}\n model = TFViTForImageClassification.from_pretrained(\n PRETRAIN_CHECKPOINT,\n num_labels=len(LABELS),\n label2id=label2id,\n id2label=id2lab...
import tarfile import wandb import gradio as gr import numpy as np from PIL import Image import tensorflow as tf from transformers import ViTFeatureExtractor PRETRAIN_CHECKPOINT = "google/vit-base-patch16-224-in21k" feature_extractor = ViTFeatureExtractor.from_pretrained(PRETRAIN_CHECKPOINT) MODEL = None RESOLTUIO...
classify_if.click( get_predictions, [wb_token_if, image_if], label_if ) demo.launch(debug=True)
{ "context_start_lineno": 0, "file": "training_pipeline/huggingface/apps/gradio/app.py", "groundtruth_start_lineno": 74, "repository": "deep-diver-TFX-WandB-05c63c4", "right_context_start_lineno": 75, "task_id": "project_cc_python/2759" }
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " any data of the type of tf.Example which is denoted as the type of sta\n ndard_artifacts.Examples in TFX. The purpose of this function is to ap\n ply Transform Graph obtained from Transform component to ...
Button()
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " pred_source = tf.gather(params=labels, indices=indices)\n probs = tf.nn.softmax(predictions.logits, axis=1)\n pred_confidence = tf.reduce_max(probs, axis=1)\n return {\"label\": pred_s...
import tarfile import wandb import gradio as gr import numpy as np from PIL import Image import tensorflow as tf from transformers import ViTFeatureExtractor PRETRAIN_CHECKPOINT = "google/vit-base-patch16-224-in21k" feature_extractor = ViTFeatureExtractor.from_pretrained(PRETRAIN_CHECKPOINT) MODEL = None RESOLTUIO...
classify_if = gr.Button() classify_if.click( get_predictions, [wb_token_if, image_if], label_if ) demo.launch(debug=True)
{ "context_start_lineno": 0, "file": "training_pipeline/huggingface/apps/gradio/app.py", "groundtruth_start_lineno": 72, "repository": "deep-diver-TFX-WandB-05c63c4", "right_context_start_lineno": 73, "task_id": "project_cc_python/2758" }
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " any data of the type of tf.Example which is denoted as the type of sta\n ndard_artifacts.Examples in TFX. The purpose of this function is to ap\n ply Transform Graph obtained from Transform component to ...
Label(num_top_classes=3)
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " pred_source = tf.gather(params=labels, indices=indices)\n probs = tf.nn.softmax(predictions.logits, axis=1)\n pred_confidence = tf.reduce_max(probs, axis=1)\n return {\"label\": pred_s...
import tarfile import wandb import gradio as gr import numpy as np from PIL import Image import tensorflow as tf from transformers import ViTFeatureExtractor PRETRAIN_CHECKPOINT = "google/vit-base-patch16-224-in21k" feature_extractor = ViTFeatureExtractor.from_pretrained(PRETRAIN_CHECKPOINT) MODEL = None RESOLTUIO...
with gr.Row(): image_if = gr.Image() label_if = gr.Label(num_top_classes=3) classify_if = gr.Button() classify_if.click( get_predictions, [wb_token_if, image_if], label_if ) demo.launch(debug=True)
{ "context_start_lineno": 0, "file": "training_pipeline/huggingface/apps/gradio/app.py", "groundtruth_start_lineno": 68, "repository": "deep-diver-TFX-WandB-05c63c4", "right_context_start_lineno": 69, "task_id": "project_cc_python/2755" }
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " any data of the type of tf.Example which is denoted as the type of sta\n ndard_artifacts.Examples in TFX. The purpose of this function is to ap\n ply Transform Graph obtained from Transform component to ...
Textbox(interactive=True, label="Your Weight & Biases API Key")
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " pred_source = tf.gather(params=labels, indices=indices)\n probs = tf.nn.softmax(predictions.logits, axis=1)\n pred_confidence = tf.reduce_max(probs, axis=1)\n return {\"label\": pred_s...
import tarfile import wandb import gradio as gr import numpy as np from PIL import Image import tensorflow as tf from transformers import ViTFeatureExtractor PRETRAIN_CHECKPOINT = "google/vit-base-patch16-224-in21k" feature_extractor = ViTFeatureExtractor.from_pretrained(PRETRAIN_CHECKPOINT) MODEL = None RESOLTUIO...
image_if = gr.Image() label_if = gr.Label(num_top_classes=3) classify_if = gr.Button() classify_if.click( get_predictions, [wb_token_if, image_if], label_if ) demo.launch(debug=True)
{ "context_start_lineno": 0, "file": "training_pipeline/huggingface/apps/gradio/app.py", "groundtruth_start_lineno": 70, "repository": "deep-diver-TFX-WandB-05c63c4", "right_context_start_lineno": 71, "task_id": "project_cc_python/2756" }
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " any data of the type of tf.Example which is denoted as the type of sta\n ndard_artifacts.Examples in TFX. The purpose of this function is to ap\n ply Transform Graph obtained from Transform component to ...
Row():
{ "list": [ { "filename": "story_development/test_outline.py", "retrieved_chunk": "- However, when Juan encounters a large butterfly, he becomes scared and runs away, leaving the butterfly in danger. The Butterfly Keeper tells Juan that having a fear of butterflies is okay, but he must learn to empath...
import openai import os from dotenv import load_dotenv from story_development.characters import Characters load_dotenv() openai.api_key = os.environ["OPENAI_API_KEY"] conditioning_info = "target audience is a 5 year old boy.The story should help the reader overcome a fear of butterflies." premise = '"The Butterfly ...
{ "context_start_lineno": 0, "file": "story_development/test_characters.py", "groundtruth_start_lineno": 52, "repository": "knexer-ai-storyteller-86da1d3", "right_context_start_lineno": 53, "task_id": "project_cc_python/2749" }
{ "list": [ { "filename": "story_development/test_feedback.py", "retrieved_chunk": "- However, when Juan encounters a large butterfly, he becomes scared and runs away, leaving the butterfly in danger. The Butterfly Keeper tells Juan that having a fear of butterflies is okay, but he must learn to empat...
make_recommendation(verbose=True)
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " pred_source = tf.gather(params=labels, indices=indices)\n probs = tf.nn.softmax(predictions.logits, axis=1)\n pred_confidence = tf.reduce_max(probs, axis=1)\n return {\"label\": pred_s...
import tarfile import wandb import gradio as gr import numpy as np from PIL import Image import tensorflow as tf from transformers import ViTFeatureExtractor PRETRAIN_CHECKPOINT = "google/vit-base-patch16-224-in21k" feature_extractor = ViTFeatureExtractor.from_pretrained(PRETRAIN_CHECKPOINT) MODEL = None RESOLTUIO...
gr.Markdown("## Simple demo for a Image Classification of the Beans Dataset with HF ViT model") wb_token_if = gr.Textbox(interactive=True, label="Your Weight & Biases API Key") with gr.Row(): image_if = gr.Image() label_if = gr.Label(num_top_classes=3) classify_if = gr.Button() ...
{ "context_start_lineno": 0, "file": "training_pipeline/huggingface/apps/gradio/app.py", "groundtruth_start_lineno": 65, "repository": "deep-diver-TFX-WandB-05c63c4", "right_context_start_lineno": 66, "task_id": "project_cc_python/2753" }
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " any data of the type of tf.Example which is denoted as the type of sta\n ndard_artifacts.Examples in TFX. The purpose of this function is to ap\n ply Transform Graph obtained from Transform component to ...
Blocks() as demo:
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " pred_source = tf.gather(params=labels, indices=indices)\n probs = tf.nn.softmax(predictions.logits, axis=1)\n pred_confidence = tf.reduce_max(probs, axis=1)\n return {\"label\": pred_s...
import tarfile import wandb import gradio as gr import numpy as np from PIL import Image import tensorflow as tf from transformers import ViTFeatureExtractor PRETRAIN_CHECKPOINT = "google/vit-base-patch16-224-in21k" feature_extractor = ViTFeatureExtractor.from_pretrained(PRETRAIN_CHECKPOINT) MODEL = None RESOLTUIO...
label_if = gr.Label(num_top_classes=3) classify_if = gr.Button() classify_if.click( get_predictions, [wb_token_if, image_if], label_if ) demo.launch(debug=True)
{ "context_start_lineno": 0, "file": "training_pipeline/huggingface/apps/gradio/app.py", "groundtruth_start_lineno": 71, "repository": "deep-diver-TFX-WandB-05c63c4", "right_context_start_lineno": 72, "task_id": "project_cc_python/2757" }
{ "list": [ { "filename": "training_pipeline/modules/signatures.py", "retrieved_chunk": " any data of the type of tf.Example which is denoted as the type of sta\n ndard_artifacts.Examples in TFX. The purpose of this function is to ap\n ply Transform Graph obtained from Transform component to ...
Image()