repo stringlengths 7 55 | path stringlengths 4 223 | url stringlengths 87 315 | code stringlengths 75 104k | code_tokens list | docstring stringlengths 1 46.9k | docstring_tokens list | language stringclasses 1
value | partition stringclasses 3
values | avg_line_len float64 7.91 980 |
|---|---|---|---|---|---|---|---|---|---|
SKA-ScienceDataProcessor/integration-prototype | sip/examples/flask_processing_controller/app/api/processing_block.py | https://github.com/SKA-ScienceDataProcessor/integration-prototype/blob/8c8006de6ad71dcd44114b0338780738079c87d4/sip/examples/flask_processing_controller/app/api/processing_block.py#L45-L61 | def delete(block_id):
"""Processing block detail resource."""
_url = get_root_url()
try:
DB.delete_processing_block(block_id)
response = dict(message='Deleted block',
id='{}'.format(block_id),
links=dict(list='{}/processing-blocks'.format(_url)... | [
"def",
"delete",
"(",
"block_id",
")",
":",
"_url",
"=",
"get_root_url",
"(",
")",
"try",
":",
"DB",
".",
"delete_processing_block",
"(",
"block_id",
")",
"response",
"=",
"dict",
"(",
"message",
"=",
"'Deleted block'",
",",
"id",
"=",
"'{}'",
".",
"form... | Processing block detail resource. | [
"Processing",
"block",
"detail",
"resource",
"."
] | python | train | 45.294118 |
SpockBotMC/SpockBot | spockbot/plugins/helpers/inventory.py | https://github.com/SpockBotMC/SpockBot/blob/f89911551f18357720034fbaa52837a0d09f66ea/spockbot/plugins/helpers/inventory.py#L24-L35 | def total_stored(self, wanted, slots=None):
"""
Calculates the total number of items of that type
in the current window or given slot range.
Args:
wanted: function(Slot) or Slot or itemID or (itemID, metadata)
"""
if slots is None:
slots = self.wi... | [
"def",
"total_stored",
"(",
"self",
",",
"wanted",
",",
"slots",
"=",
"None",
")",
":",
"if",
"slots",
"is",
"None",
":",
"slots",
"=",
"self",
".",
"window",
".",
"slots",
"wanted",
"=",
"make_slot_check",
"(",
"wanted",
")",
"return",
"sum",
"(",
"... | Calculates the total number of items of that type
in the current window or given slot range.
Args:
wanted: function(Slot) or Slot or itemID or (itemID, metadata) | [
"Calculates",
"the",
"total",
"number",
"of",
"items",
"of",
"that",
"type",
"in",
"the",
"current",
"window",
"or",
"given",
"slot",
"range",
"."
] | python | train | 35.5 |
DLR-RM/RAFCON | source/rafcon/gui/controllers/utils/tree_view_controller.py | https://github.com/DLR-RM/RAFCON/blob/24942ef1a904531f49ab8830a1dbb604441be498/source/rafcon/gui/controllers/utils/tree_view_controller.py#L614-L623 | def select_entry(self, core_element_id, by_cursor=True):
"""Selects the row entry belonging to the given core_element_id by cursor or tree selection"""
for row_num, element_row in enumerate(self.list_store):
# Compare data port ids
if element_row[self.ID_STORAGE_ID] == core_eleme... | [
"def",
"select_entry",
"(",
"self",
",",
"core_element_id",
",",
"by_cursor",
"=",
"True",
")",
":",
"for",
"row_num",
",",
"element_row",
"in",
"enumerate",
"(",
"self",
".",
"list_store",
")",
":",
"# Compare data port ids",
"if",
"element_row",
"[",
"self",... | Selects the row entry belonging to the given core_element_id by cursor or tree selection | [
"Selects",
"the",
"row",
"entry",
"belonging",
"to",
"the",
"given",
"core_element_id",
"by",
"cursor",
"or",
"tree",
"selection"
] | python | train | 52.2 |
mitsei/dlkit | dlkit/handcar/learning/sessions.py | https://github.com/mitsei/dlkit/blob/445f968a175d61c8d92c0f617a3c17dc1dc7c584/dlkit/handcar/learning/sessions.py#L2576-L2599 | def get_activities_for_objective(self, objective_id=None):
"""Gets the activities for the given objective.
In plenary mode, the returned list contains all of the
activities mapped to the objective Id or an error results if an
Id in the supplied list is not found or inaccessible. Otherwis... | [
"def",
"get_activities_for_objective",
"(",
"self",
",",
"objective_id",
"=",
"None",
")",
":",
"if",
"objective_id",
"is",
"None",
":",
"raise",
"NullArgument",
"(",
")",
"# Should also check if objective_id exists?",
"url_path",
"=",
"construct_url",
"(",
"'activiti... | Gets the activities for the given objective.
In plenary mode, the returned list contains all of the
activities mapped to the objective Id or an error results if an
Id in the supplied list is not found or inaccessible. Otherwise,
inaccessible Activities may be omitted from the list and ma... | [
"Gets",
"the",
"activities",
"for",
"the",
"given",
"objective",
".",
"In",
"plenary",
"mode",
"the",
"returned",
"list",
"contains",
"all",
"of",
"the",
"activities",
"mapped",
"to",
"the",
"objective",
"Id",
"or",
"an",
"error",
"results",
"if",
"an",
"I... | python | train | 51.083333 |
StevenMaude/bbc-radio-tracklisting-downloader | bbc_tracklist.py | https://github.com/StevenMaude/bbc-radio-tracklisting-downloader/blob/9fe9096b4d889888f65756444e4fd71352b92458/bbc_tracklist.py#L72-L79 | def get_broadcast_date(pid):
"""Take BBC pid (string); extract and return broadcast date as string."""
print("Extracting first broadcast date...")
broadcast_etree = open_listing_page(pid + '/broadcasts.inc')
original_broadcast_date, = broadcast_etree.xpath(
'(//div[@class="grid__inner"]//div'
... | [
"def",
"get_broadcast_date",
"(",
"pid",
")",
":",
"print",
"(",
"\"Extracting first broadcast date...\"",
")",
"broadcast_etree",
"=",
"open_listing_page",
"(",
"pid",
"+",
"'/broadcasts.inc'",
")",
"original_broadcast_date",
",",
"=",
"broadcast_etree",
".",
"xpath",
... | Take BBC pid (string); extract and return broadcast date as string. | [
"Take",
"BBC",
"pid",
"(",
"string",
")",
";",
"extract",
"and",
"return",
"broadcast",
"date",
"as",
"string",
"."
] | python | train | 50.625 |
SoCo/SoCo | soco/core.py | https://github.com/SoCo/SoCo/blob/671937e07d7973b78c0cbee153d4f3ad68ec48c6/soco/core.py#L629-L643 | def seek(self, timestamp):
"""Seek to a given timestamp in the current track, specified in the
format of HH:MM:SS or H:MM:SS.
Raises:
ValueError: if the given timestamp is invalid.
"""
if not re.match(r'^[0-9][0-9]?:[0-9][0-9]:[0-9][0-9]$', timestamp):
ra... | [
"def",
"seek",
"(",
"self",
",",
"timestamp",
")",
":",
"if",
"not",
"re",
".",
"match",
"(",
"r'^[0-9][0-9]?:[0-9][0-9]:[0-9][0-9]$'",
",",
"timestamp",
")",
":",
"raise",
"ValueError",
"(",
"'invalid timestamp, use HH:MM:SS format'",
")",
"self",
".",
"avTranspo... | Seek to a given timestamp in the current track, specified in the
format of HH:MM:SS or H:MM:SS.
Raises:
ValueError: if the given timestamp is invalid. | [
"Seek",
"to",
"a",
"given",
"timestamp",
"in",
"the",
"current",
"track",
"specified",
"in",
"the",
"format",
"of",
"HH",
":",
"MM",
":",
"SS",
"or",
"H",
":",
"MM",
":",
"SS",
"."
] | python | train | 33.666667 |
wdecoster/nanoget | nanoget/extraction_functions.py | https://github.com/wdecoster/nanoget/blob/fb7306220e261849b96785fab02dd2f35a0e3b60/nanoget/extraction_functions.py#L309-L322 | def stream_fastq_full(fastq, threads):
"""Generator for returning metrics extracted from fastq.
Extract from a fastq file:
-readname
-average and median quality
-read_lenght
"""
logging.info("Nanoget: Starting to collect full metrics from plain fastq file.")
inputfastq = handle_compress... | [
"def",
"stream_fastq_full",
"(",
"fastq",
",",
"threads",
")",
":",
"logging",
".",
"info",
"(",
"\"Nanoget: Starting to collect full metrics from plain fastq file.\"",
")",
"inputfastq",
"=",
"handle_compressed_input",
"(",
"fastq",
")",
"with",
"cfutures",
".",
"Proce... | Generator for returning metrics extracted from fastq.
Extract from a fastq file:
-readname
-average and median quality
-read_lenght | [
"Generator",
"for",
"returning",
"metrics",
"extracted",
"from",
"fastq",
"."
] | python | train | 42.714286 |
squdle/baseconvert | baseconvert/baseconvert.py | https://github.com/squdle/baseconvert/blob/26c9a2c07c2ffcde7d078fb812419ca6d388900b/baseconvert/baseconvert.py#L571-L592 | def expand_recurring(number, repeat=5):
"""
Expands a recurring pattern within a number.
Args:
number(tuple): the number to process in the form:
(int, int, int, ... ".", ... , int int int)
repeat: the number of times to expand the pattern.
Returns:
The ori... | [
"def",
"expand_recurring",
"(",
"number",
",",
"repeat",
"=",
"5",
")",
":",
"if",
"\"[\"",
"in",
"number",
":",
"pattern_index",
"=",
"number",
".",
"index",
"(",
"\"[\"",
")",
"pattern",
"=",
"number",
"[",
"pattern_index",
"+",
"1",
":",
"-",
"1",
... | Expands a recurring pattern within a number.
Args:
number(tuple): the number to process in the form:
(int, int, int, ... ".", ... , int int int)
repeat: the number of times to expand the pattern.
Returns:
The original number with recurring pattern expanded.
E... | [
"Expands",
"a",
"recurring",
"pattern",
"within",
"a",
"number",
".",
"Args",
":",
"number",
"(",
"tuple",
")",
":",
"the",
"number",
"to",
"process",
"in",
"the",
"form",
":",
"(",
"int",
"int",
"int",
"...",
".",
"...",
"int",
"int",
"int",
")",
... | python | train | 31.454545 |
DataONEorg/d1_python | gmn/src/d1_gmn/app/did.py | https://github.com/DataONEorg/d1_python/blob/3ac4d4f3ca052d3e8641a6a329cab526c8ddcb0d/gmn/src/d1_gmn/app/did.py#L207-L212 | def is_revision_chain_placeholder(pid):
"""For replicas, the PIDs referenced in revision chains are reserved for use by
other replicas."""
return d1_gmn.app.models.ReplicaRevisionChainReference.objects.filter(
pid__did=pid
).exists() | [
"def",
"is_revision_chain_placeholder",
"(",
"pid",
")",
":",
"return",
"d1_gmn",
".",
"app",
".",
"models",
".",
"ReplicaRevisionChainReference",
".",
"objects",
".",
"filter",
"(",
"pid__did",
"=",
"pid",
")",
".",
"exists",
"(",
")"
] | For replicas, the PIDs referenced in revision chains are reserved for use by
other replicas. | [
"For",
"replicas",
"the",
"PIDs",
"referenced",
"in",
"revision",
"chains",
"are",
"reserved",
"for",
"use",
"by",
"other",
"replicas",
"."
] | python | train | 42 |
HewlettPackard/python-hpOneView | hpOneView/resources/networking/interconnects.py | https://github.com/HewlettPackard/python-hpOneView/blob/3c6219723ef25e6e0c83d44a89007f89bc325b89/hpOneView/resources/networking/interconnects.py#L203-L220 | def update_ports(self, ports, id_or_uri, timeout=-1):
"""
Updates the interconnect ports.
Args:
id_or_uri: Can be either the interconnect id or the interconnect uri.
ports (list): Ports to update.
timeout: Timeout in seconds. Wait for task completion by defau... | [
"def",
"update_ports",
"(",
"self",
",",
"ports",
",",
"id_or_uri",
",",
"timeout",
"=",
"-",
"1",
")",
":",
"resources",
"=",
"merge_default_values",
"(",
"ports",
",",
"{",
"'type'",
":",
"'port'",
"}",
")",
"uri",
"=",
"self",
".",
"_client",
".",
... | Updates the interconnect ports.
Args:
id_or_uri: Can be either the interconnect id or the interconnect uri.
ports (list): Ports to update.
timeout: Timeout in seconds. Wait for task completion by default. The timeout does not abort the operation
in OneView; i... | [
"Updates",
"the",
"interconnect",
"ports",
"."
] | python | train | 37.611111 |
silverlogic/djangorestframework-expander | expander/parse_qs.py | https://github.com/silverlogic/djangorestframework-expander/blob/b1cf60c7076169cbd6ad65350841c86080564f97/expander/parse_qs.py#L29-L41 | def qs_from_dict(qsdict, prefix=""):
''' Same as dict_from_qs, but in reverse
i.e. {"period": {"di": {}, "fhr": {}}} => "period.di,period.fhr"
'''
prefix = prefix + '.' if prefix else ""
def descend(qsd):
for key, val in sorted(qsd.items()):
if val:
yield qs_... | [
"def",
"qs_from_dict",
"(",
"qsdict",
",",
"prefix",
"=",
"\"\"",
")",
":",
"prefix",
"=",
"prefix",
"+",
"'.'",
"if",
"prefix",
"else",
"\"\"",
"def",
"descend",
"(",
"qsd",
")",
":",
"for",
"key",
",",
"val",
"in",
"sorted",
"(",
"qsd",
".",
"ite... | Same as dict_from_qs, but in reverse
i.e. {"period": {"di": {}, "fhr": {}}} => "period.di,period.fhr" | [
"Same",
"as",
"dict_from_qs",
"but",
"in",
"reverse",
"i",
".",
"e",
".",
"{",
"period",
":",
"{",
"di",
":",
"{}",
"fhr",
":",
"{}",
"}}",
"=",
">",
"period",
".",
"di",
"period",
".",
"fhr"
] | python | train | 32.769231 |
CamDavidsonPilon/lifetimes | lifetimes/plotting.py | https://github.com/CamDavidsonPilon/lifetimes/blob/f926308bc03c17c1d12fead729de43885cf13321/lifetimes/plotting.py#L570-L618 | def plot_transaction_rate_heterogeneity(
model,
suptitle="Heterogeneity in Transaction Rate",
xlabel="Transaction Rate",
ylabel="Density",
suptitle_fontsize=14,
**kwargs
):
"""
Plot the estimated gamma distribution of lambda (customers' propensities to purchase).
Parameters
----... | [
"def",
"plot_transaction_rate_heterogeneity",
"(",
"model",
",",
"suptitle",
"=",
"\"Heterogeneity in Transaction Rate\"",
",",
"xlabel",
"=",
"\"Transaction Rate\"",
",",
"ylabel",
"=",
"\"Density\"",
",",
"suptitle_fontsize",
"=",
"14",
",",
"*",
"*",
"kwargs",
")",... | Plot the estimated gamma distribution of lambda (customers' propensities to purchase).
Parameters
----------
model: lifetimes model
A fitted lifetimes model, for now only for BG/NBD
suptitle: str, optional
Figure suptitle
xlabel: str, optional
Figure xlabel
ylabel: str, ... | [
"Plot",
"the",
"estimated",
"gamma",
"distribution",
"of",
"lambda",
"(",
"customers",
"propensities",
"to",
"purchase",
")",
"."
] | python | train | 25.469388 |
RockFeng0/rtsf-web | webuidriver/actions.py | https://github.com/RockFeng0/rtsf-web/blob/ceabcf62ddf1c969a97b5c7a4a4c547198b6ea71/webuidriver/actions.py#L637-L643 | def Ctrl(cls, key):
""" 在指定元素上执行ctrl组合键事件
@note: key event -> control + key
@param key: 如'X'
"""
element = cls._element()
element.send_keys(Keys.CONTROL, key) | [
"def",
"Ctrl",
"(",
"cls",
",",
"key",
")",
":",
"element",
"=",
"cls",
".",
"_element",
"(",
")",
"element",
".",
"send_keys",
"(",
"Keys",
".",
"CONTROL",
",",
"key",
")"
] | 在指定元素上执行ctrl组合键事件
@note: key event -> control + key
@param key: 如'X' | [
"在指定元素上执行ctrl组合键事件"
] | python | train | 31.142857 |
bloomreach/s4cmd | s4cmd.py | https://github.com/bloomreach/s4cmd/blob/bb51075bf43703e7cd95aa39288cf7732ec13a6d/s4cmd.py#L680-L688 | def connect(self):
'''Connect to S3 storage'''
try:
if S3Handler.S3_KEYS:
self.s3 = BotoClient(self.opt, S3Handler.S3_KEYS[0], S3Handler.S3_KEYS[1])
else:
self.s3 = BotoClient(self.opt)
except Exception as e:
raise RetryFailure('Unable to connect to s3: %s' % e) | [
"def",
"connect",
"(",
"self",
")",
":",
"try",
":",
"if",
"S3Handler",
".",
"S3_KEYS",
":",
"self",
".",
"s3",
"=",
"BotoClient",
"(",
"self",
".",
"opt",
",",
"S3Handler",
".",
"S3_KEYS",
"[",
"0",
"]",
",",
"S3Handler",
".",
"S3_KEYS",
"[",
"1",... | Connect to S3 storage | [
"Connect",
"to",
"S3",
"storage"
] | python | test | 33.333333 |
materialsproject/pymatgen | pymatgen/phonon/dos.py | https://github.com/materialsproject/pymatgen/blob/4ca558cf72f8d5f8a1f21dfdfc0181a971c186da/pymatgen/phonon/dos.py#L216-L250 | def internal_energy(self, t, structure=None):
"""
Phonon contribution to the internal energy at temperature T obtained from the integration of the DOS.
Only positive frequencies will be used.
Result in J/mol-c. A mol-c is the abbreviation of a mole-cell, that is, the number
of Av... | [
"def",
"internal_energy",
"(",
"self",
",",
"t",
",",
"structure",
"=",
"None",
")",
":",
"if",
"t",
"==",
"0",
":",
"return",
"self",
".",
"zero_point_energy",
"(",
"structure",
"=",
"structure",
")",
"freqs",
"=",
"self",
".",
"_positive_frequencies",
... | Phonon contribution to the internal energy at temperature T obtained from the integration of the DOS.
Only positive frequencies will be used.
Result in J/mol-c. A mol-c is the abbreviation of a mole-cell, that is, the number
of Avogadro times the atoms in a unit cell. To compare with experimenta... | [
"Phonon",
"contribution",
"to",
"the",
"internal",
"energy",
"at",
"temperature",
"T",
"obtained",
"from",
"the",
"integration",
"of",
"the",
"DOS",
".",
"Only",
"positive",
"frequencies",
"will",
"be",
"used",
".",
"Result",
"in",
"J",
"/",
"mol",
"-",
"c... | python | train | 38.742857 |
saltstack/salt | salt/utils/kickstart.py | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/utils/kickstart.py#L560-L601 | def parse_raid(rule):
'''
Parse the raid line
'''
parser = argparse.ArgumentParser()
rules = shlex.split(rule)
rules.pop(0)
partitions = []
newrules = []
for count in range(0, len(rules)):
if count == 0:
newrules.append(rules[count])
continue
... | [
"def",
"parse_raid",
"(",
"rule",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
")",
"rules",
"=",
"shlex",
".",
"split",
"(",
"rule",
")",
"rules",
".",
"pop",
"(",
"0",
")",
"partitions",
"=",
"[",
"]",
"newrules",
"=",
"[",
"]... | Parse the raid line | [
"Parse",
"the",
"raid",
"line"
] | python | train | 37.285714 |
DomainTools/python_api | domaintools/api.py | https://github.com/DomainTools/python_api/blob/17be85fd4913fbe14d7660a4f4829242f1663e60/domaintools/api.py#L163-L168 | def reverse_whois(self, query, exclude=[], scope='current', mode=None, **kwargs):
"""List of one or more terms to search for in the Whois record,
as a Python list or separated with the pipe character ( | ).
"""
return self._results('reverse-whois', '/v1/reverse-whois', terms=delimited... | [
"def",
"reverse_whois",
"(",
"self",
",",
"query",
",",
"exclude",
"=",
"[",
"]",
",",
"scope",
"=",
"'current'",
",",
"mode",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"self",
".",
"_results",
"(",
"'reverse-whois'",
",",
"'/v1/reverse... | List of one or more terms to search for in the Whois record,
as a Python list or separated with the pipe character ( | ). | [
"List",
"of",
"one",
"or",
"more",
"terms",
"to",
"search",
"for",
"in",
"the",
"Whois",
"record",
"as",
"a",
"Python",
"list",
"or",
"separated",
"with",
"the",
"pipe",
"character",
"(",
"|",
")",
"."
] | python | train | 68.833333 |
codelv/enaml-native | src/enamlnative/core/dev.py | https://github.com/codelv/enaml-native/blob/c33986e9eda468c508806e0a3e73c771401e5718/src/enamlnative/core/dev.py#L491-L514 | def render_files(self, root=None):
""" Render the file path as accordions
"""
if root is None:
tmp = os.environ.get('TMP')
root = sys.path[1 if tmp and tmp in sys.path else 0]
items = []
for filename in os.listdir(root):
# for subdirname in di... | [
"def",
"render_files",
"(",
"self",
",",
"root",
"=",
"None",
")",
":",
"if",
"root",
"is",
"None",
":",
"tmp",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'TMP'",
")",
"root",
"=",
"sys",
".",
"path",
"[",
"1",
"if",
"tmp",
"and",
"tmp",
"in"... | Render the file path as accordions | [
"Render",
"the",
"file",
"path",
"as",
"accordions"
] | python | train | 35.375 |
dnanexus/dx-toolkit | src/python/dxpy/api.py | https://github.com/dnanexus/dx-toolkit/blob/74befb53ad90fcf902d8983ae6d74580f402d619/src/python/dxpy/api.py#L915-L921 | def org_update(object_id, input_params={}, always_retry=True, **kwargs):
"""
Invokes the /org-xxxx/update API method.
For more info, see: https://wiki.dnanexus.com/API-Specification-v1.0.0/Organizations#API-method%3A-%2Forg-xxxx%2Fupdate
"""
return DXHTTPRequest('/%s/update' % object_id, input_para... | [
"def",
"org_update",
"(",
"object_id",
",",
"input_params",
"=",
"{",
"}",
",",
"always_retry",
"=",
"True",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"DXHTTPRequest",
"(",
"'/%s/update'",
"%",
"object_id",
",",
"input_params",
",",
"always_retry",
"=",
... | Invokes the /org-xxxx/update API method.
For more info, see: https://wiki.dnanexus.com/API-Specification-v1.0.0/Organizations#API-method%3A-%2Forg-xxxx%2Fupdate | [
"Invokes",
"the",
"/",
"org",
"-",
"xxxx",
"/",
"update",
"API",
"method",
"."
] | python | train | 50.571429 |
ThreatResponse/margaritashotgun | margaritashotgun/remote_shell.py | https://github.com/ThreatResponse/margaritashotgun/blob/6dee53ef267959b214953439968244cc46a19690/margaritashotgun/remote_shell.py#L200-L218 | def upload_file(self, local_path, remote_path):
"""
Upload a file from the local filesystem to the remote host
:type local_path: str
:param local_path: path of local file to upload
:type remote_path: str
:param remote_path: destination path of upload on remote host
... | [
"def",
"upload_file",
"(",
"self",
",",
"local_path",
",",
"remote_path",
")",
":",
"logger",
".",
"debug",
"(",
"\"{0}: uploading {1} to {0}:{2}\"",
".",
"format",
"(",
"self",
".",
"target_address",
",",
"local_path",
",",
"remote_path",
")",
")",
"try",
":"... | Upload a file from the local filesystem to the remote host
:type local_path: str
:param local_path: path of local file to upload
:type remote_path: str
:param remote_path: destination path of upload on remote host | [
"Upload",
"a",
"file",
"from",
"the",
"local",
"filesystem",
"to",
"the",
"remote",
"host"
] | python | train | 45.526316 |
ronaldguillen/wave | wave/compat.py | https://github.com/ronaldguillen/wave/blob/20bb979c917f7634d8257992e6d449dc751256a9/wave/compat.py#L251-L261 | def get_all_related_many_to_many_objects(opts):
"""
Django 1.8 changed meta api, see docstr in compat.get_all_related_objects()
:param opts: Options instance
:return: list of many-to-many relations
"""
if django.VERSION < (1, 9):
return opts.get_all_related_many_to_many_objects()
el... | [
"def",
"get_all_related_many_to_many_objects",
"(",
"opts",
")",
":",
"if",
"django",
".",
"VERSION",
"<",
"(",
"1",
",",
"9",
")",
":",
"return",
"opts",
".",
"get_all_related_many_to_many_objects",
"(",
")",
"else",
":",
"return",
"[",
"r",
"for",
"r",
"... | Django 1.8 changed meta api, see docstr in compat.get_all_related_objects()
:param opts: Options instance
:return: list of many-to-many relations | [
"Django",
"1",
".",
"8",
"changed",
"meta",
"api",
"see",
"docstr",
"in",
"compat",
".",
"get_all_related_objects",
"()"
] | python | train | 35.090909 |
mfitzp/padua | padua/process.py | https://github.com/mfitzp/padua/blob/8b14bf4d2f895da6aea5d7885d409315bd303ec6/padua/process.py#L330-L377 | def fold_columns_to_rows(df, levels_from=2):
"""
Take a levels from the columns and fold down into the row index.
This destroys the existing index; existing rows will appear as
columns under the new column index
:param df:
:param levels_from: The level (inclusive) from which column index will b... | [
"def",
"fold_columns_to_rows",
"(",
"df",
",",
"levels_from",
"=",
"2",
")",
":",
"df",
"=",
"df",
".",
"copy",
"(",
")",
"df",
".",
"reset_index",
"(",
"inplace",
"=",
"True",
",",
"drop",
"=",
"True",
")",
"# Wipe out the current index",
"df",
"=",
"... | Take a levels from the columns and fold down into the row index.
This destroys the existing index; existing rows will appear as
columns under the new column index
:param df:
:param levels_from: The level (inclusive) from which column index will be folded
:return: | [
"Take",
"a",
"levels",
"from",
"the",
"columns",
"and",
"fold",
"down",
"into",
"the",
"row",
"index",
".",
"This",
"destroys",
"the",
"existing",
"index",
";",
"existing",
"rows",
"will",
"appear",
"as",
"columns",
"under",
"the",
"new",
"column",
"index"... | python | train | 25.5625 |
markperdue/pyvesync | src/pyvesync/vesync.py | https://github.com/markperdue/pyvesync/blob/7552dd1a6dd5ebc452acf78e33fd8f6e721e8cfc/src/pyvesync/vesync.py#L186-L189 | def update_energy(self, bypass_check=False):
"""Fetch updated energy information about devices"""
for outlet in self.outlets:
outlet.update_energy(bypass_check) | [
"def",
"update_energy",
"(",
"self",
",",
"bypass_check",
"=",
"False",
")",
":",
"for",
"outlet",
"in",
"self",
".",
"outlets",
":",
"outlet",
".",
"update_energy",
"(",
"bypass_check",
")"
] | Fetch updated energy information about devices | [
"Fetch",
"updated",
"energy",
"information",
"about",
"devices"
] | python | train | 46.25 |
phoebe-project/phoebe2 | phoebe/parameters/dataset.py | https://github.com/phoebe-project/phoebe2/blob/e64b8be683977064e2d55dd1b3ac400f64c3e379/phoebe/parameters/dataset.py#L345-L367 | def mesh(**kwargs):
"""
Create parameters for a new mesh dataset.
Generally, this will be used as an input to the kind argument in
:meth:`phoebe.frontend.bundle.Bundle.add_dataset`
:parameter **kwargs: defaults for the values of any of the parameters
:return: a :class:`phoebe.parameters.parame... | [
"def",
"mesh",
"(",
"*",
"*",
"kwargs",
")",
":",
"obs_params",
"=",
"[",
"]",
"syn_params",
",",
"constraints",
"=",
"mesh_syn",
"(",
"syn",
"=",
"False",
",",
"*",
"*",
"kwargs",
")",
"obs_params",
"+=",
"syn_params",
".",
"to_list",
"(",
")",
"obs... | Create parameters for a new mesh dataset.
Generally, this will be used as an input to the kind argument in
:meth:`phoebe.frontend.bundle.Bundle.add_dataset`
:parameter **kwargs: defaults for the values of any of the parameters
:return: a :class:`phoebe.parameters.parameters.ParameterSet` of all newly
... | [
"Create",
"parameters",
"for",
"a",
"new",
"mesh",
"dataset",
"."
] | python | train | 42.782609 |
frejanordsiek/hdf5storage | hdf5storage/__init__.py | https://github.com/frejanordsiek/hdf5storage/blob/539275141dd3a4efbbbfd9bdb978f3ed59e3f05d/hdf5storage/__init__.py#L1727-L1836 | def reads(paths, filename='data.h5', options=None, **keywords):
""" Reads data from an HDF5 file (high level).
High level function to read one or more pieces of data from an HDF5
file located at the paths specified in `paths` into Python
types. Each path is specified as a POSIX style path where the dat... | [
"def",
"reads",
"(",
"paths",
",",
"filename",
"=",
"'data.h5'",
",",
"options",
"=",
"None",
",",
"*",
"*",
"keywords",
")",
":",
"# Pack the different options into an Options class if an Options was",
"# not given. By default, the matlab_compatible option is set to",
"# Fal... | Reads data from an HDF5 file (high level).
High level function to read one or more pieces of data from an HDF5
file located at the paths specified in `paths` into Python
types. Each path is specified as a POSIX style path where the data
to read is located.
There are various options that can be use... | [
"Reads",
"data",
"from",
"an",
"HDF5",
"file",
"(",
"high",
"level",
")",
"."
] | python | train | 37.436364 |
krukas/Trionyx | trionyx/navigation.py | https://github.com/krukas/Trionyx/blob/edac132cc0797190153f2e60bc7e88cb50e80da6/trionyx/navigation.py#L238-L266 | def register(self, model_alias, code='general', name=None, order=None, display_filter=None):
"""
Register new tab
:param model_alias:
:param code:
:param name:
:param order:
:return:
"""
model_alias = self.get_model_alias(model_alias)
def... | [
"def",
"register",
"(",
"self",
",",
"model_alias",
",",
"code",
"=",
"'general'",
",",
"name",
"=",
"None",
",",
"order",
"=",
"None",
",",
"display_filter",
"=",
"None",
")",
":",
"model_alias",
"=",
"self",
".",
"get_model_alias",
"(",
"model_alias",
... | Register new tab
:param model_alias:
:param code:
:param name:
:param order:
:return: | [
"Register",
"new",
"tab"
] | python | train | 31.241379 |
pmacosta/peng | docs/support/requirements_to_rst.py | https://github.com/pmacosta/peng/blob/976935377adaa3de26fc5677aceb2cdfbd6f93a7/docs/support/requirements_to_rst.py#L77-L102 | def ops_to_words(item):
"""Translate requirement specification to words."""
unsupp_ops = ["~=", "==="]
# Ordered for "pleasant" word specification
supp_ops = [">=", ">", "==", "<=", "<", "!="]
tokens = sorted(item.split(","), reverse=True)
actual_tokens = []
for req in tokens:
for o... | [
"def",
"ops_to_words",
"(",
"item",
")",
":",
"unsupp_ops",
"=",
"[",
"\"~=\"",
",",
"\"===\"",
"]",
"# Ordered for \"pleasant\" word specification",
"supp_ops",
"=",
"[",
"\">=\"",
",",
"\">\"",
",",
"\"==\"",
",",
"\"<=\"",
",",
"\"<\"",
",",
"\"!=\"",
"]",... | Translate requirement specification to words. | [
"Translate",
"requirement",
"specification",
"to",
"words",
"."
] | python | test | 40 |
synw/goerr | goerr/messages.py | https://github.com/synw/goerr/blob/08b3809d6715bffe26899a769d96fa5de8573faf/goerr/messages.py#L18-L25 | def error(self, i: int=None) -> str:
"""
Returns an error message
"""
head = "[" + colors.red("error") + "]"
if i is not None:
head = str(i) + " " + head
return head | [
"def",
"error",
"(",
"self",
",",
"i",
":",
"int",
"=",
"None",
")",
"->",
"str",
":",
"head",
"=",
"\"[\"",
"+",
"colors",
".",
"red",
"(",
"\"error\"",
")",
"+",
"\"]\"",
"if",
"i",
"is",
"not",
"None",
":",
"head",
"=",
"str",
"(",
"i",
")... | Returns an error message | [
"Returns",
"an",
"error",
"message"
] | python | train | 27.25 |
archman/beamline | beamline/models.py | https://github.com/archman/beamline/blob/417bc5dc13e754bc89d246427984590fced64d07/beamline/models.py#L303-L354 | def anoteElements(ax, anotelist, showAccName=False, efilter=None, textypos=None, **kwargs):
""" annotate elements to axes
:param ax: matplotlib axes object
:param anotelist: element annotation object list
:param showAccName: tag name for accelerator tubes? defaul... | [
"def",
"anoteElements",
"(",
"ax",
",",
"anotelist",
",",
"showAccName",
"=",
"False",
",",
"efilter",
"=",
"None",
",",
"textypos",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"defaultstyle",
"=",
"{",
"'alpha'",
":",
"0.8",
",",
"'arrowprops'",
"... | annotate elements to axes
:param ax: matplotlib axes object
:param anotelist: element annotation object list
:param showAccName: tag name for accelerator tubes? default is False,
show acceleration band type, e.g. 'S', 'C', 'X', or for '[S,C,X]D' for cavi... | [
"annotate",
"elements",
"to",
"axes",
":",
"param",
"ax",
":",
"matplotlib",
"axes",
"object",
":",
"param",
"anotelist",
":",
"element",
"annotation",
"object",
"list",
":",
"param",
"showAccName",
":",
"tag",
"name",
"for",
"accelerator",
"tubes?",
"default"... | python | train | 54.346154 |
callowayproject/Transmogrify | transmogrify/images2gif.py | https://github.com/callowayproject/Transmogrify/blob/f1f891b8b923b3a1ede5eac7f60531c1c472379e/transmogrify/images2gif.py#L977-L991 | def palette_image(self):
""" PIL weird interface for making a paletted image: create an image which
already has the palette, and use that in Image.quantize. This function
returns this palette image. """
if self.pimage is None:
palette = []
for i in range(s... | [
"def",
"palette_image",
"(",
"self",
")",
":",
"if",
"self",
".",
"pimage",
"is",
"None",
":",
"palette",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"self",
".",
"NETSIZE",
")",
":",
"palette",
".",
"extend",
"(",
"self",
".",
"colormap",
"[",
... | PIL weird interface for making a paletted image: create an image which
already has the palette, and use that in Image.quantize. This function
returns this palette image. | [
"PIL",
"weird",
"interface",
"for",
"making",
"a",
"paletted",
"image",
":",
"create",
"an",
"image",
"which",
"already",
"has",
"the",
"palette",
"and",
"use",
"that",
"in",
"Image",
".",
"quantize",
".",
"This",
"function",
"returns",
"this",
"palette",
... | python | train | 40.2 |
asyncee/django-easy-select2 | docs/source/_ext/djangodocs.py | https://github.com/asyncee/django-easy-select2/blob/f81bbaa91d0266029be7ef6d075d85f13273e3a5/docs/source/_ext/djangodocs.py#L364-L403 | def visit_console_html(self, node):
"""Generate HTML for the console directive."""
if self.builder.name in ('djangohtml', 'json') and node['win_console_text']:
# Put a mark on the document object signaling the fact the directive
# has been used on it.
self.document._console_directive_use... | [
"def",
"visit_console_html",
"(",
"self",
",",
"node",
")",
":",
"if",
"self",
".",
"builder",
".",
"name",
"in",
"(",
"'djangohtml'",
",",
"'json'",
")",
"and",
"node",
"[",
"'win_console_text'",
"]",
":",
"# Put a mark on the document object signaling the fact t... | Generate HTML for the console directive. | [
"Generate",
"HTML",
"for",
"the",
"console",
"directive",
"."
] | python | train | 43.4 |
saltstack/salt | salt/modules/genesis.py | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/genesis.py#L264-L278 | def _mkfs(root, fs_format, fs_opts=None):
'''
Make a filesystem using the appropriate module
.. versionadded:: Beryllium
'''
if fs_opts is None:
fs_opts = {}
if fs_format in ('ext2', 'ext3', 'ext4'):
__salt__['extfs.mkfs'](root, fs_format, **fs_opts)
elif fs_format in ('btr... | [
"def",
"_mkfs",
"(",
"root",
",",
"fs_format",
",",
"fs_opts",
"=",
"None",
")",
":",
"if",
"fs_opts",
"is",
"None",
":",
"fs_opts",
"=",
"{",
"}",
"if",
"fs_format",
"in",
"(",
"'ext2'",
",",
"'ext3'",
",",
"'ext4'",
")",
":",
"__salt__",
"[",
"'e... | Make a filesystem using the appropriate module
.. versionadded:: Beryllium | [
"Make",
"a",
"filesystem",
"using",
"the",
"appropriate",
"module"
] | python | train | 29.2 |
Erotemic/utool | utool/util_numpy.py | https://github.com/Erotemic/utool/blob/3b27e1f4e6e6fb23cd8744af7b7195b57d99e03a/utool/util_numpy.py#L142-L175 | def random_indexes(max_index, subset_size=None, seed=None, rng=None):
""" random unrepeated indicies
Args:
max_index (?):
subset_size (None): (default = None)
seed (None): (default = None)
rng (RandomState): random number generator(default = None)
Returns:
?: subst... | [
"def",
"random_indexes",
"(",
"max_index",
",",
"subset_size",
"=",
"None",
",",
"seed",
"=",
"None",
",",
"rng",
"=",
"None",
")",
":",
"subst_",
"=",
"np",
".",
"arange",
"(",
"0",
",",
"max_index",
")",
"rng",
"=",
"ensure_rng",
"(",
"seed",
"if",... | random unrepeated indicies
Args:
max_index (?):
subset_size (None): (default = None)
seed (None): (default = None)
rng (RandomState): random number generator(default = None)
Returns:
?: subst
CommandLine:
python -m utool.util_numpy --exec-random_indexes
... | [
"random",
"unrepeated",
"indicies"
] | python | train | 28.882353 |
arogozhnikov/einops | einops/einops.py | https://github.com/arogozhnikov/einops/blob/9698f0f5efa6c5a79daa75253137ba5d79a95615/einops/einops.py#L191-L254 | def parse_expression(expression: str) -> Tuple[Set[str], List[CompositeAxis]]:
"""
Parses an indexing expression (for a single tensor).
Checks uniqueness of names, checks usage of '...' (allowed only once)
Returns set of all used identifiers and a list of axis groups
"""
identifiers = set()
... | [
"def",
"parse_expression",
"(",
"expression",
":",
"str",
")",
"->",
"Tuple",
"[",
"Set",
"[",
"str",
"]",
",",
"List",
"[",
"CompositeAxis",
"]",
"]",
":",
"identifiers",
"=",
"set",
"(",
")",
"composite_axes",
"=",
"[",
"]",
"if",
"'.'",
"in",
"exp... | Parses an indexing expression (for a single tensor).
Checks uniqueness of names, checks usage of '...' (allowed only once)
Returns set of all used identifiers and a list of axis groups | [
"Parses",
"an",
"indexing",
"expression",
"(",
"for",
"a",
"single",
"tensor",
")",
".",
"Checks",
"uniqueness",
"of",
"names",
"checks",
"usage",
"of",
"...",
"(",
"allowed",
"only",
"once",
")",
"Returns",
"set",
"of",
"all",
"used",
"identifiers",
"and"... | python | train | 43.296875 |
andreikop/qutepart | qutepart/completer.py | https://github.com/andreikop/qutepart/blob/109d76b239751318bcef06f39b2fbbf18687a40b/qutepart/completer.py#L408-L433 | def invokeCompletionIfAvailable(self, requestedByUser=False):
"""Invoke completion, if available. Called after text has been typed in qpart
Returns True, if invoked
"""
if self._qpart.completionEnabled and self._wordSet is not None:
wordBeforeCursor = self._wordBeforeCursor()... | [
"def",
"invokeCompletionIfAvailable",
"(",
"self",
",",
"requestedByUser",
"=",
"False",
")",
":",
"if",
"self",
".",
"_qpart",
".",
"completionEnabled",
"and",
"self",
".",
"_wordSet",
"is",
"not",
"None",
":",
"wordBeforeCursor",
"=",
"self",
".",
"_wordBefo... | Invoke completion, if available. Called after text has been typed in qpart
Returns True, if invoked | [
"Invoke",
"completion",
"if",
"available",
".",
"Called",
"after",
"text",
"has",
"been",
"typed",
"in",
"qpart",
"Returns",
"True",
"if",
"invoked"
] | python | train | 47.769231 |
edx/edx-django-utils | edx_django_utils/cache/utils.py | https://github.com/edx/edx-django-utils/blob/16cb4ac617e53c572bf68ccd19d24afeff1ca769/edx_django_utils/cache/utils.py#L240-L250 | def _set_request_cache_if_django_cache_hit(key, django_cached_response):
"""
Sets the value in the request cache if the django cached response was a hit.
Args:
key (string)
django_cached_response (CachedResponse)
"""
if django_cached_response.is_found:
... | [
"def",
"_set_request_cache_if_django_cache_hit",
"(",
"key",
",",
"django_cached_response",
")",
":",
"if",
"django_cached_response",
".",
"is_found",
":",
"DEFAULT_REQUEST_CACHE",
".",
"set",
"(",
"key",
",",
"django_cached_response",
".",
"value",
")"
] | Sets the value in the request cache if the django cached response was a hit.
Args:
key (string)
django_cached_response (CachedResponse) | [
"Sets",
"the",
"value",
"in",
"the",
"request",
"cache",
"if",
"the",
"django",
"cached",
"response",
"was",
"a",
"hit",
"."
] | python | train | 34.636364 |
pantsbuild/pants | src/python/pants/backend/jvm/tasks/jvm_dependency_usage.py | https://github.com/pantsbuild/pants/blob/b72e650da0df685824ffdcc71988b8c282d0962d/src/python/pants/backend/jvm/tasks/jvm_dependency_usage.py#L422-L443 | def to_json(self):
"""Outputs the entire graph."""
res_dict = {}
def gen_dep_edge(node, edge, dep_tgt, aliases):
return {
'target': dep_tgt.address.spec,
'dependency_type': self._edge_type(node.concrete_target, edge, dep_tgt),
'products_used': len(edge.products_used),
... | [
"def",
"to_json",
"(",
"self",
")",
":",
"res_dict",
"=",
"{",
"}",
"def",
"gen_dep_edge",
"(",
"node",
",",
"edge",
",",
"dep_tgt",
",",
"aliases",
")",
":",
"return",
"{",
"'target'",
":",
"dep_tgt",
".",
"address",
".",
"spec",
",",
"'dependency_typ... | Outputs the entire graph. | [
"Outputs",
"the",
"entire",
"graph",
"."
] | python | train | 41.636364 |
Kortemme-Lab/klab | klab/db/mysql.py | https://github.com/Kortemme-Lab/klab/blob/6d410ad08f1bd9f7cbbb28d7d946e94fbaaa2b6b/klab/db/mysql.py#L327-L347 | def run_transaction(self, command_list, do_commit=True):
'''This can be used to stage multiple commands and roll back the transaction if an error occurs. This is useful
if you want to remove multiple records in multiple tables for one entity but do not want the deletion to occur
if the e... | [
"def",
"run_transaction",
"(",
"self",
",",
"command_list",
",",
"do_commit",
"=",
"True",
")",
":",
"pass",
"# I decided against creating this for now.",
"# It may be more useful to create a stored procedure like in e.g. _create_protein_deletion_stored_procedure",
"# in the DDGadmin p... | This can be used to stage multiple commands and roll back the transaction if an error occurs. This is useful
if you want to remove multiple records in multiple tables for one entity but do not want the deletion to occur
if the entity is tied to table not specified in the list of commands. Perfor... | [
"This",
"can",
"be",
"used",
"to",
"stage",
"multiple",
"commands",
"and",
"roll",
"back",
"the",
"transaction",
"if",
"an",
"error",
"occurs",
".",
"This",
"is",
"useful",
"if",
"you",
"want",
"to",
"remove",
"multiple",
"records",
"in",
"multiple",
"tabl... | python | train | 59.285714 |
koordinates/python-client | koordinates/base.py | https://github.com/koordinates/python-client/blob/f3dc7cd164f5a9499b2454cd1d4516e9d4b3c252/koordinates/base.py#L492-L505 | def _serialize_value(self, value):
"""
Called by :py:meth:`._serialize` to serialise an individual value.
"""
if isinstance(value, (list, tuple, set)):
return [self._serialize_value(v) for v in value]
elif isinstance(value, dict):
return dict([(k, self._se... | [
"def",
"_serialize_value",
"(",
"self",
",",
"value",
")",
":",
"if",
"isinstance",
"(",
"value",
",",
"(",
"list",
",",
"tuple",
",",
"set",
")",
")",
":",
"return",
"[",
"self",
".",
"_serialize_value",
"(",
"v",
")",
"for",
"v",
"in",
"value",
"... | Called by :py:meth:`._serialize` to serialise an individual value. | [
"Called",
"by",
":",
"py",
":",
"meth",
":",
".",
"_serialize",
"to",
"serialise",
"an",
"individual",
"value",
"."
] | python | train | 41.857143 |
AndrewAnnex/SpiceyPy | spiceypy/spiceypy.py | https://github.com/AndrewAnnex/SpiceyPy/blob/fc20a9b9de68b58eed5b332f0c051fb343a6e335/spiceypy/spiceypy.py#L7711-L7732 | def latcyl(radius, lon, lat):
"""
Convert from latitudinal coordinates to cylindrical coordinates.
http://naif.jpl.nasa.gov/pub/naif/toolkit_docs/C/cspice/latcyl_c.html
:param radius: Distance of a point from the origin.
:type radius:
:param lon: Angle of the point from the XZ plane in radians... | [
"def",
"latcyl",
"(",
"radius",
",",
"lon",
",",
"lat",
")",
":",
"radius",
"=",
"ctypes",
".",
"c_double",
"(",
"radius",
")",
"lon",
"=",
"ctypes",
".",
"c_double",
"(",
"lon",
")",
"lat",
"=",
"ctypes",
".",
"c_double",
"(",
"lat",
")",
"r",
"... | Convert from latitudinal coordinates to cylindrical coordinates.
http://naif.jpl.nasa.gov/pub/naif/toolkit_docs/C/cspice/latcyl_c.html
:param radius: Distance of a point from the origin.
:type radius:
:param lon: Angle of the point from the XZ plane in radians.
:param lat: Angle of the point from ... | [
"Convert",
"from",
"latitudinal",
"coordinates",
"to",
"cylindrical",
"coordinates",
"."
] | python | train | 34.227273 |
mabuchilab/QNET | src/qnet/algebra/_rules.py | https://github.com/mabuchilab/QNET/blob/cc20d26dad78691d34c67173e5cd67dcac94208a/src/qnet/algebra/_rules.py#L965-L971 | def _pull_out_perm_rhs(rest, rhs, out_port, in_port):
"""Similar to :func:`_pull_out_perm_lhs` but on the RHS of a series
product self-feedback."""
in_im, rhs_red = rhs._factor_rhs(in_port)
return (Feedback.create(
SeriesProduct.create(*rest),
out_port=out_port, in_port=i... | [
"def",
"_pull_out_perm_rhs",
"(",
"rest",
",",
"rhs",
",",
"out_port",
",",
"in_port",
")",
":",
"in_im",
",",
"rhs_red",
"=",
"rhs",
".",
"_factor_rhs",
"(",
"in_port",
")",
"return",
"(",
"Feedback",
".",
"create",
"(",
"SeriesProduct",
".",
"create",
... | Similar to :func:`_pull_out_perm_lhs` but on the RHS of a series
product self-feedback. | [
"Similar",
"to",
":",
"func",
":",
"_pull_out_perm_lhs",
"but",
"on",
"the",
"RHS",
"of",
"a",
"series",
"product",
"self",
"-",
"feedback",
"."
] | python | train | 47.285714 |
PyCQA/astroid | astroid/node_classes.py | https://github.com/PyCQA/astroid/blob/e0a298df55b15abcb77c2a93253f5ab7be52d0fb/astroid/node_classes.py#L4558-L4572 | def postinit(self, value, conversion=None, format_spec=None):
"""Do some setup after initialisation.
:param value: The value to be formatted into the string.
:type value: NodeNG
:param conversion: The type of formatting to be applied to the value.
:type conversion: int or None
... | [
"def",
"postinit",
"(",
"self",
",",
"value",
",",
"conversion",
"=",
"None",
",",
"format_spec",
"=",
"None",
")",
":",
"self",
".",
"value",
"=",
"value",
"self",
".",
"conversion",
"=",
"conversion",
"self",
".",
"format_spec",
"=",
"format_spec"
] | Do some setup after initialisation.
:param value: The value to be formatted into the string.
:type value: NodeNG
:param conversion: The type of formatting to be applied to the value.
:type conversion: int or None
:param format_spec: The formatting to be applied to the value.
... | [
"Do",
"some",
"setup",
"after",
"initialisation",
"."
] | python | train | 35.8 |
assamite/creamas | creamas/examples/spiro/spiro.py | https://github.com/assamite/creamas/blob/54dc3e31c97a3f938e58272f8ab80b6bcafeff58/creamas/examples/spiro/spiro.py#L39-L53 | def spiro_image(R, r, r_, resolution=2*PI/1000, spins=50, size=[32, 32]):
'''Create image with given Spirograph parameters using numpy and scipy.
'''
x, y = give_dots(200, r, r_, spins=20)
xy = np.array([x, y]).T
xy = np.array(np.around(xy), dtype=np.int64)
xy = xy[(xy[:, 0] >= -250) & (xy[:, 1]... | [
"def",
"spiro_image",
"(",
"R",
",",
"r",
",",
"r_",
",",
"resolution",
"=",
"2",
"*",
"PI",
"/",
"1000",
",",
"spins",
"=",
"50",
",",
"size",
"=",
"[",
"32",
",",
"32",
"]",
")",
":",
"x",
",",
"y",
"=",
"give_dots",
"(",
"200",
",",
"r",... | Create image with given Spirograph parameters using numpy and scipy. | [
"Create",
"image",
"with",
"given",
"Spirograph",
"parameters",
"using",
"numpy",
"and",
"scipy",
"."
] | python | train | 36.866667 |
scopely-devops/skew | skew/arn/__init__.py | https://github.com/scopely-devops/skew/blob/e90d5e2220b2284502a06430bb94b4aba9ea60db/skew/arn/__init__.py#L53-L71 | def match(self, pattern, context=None):
"""
This method returns a (possibly empty) list of strings that
match the regular expression ``pattern`` provided. You can
also provide a ``context`` as described above.
This method calls ``choices`` to get a list of all possible
... | [
"def",
"match",
"(",
"self",
",",
"pattern",
",",
"context",
"=",
"None",
")",
":",
"matches",
"=",
"[",
"]",
"regex",
"=",
"pattern",
"if",
"regex",
"==",
"'*'",
":",
"regex",
"=",
"'.*'",
"regex",
"=",
"re",
".",
"compile",
"(",
"regex",
")",
"... | This method returns a (possibly empty) list of strings that
match the regular expression ``pattern`` provided. You can
also provide a ``context`` as described above.
This method calls ``choices`` to get a list of all possible
choices and then filters the list by performing a regular
... | [
"This",
"method",
"returns",
"a",
"(",
"possibly",
"empty",
")",
"list",
"of",
"strings",
"that",
"match",
"the",
"regular",
"expression",
"pattern",
"provided",
".",
"You",
"can",
"also",
"provide",
"a",
"context",
"as",
"described",
"above",
"."
] | python | train | 37.736842 |
developersociety/django-glitter | glitter/reminders/admin.py | https://github.com/developersociety/django-glitter/blob/2c0280ec83afee80deee94ee3934fc54239c2e87/glitter/reminders/admin.py#L15-L20 | def get_formset(self, request, obj=None, **kwargs):
""" Default user to the current version owner. """
data = super().get_formset(request, obj, **kwargs)
if obj:
data.form.base_fields['user'].initial = request.user.id
return data | [
"def",
"get_formset",
"(",
"self",
",",
"request",
",",
"obj",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"data",
"=",
"super",
"(",
")",
".",
"get_formset",
"(",
"request",
",",
"obj",
",",
"*",
"*",
"kwargs",
")",
"if",
"obj",
":",
"data",... | Default user to the current version owner. | [
"Default",
"user",
"to",
"the",
"current",
"version",
"owner",
"."
] | python | train | 44.666667 |
mpenning/polymer | polymer/Polymer.py | https://github.com/mpenning/polymer/blob/1cdf4ed2573c894bde9d398fa173816b6b47e9f3/polymer/Polymer.py#L73-L161 | def r_q_send(self, msg_dict):
"""Send message dicts through r_q, and throw explicit errors for
pickle problems"""
# Check whether msg_dict can be pickled...
no_pickle_keys = self.invalid_dict_pickle_keys(msg_dict)
if no_pickle_keys == []:
self.r_q.put(msg_dict)
... | [
"def",
"r_q_send",
"(",
"self",
",",
"msg_dict",
")",
":",
"# Check whether msg_dict can be pickled...",
"no_pickle_keys",
"=",
"self",
".",
"invalid_dict_pickle_keys",
"(",
"msg_dict",
")",
"if",
"no_pickle_keys",
"==",
"[",
"]",
":",
"self",
".",
"r_q",
".",
"... | Send message dicts through r_q, and throw explicit errors for
pickle problems | [
"Send",
"message",
"dicts",
"through",
"r_q",
"and",
"throw",
"explicit",
"errors",
"for",
"pickle",
"problems"
] | python | test | 39.988764 |
srittau/python-asserts | asserts/__init__.py | https://github.com/srittau/python-asserts/blob/1d5c797031c68ee27552d1c94e7f918c3d3d0453/asserts/__init__.py#L177-L194 | def assert_not_equal(first, second, msg_fmt="{msg}"):
"""Fail if first equals second, as determined by the '==' operator.
>>> assert_not_equal(5, 8)
>>> assert_not_equal(-7, -7.0)
Traceback (most recent call last):
...
AssertionError: -7 == -7.0
The following msg_fmt arguments are supp... | [
"def",
"assert_not_equal",
"(",
"first",
",",
"second",
",",
"msg_fmt",
"=",
"\"{msg}\"",
")",
":",
"if",
"first",
"==",
"second",
":",
"msg",
"=",
"\"{!r} == {!r}\"",
".",
"format",
"(",
"first",
",",
"second",
")",
"fail",
"(",
"msg_fmt",
".",
"format"... | Fail if first equals second, as determined by the '==' operator.
>>> assert_not_equal(5, 8)
>>> assert_not_equal(-7, -7.0)
Traceback (most recent call last):
...
AssertionError: -7 == -7.0
The following msg_fmt arguments are supported:
* msg - the default error message
* first - th... | [
"Fail",
"if",
"first",
"equals",
"second",
"as",
"determined",
"by",
"the",
"==",
"operator",
"."
] | python | train | 31.388889 |
mfcloud/python-zvm-sdk | zvmsdk/smtclient.py | https://github.com/mfcloud/python-zvm-sdk/blob/de9994ceca764f5460ce51bd74237986341d8e3c/zvmsdk/smtclient.py#L374-L384 | def guest_stop(self, userid, **kwargs):
"""Power off VM."""
requestData = "PowerVM " + userid + " off"
if 'timeout' in kwargs.keys() and kwargs['timeout']:
requestData += ' --maxwait ' + str(kwargs['timeout'])
if 'poll_interval' in kwargs.keys() and kwargs['poll_interval']:
... | [
"def",
"guest_stop",
"(",
"self",
",",
"userid",
",",
"*",
"*",
"kwargs",
")",
":",
"requestData",
"=",
"\"PowerVM \"",
"+",
"userid",
"+",
"\" off\"",
"if",
"'timeout'",
"in",
"kwargs",
".",
"keys",
"(",
")",
"and",
"kwargs",
"[",
"'timeout'",
"]",
":... | Power off VM. | [
"Power",
"off",
"VM",
"."
] | python | train | 43.454545 |
coinkite/connectrum | connectrum/svr_info.py | https://github.com/coinkite/connectrum/blob/99948f92cc5c3ecb1a8a70146294014e608e50fc/connectrum/svr_info.py#L180-L201 | def from_irc(self, irc_nickname=None, irc_password=None):
'''
Connect to the IRC channel and find all servers presently connected.
Slow; takes 30+ seconds but authoritative and current.
OBSOLETE.
'''
if have_bottom:
from .findall import IrcListen... | [
"def",
"from_irc",
"(",
"self",
",",
"irc_nickname",
"=",
"None",
",",
"irc_password",
"=",
"None",
")",
":",
"if",
"have_bottom",
":",
"from",
".",
"findall",
"import",
"IrcListener",
"# connect and fetch current set of servers who are",
"# on #electrum channel at free... | Connect to the IRC channel and find all servers presently connected.
Slow; takes 30+ seconds but authoritative and current.
OBSOLETE. | [
"Connect",
"to",
"the",
"IRC",
"channel",
"and",
"find",
"all",
"servers",
"presently",
"connected",
"."
] | python | train | 31.954545 |
django-fluent/django-fluent-contents | fluent_contents/rendering/media.py | https://github.com/django-fluent/django-fluent-contents/blob/896f14add58471b98d7aa295b2c9e6abedec9003/fluent_contents/rendering/media.py#L6-L14 | def register_frontend_media(request, media):
"""
Add a :class:`~django.forms.Media` class to the current request.
This will be rendered by the ``render_plugin_media`` template tag.
"""
if not hasattr(request, '_fluent_contents_frontend_media'):
request._fluent_contents_frontend_media = Media... | [
"def",
"register_frontend_media",
"(",
"request",
",",
"media",
")",
":",
"if",
"not",
"hasattr",
"(",
"request",
",",
"'_fluent_contents_frontend_media'",
")",
":",
"request",
".",
"_fluent_contents_frontend_media",
"=",
"Media",
"(",
")",
"add_media",
"(",
"requ... | Add a :class:`~django.forms.Media` class to the current request.
This will be rendered by the ``render_plugin_media`` template tag. | [
"Add",
"a",
":",
"class",
":",
"~django",
".",
"forms",
".",
"Media",
"class",
"to",
"the",
"current",
"request",
".",
"This",
"will",
"be",
"rendered",
"by",
"the",
"render_plugin_media",
"template",
"tag",
"."
] | python | train | 41.888889 |
jobovy/galpy | galpy/df/streamdf.py | https://github.com/jobovy/galpy/blob/9c5b9fe65d58835624dffe432be282060918ee08/galpy/df/streamdf.py#L2606-L2639 | def _parse_call_args(self,*args,**kwargs):
"""Helper function to parse the arguments to the __call__ and related functions,
return [6,nobj] array of frequencies (:3) and angles (3:)"""
interp= kwargs.get('interp',self._useInterp)
if len(args) == 5:
raise IOError("Must specify... | [
"def",
"_parse_call_args",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"interp",
"=",
"kwargs",
".",
"get",
"(",
"'interp'",
",",
"self",
".",
"_useInterp",
")",
"if",
"len",
"(",
"args",
")",
"==",
"5",
":",
"raise",
"IOError... | Helper function to parse the arguments to the __call__ and related functions,
return [6,nobj] array of frequencies (:3) and angles (3:) | [
"Helper",
"function",
"to",
"parse",
"the",
"arguments",
"to",
"the",
"__call__",
"and",
"related",
"functions",
"return",
"[",
"6",
"nobj",
"]",
"array",
"of",
"frequencies",
"(",
":",
"3",
")",
"and",
"angles",
"(",
"3",
":",
")"
] | python | train | 45.852941 |
horazont/aioxmpp | aioxmpp/pubsub/service.py | https://github.com/horazont/aioxmpp/blob/22a68e5e1d23f2a4dee470092adbd4672f9ef061/aioxmpp/pubsub/service.py#L985-L1020 | def change_node_subscriptions(self, jid, node, subscriptions_to_set):
"""
Update the subscriptions at a node.
:param jid: Address of the PubSub service.
:type jid: :class:`aioxmpp.JID`
:param node: Name of the node to modify
:type node: :class:`str`
:param subscr... | [
"def",
"change_node_subscriptions",
"(",
"self",
",",
"jid",
",",
"node",
",",
"subscriptions_to_set",
")",
":",
"iq",
"=",
"aioxmpp",
".",
"stanza",
".",
"IQ",
"(",
"type_",
"=",
"aioxmpp",
".",
"structs",
".",
"IQType",
".",
"SET",
",",
"to",
"=",
"j... | Update the subscriptions at a node.
:param jid: Address of the PubSub service.
:type jid: :class:`aioxmpp.JID`
:param node: Name of the node to modify
:type node: :class:`str`
:param subscriptions_to_set: The subscriptions to set at the node.
:type subscriptions_to_set: ... | [
"Update",
"the",
"subscriptions",
"at",
"a",
"node",
"."
] | python | train | 37.722222 |
yahoo/TensorFlowOnSpark | examples/imagenet/inception/image_processing.py | https://github.com/yahoo/TensorFlowOnSpark/blob/5e4b6c185ab722fd0104ede0377e1149ea8d6f7c/examples/imagenet/inception/image_processing.py#L304-L336 | def image_preprocessing(image_buffer, bbox, train, thread_id=0):
"""Decode and preprocess one image for evaluation or training.
Args:
image_buffer: JPEG encoded string Tensor
bbox: 3-D float Tensor of bounding boxes arranged [1, num_boxes, coords]
where each coordinate is [0, 1) and the coordinates a... | [
"def",
"image_preprocessing",
"(",
"image_buffer",
",",
"bbox",
",",
"train",
",",
"thread_id",
"=",
"0",
")",
":",
"if",
"bbox",
"is",
"None",
":",
"raise",
"ValueError",
"(",
"'Please supply a bounding box.'",
")",
"image",
"=",
"decode_jpeg",
"(",
"image_bu... | Decode and preprocess one image for evaluation or training.
Args:
image_buffer: JPEG encoded string Tensor
bbox: 3-D float Tensor of bounding boxes arranged [1, num_boxes, coords]
where each coordinate is [0, 1) and the coordinates are arranged as
[ymin, xmin, ymax, xmax].
train: boolean
... | [
"Decode",
"and",
"preprocess",
"one",
"image",
"for",
"evaluation",
"or",
"training",
"."
] | python | train | 29.757576 |
dArignac/pelican-extended-sitemap | extended_sitemap/__init__.py | https://github.com/dArignac/pelican-extended-sitemap/blob/1cf7746c071c303db3b321955a91b3a78d9585f8/extended_sitemap/__init__.py#L218-L251 | def __create_url_node_for_content(self, content, content_type, url=None, modification_time=None):
"""
Creates the required <url> node for the sitemap xml.
:param content: the content class to handle
:type content: pelican.contents.Content | None
:param content_type: the type of t... | [
"def",
"__create_url_node_for_content",
"(",
"self",
",",
"content",
",",
"content_type",
",",
"url",
"=",
"None",
",",
"modification_time",
"=",
"None",
")",
":",
"loc",
"=",
"url",
"if",
"loc",
"is",
"None",
":",
"loc",
"=",
"urljoin",
"(",
"self",
"."... | Creates the required <url> node for the sitemap xml.
:param content: the content class to handle
:type content: pelican.contents.Content | None
:param content_type: the type of the given content to match settings.EXTENDED_SITEMAP_PLUGIN
:type content_type; str
:param url; if give... | [
"Creates",
"the",
"required",
"<url",
">",
"node",
"for",
"the",
"sitemap",
"xml",
".",
":",
"param",
"content",
":",
"the",
"content",
"class",
"to",
"handle",
":",
"type",
"content",
":",
"pelican",
".",
"contents",
".",
"Content",
"|",
"None",
":",
... | python | train | 51.382353 |
stain/forgetSQL | lib/forgetSQL.py | https://github.com/stain/forgetSQL/blob/2e13f983020b121fd75a95fcafce3ea75573fb6b/lib/forgetSQL.py#L608-L629 | def _nextSequence(cls, name=None):
"""Return a new sequence number for insertion in self._sqlTable.
Note that if your sequences are not named
tablename_primarykey_seq (ie. for table 'blapp' with primary
key 'john_id', sequence name blapp_john_id_seq) you must give
the full se... | [
"def",
"_nextSequence",
"(",
"cls",
",",
"name",
"=",
"None",
")",
":",
"if",
"not",
"name",
":",
"name",
"=",
"cls",
".",
"_sqlSequence",
"if",
"not",
"name",
":",
"# Assume it's tablename_primarykey_seq",
"if",
"len",
"(",
"cls",
".",
"_sqlPrimary",
")",... | Return a new sequence number for insertion in self._sqlTable.
Note that if your sequences are not named
tablename_primarykey_seq (ie. for table 'blapp' with primary
key 'john_id', sequence name blapp_john_id_seq) you must give
the full sequence name as an optional argument to _nextSe... | [
"Return",
"a",
"new",
"sequence",
"number",
"for",
"insertion",
"in",
"self",
".",
"_sqlTable",
"."
] | python | train | 42.954545 |
DataDog/integrations-core | datadog_checks_base/datadog_checks/base/checks/win/wmi/__init__.py | https://github.com/DataDog/integrations-core/blob/ebd41c873cf9f97a8c51bf9459bc6a7536af8acd/datadog_checks_base/datadog_checks/base/checks/win/wmi/__init__.py#L263-L311 | def from_time(
year=None, month=None, day=None, hours=None, minutes=None, seconds=None, microseconds=None, timezone=None
):
"""Convenience wrapper to take a series of date/time elements and return a WMI time
of the form `yyyymmddHHMMSS.mmmmmm+UUU`. All elements may be int, string or
omitted altogether. ... | [
"def",
"from_time",
"(",
"year",
"=",
"None",
",",
"month",
"=",
"None",
",",
"day",
"=",
"None",
",",
"hours",
"=",
"None",
",",
"minutes",
"=",
"None",
",",
"seconds",
"=",
"None",
",",
"microseconds",
"=",
"None",
",",
"timezone",
"=",
"None",
"... | Convenience wrapper to take a series of date/time elements and return a WMI time
of the form `yyyymmddHHMMSS.mmmmmm+UUU`. All elements may be int, string or
omitted altogether. If omitted, they will be replaced in the output string
by a series of stars of the appropriate length.
:param year: The year el... | [
"Convenience",
"wrapper",
"to",
"take",
"a",
"series",
"of",
"date",
"/",
"time",
"elements",
"and",
"return",
"a",
"WMI",
"time",
"of",
"the",
"form",
"yyyymmddHHMMSS",
".",
"mmmmmm",
"+",
"UUU",
".",
"All",
"elements",
"may",
"be",
"int",
"string",
"or... | python | train | 36.163265 |
robotools/fontParts | Lib/fontParts/base/normalizers.py | https://github.com/robotools/fontParts/blob/d2ff106fe95f9d566161d936a645157626568712/Lib/fontParts/base/normalizers.py#L379-L389 | def normalizeGlyphHeight(value):
"""
Normalizes glyph height.
* **value** must be a :ref:`type-int-float`.
* Returned value is the same type as the input value.
"""
if not isinstance(value, (int, float)):
raise TypeError("Glyph height must be an :ref:`type-int-float`, not "
... | [
"def",
"normalizeGlyphHeight",
"(",
"value",
")",
":",
"if",
"not",
"isinstance",
"(",
"value",
",",
"(",
"int",
",",
"float",
")",
")",
":",
"raise",
"TypeError",
"(",
"\"Glyph height must be an :ref:`type-int-float`, not \"",
"\"%s.\"",
"%",
"type",
"(",
"valu... | Normalizes glyph height.
* **value** must be a :ref:`type-int-float`.
* Returned value is the same type as the input value. | [
"Normalizes",
"glyph",
"height",
"."
] | python | train | 33.454545 |
mcrute/pydora | pydora/player.py | https://github.com/mcrute/pydora/blob/d9e353e7f19da741dcf372246b4d5640cb788488/pydora/player.py#L223-L232 | def input(self, input, song):
"""Input callback, handles key presses
"""
try:
cmd = getattr(self, self.CMD_MAP[input][1])
except (IndexError, KeyError):
return self.screen.print_error(
"Invalid command {!r}!".format(input))
cmd(song) | [
"def",
"input",
"(",
"self",
",",
"input",
",",
"song",
")",
":",
"try",
":",
"cmd",
"=",
"getattr",
"(",
"self",
",",
"self",
".",
"CMD_MAP",
"[",
"input",
"]",
"[",
"1",
"]",
")",
"except",
"(",
"IndexError",
",",
"KeyError",
")",
":",
"return"... | Input callback, handles key presses | [
"Input",
"callback",
"handles",
"key",
"presses"
] | python | valid | 30.5 |
genialis/resolwe | resolwe/process/runtime.py | https://github.com/genialis/resolwe/blob/f7bb54932c81ec0cfc5b5e80d238fceaeaa48d86/resolwe/process/runtime.py#L177-L193 | def run_process(self, slug, inputs):
"""Run a new process from a running process."""
def export_files(value):
"""Export input files of spawned process."""
if isinstance(value, str) and os.path.isfile(value):
# TODO: Use the protocol to export files and get the
... | [
"def",
"run_process",
"(",
"self",
",",
"slug",
",",
"inputs",
")",
":",
"def",
"export_files",
"(",
"value",
")",
":",
"\"\"\"Export input files of spawned process.\"\"\"",
"if",
"isinstance",
"(",
"value",
",",
"str",
")",
"and",
"os",
".",
"path",
".",
"i... | Run a new process from a running process. | [
"Run",
"a",
"new",
"process",
"from",
"a",
"running",
"process",
"."
] | python | train | 45.647059 |
crs4/pydoop | pydoop/hadut.py | https://github.com/crs4/pydoop/blob/f375be2a06f9c67eaae3ce6f605195dbca143b2b/pydoop/hadut.py#L338-L395 | def run_pipes(executable, input_path, output_path, more_args=None,
properties=None, force_pydoop_submitter=False,
hadoop_conf_dir=None, logger=None, keep_streams=False):
"""
Run a pipes command.
``more_args`` (after setting input/output path) and ``properties``
are passed to... | [
"def",
"run_pipes",
"(",
"executable",
",",
"input_path",
",",
"output_path",
",",
"more_args",
"=",
"None",
",",
"properties",
"=",
"None",
",",
"force_pydoop_submitter",
"=",
"False",
",",
"hadoop_conf_dir",
"=",
"None",
",",
"logger",
"=",
"None",
",",
"k... | Run a pipes command.
``more_args`` (after setting input/output path) and ``properties``
are passed to :func:`run_cmd`.
If not specified otherwise, this function sets the properties
``mapreduce.pipes.isjavarecordreader`` and
``mapreduce.pipes.isjavarecordwriter`` to ``"true"``.
This function w... | [
"Run",
"a",
"pipes",
"command",
"."
] | python | train | 42.741379 |
brocade/pynos | pynos/versions/ver_7/ver_7_1_0/yang/brocade_lldp_ext.py | https://github.com/brocade/pynos/blob/bd8a34e98f322de3fc06750827d8bbc3a0c00380/pynos/versions/ver_7/ver_7_1_0/yang/brocade_lldp_ext.py#L338-L354 | def get_lldp_neighbor_detail_output_lldp_neighbor_detail_lldp_pdu_received(self, **kwargs):
"""Auto Generated Code
"""
config = ET.Element("config")
get_lldp_neighbor_detail = ET.Element("get_lldp_neighbor_detail")
config = get_lldp_neighbor_detail
output = ET.SubElement(... | [
"def",
"get_lldp_neighbor_detail_output_lldp_neighbor_detail_lldp_pdu_received",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"config",
"=",
"ET",
".",
"Element",
"(",
"\"config\"",
")",
"get_lldp_neighbor_detail",
"=",
"ET",
".",
"Element",
"(",
"\"get_lldp_neighb... | Auto Generated Code | [
"Auto",
"Generated",
"Code"
] | python | train | 58.882353 |
farshidce/touchworks-python | touchworks/api/http.py | https://github.com/farshidce/touchworks-python/blob/ea8f93a0f4273de1317a318e945a571f5038ba62/touchworks/api/http.py#L869-L897 | def get_schedule(self, ehr_username, start_date,
changed_since, include_pix, other_user='All',
end_date='',
appointment_types=None, status_filter='All'):
"""
invokes TouchWorksMagicConstants.ACTION_GET_SCHEDULE action
:return: JSON r... | [
"def",
"get_schedule",
"(",
"self",
",",
"ehr_username",
",",
"start_date",
",",
"changed_since",
",",
"include_pix",
",",
"other_user",
"=",
"'All'",
",",
"end_date",
"=",
"''",
",",
"appointment_types",
"=",
"None",
",",
"status_filter",
"=",
"'All'",
")",
... | invokes TouchWorksMagicConstants.ACTION_GET_SCHEDULE action
:return: JSON response | [
"invokes",
"TouchWorksMagicConstants",
".",
"ACTION_GET_SCHEDULE",
"action",
":",
"return",
":",
"JSON",
"response"
] | python | train | 47.448276 |
RRZE-HPC/kerncraft | kerncraft/kerncraft.py | https://github.com/RRZE-HPC/kerncraft/blob/c60baf8043e4da8d8d66da7575021c2f4c6c78af/kerncraft/kerncraft.py#L70-L80 | def get_last_modified_datetime(dir_path=os.path.dirname(__file__)):
"""Return datetime object of latest change in kerncraft module directory."""
max_mtime = 0
for root, dirs, files in os.walk(dir_path):
for f in files:
p = os.path.join(root, f)
try:
max_mtime ... | [
"def",
"get_last_modified_datetime",
"(",
"dir_path",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
")",
":",
"max_mtime",
"=",
"0",
"for",
"root",
",",
"dirs",
",",
"files",
"in",
"os",
".",
"walk",
"(",
"dir_path",
")",
":",
"for",
... | Return datetime object of latest change in kerncraft module directory. | [
"Return",
"datetime",
"object",
"of",
"latest",
"change",
"in",
"kerncraft",
"module",
"directory",
"."
] | python | test | 41.272727 |
tensorflow/cleverhans | cleverhans/dataset.py | https://github.com/tensorflow/cleverhans/blob/97488e215760547b81afc53f5e5de8ba7da5bd98/cleverhans/dataset.py#L269-L307 | def data_cifar10(train_start=0, train_end=50000, test_start=0, test_end=10000):
"""
Preprocess CIFAR10 dataset
:return:
"""
# These values are specific to CIFAR10
img_rows = 32
img_cols = 32
nb_classes = 10
# the data, shuffled and split between train and test sets
(x_train, y_train), (x_test, y_... | [
"def",
"data_cifar10",
"(",
"train_start",
"=",
"0",
",",
"train_end",
"=",
"50000",
",",
"test_start",
"=",
"0",
",",
"test_end",
"=",
"10000",
")",
":",
"# These values are specific to CIFAR10",
"img_rows",
"=",
"32",
"img_cols",
"=",
"32",
"nb_classes",
"="... | Preprocess CIFAR10 dataset
:return: | [
"Preprocess",
"CIFAR10",
"dataset",
":",
"return",
":"
] | python | train | 32.846154 |
saltstack/salt | salt/utils/zfs.py | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/utils/zfs.py#L495-L527 | def to_size(value, convert_to_human=True):
'''
Convert python int (bytes) to zfs size
NOTE: http://src.illumos.org/source/xref/illumos-gate/usr/src/lib/pyzfs/common/util.py#114
'''
value = from_size(value)
if value is None:
value = 'none'
if isinstance(value, Number) and value > 10... | [
"def",
"to_size",
"(",
"value",
",",
"convert_to_human",
"=",
"True",
")",
":",
"value",
"=",
"from_size",
"(",
"value",
")",
"if",
"value",
"is",
"None",
":",
"value",
"=",
"'none'",
"if",
"isinstance",
"(",
"value",
",",
"Number",
")",
"and",
"value"... | Convert python int (bytes) to zfs size
NOTE: http://src.illumos.org/source/xref/illumos-gate/usr/src/lib/pyzfs/common/util.py#114 | [
"Convert",
"python",
"int",
"(",
"bytes",
")",
"to",
"zfs",
"size"
] | python | train | 32.969697 |
CityOfZion/neo-python | examples/build_raw_transactions.py | https://github.com/CityOfZion/neo-python/blob/fe90f62e123d720d4281c79af0598d9df9e776fb/examples/build_raw_transactions.py#L219-L232 | def address_to_scripthash(address: str) -> UInt160:
"""Just a helper method"""
AddressVersion = 23 # fixed at this point
data = b58decode(address)
if len(data) != 25:
raise ValueError('Not correct Address, wrong length.')
if data[0] != AddressVersion:
raise ValueError('Not correct C... | [
"def",
"address_to_scripthash",
"(",
"address",
":",
"str",
")",
"->",
"UInt160",
":",
"AddressVersion",
"=",
"23",
"# fixed at this point",
"data",
"=",
"b58decode",
"(",
"address",
")",
"if",
"len",
"(",
"data",
")",
"!=",
"25",
":",
"raise",
"ValueError",... | Just a helper method | [
"Just",
"a",
"helper",
"method"
] | python | train | 39.142857 |
openego/ding0 | ding0/core/network/grids.py | https://github.com/openego/ding0/blob/e2d6528f96255e4bb22ba15514a4f1883564ed5d/ding0/core/network/grids.py#L158-L165 | def remove_cable_distributor(self, cable_dist):
"""Removes a cable distributor from _cable_distributors if existing"""
if cable_dist in self.cable_distributors() and isinstance(cable_dist,
MVCableDistributorDing0):
# remove fr... | [
"def",
"remove_cable_distributor",
"(",
"self",
",",
"cable_dist",
")",
":",
"if",
"cable_dist",
"in",
"self",
".",
"cable_distributors",
"(",
")",
"and",
"isinstance",
"(",
"cable_dist",
",",
"MVCableDistributorDing0",
")",
":",
"# remove from array and graph",
"se... | Removes a cable distributor from _cable_distributors if existing | [
"Removes",
"a",
"cable",
"distributor",
"from",
"_cable_distributors",
"if",
"existing"
] | python | train | 61 |
chrislit/abydos | abydos/stats/_confusion_table.py | https://github.com/chrislit/abydos/blob/165466b3ff6afd8024a4c8660421b0c4e7773db9/abydos/stats/_confusion_table.py#L1347-L1378 | def kappa_statistic(self):
r"""Return κ statistic.
The κ statistic is defined as:
:math:`\kappa = \frac{accuracy - random~ accuracy}
{1 - random~ accuracy}`
The κ statistic compares the performance of the classifier relative to
the performance of a random classifier. :m... | [
"def",
"kappa_statistic",
"(",
"self",
")",
":",
"if",
"self",
".",
"population",
"(",
")",
"==",
"0",
":",
"return",
"float",
"(",
"'NaN'",
")",
"random_accuracy",
"=",
"(",
"(",
"self",
".",
"_tn",
"+",
"self",
".",
"_fp",
")",
"*",
"(",
"self",
... | r"""Return κ statistic.
The κ statistic is defined as:
:math:`\kappa = \frac{accuracy - random~ accuracy}
{1 - random~ accuracy}`
The κ statistic compares the performance of the classifier relative to
the performance of a random classifier. :math:`\kappa` = 0 indicates
... | [
"r",
"Return",
"κ",
"statistic",
"."
] | python | valid | 33.21875 |
klahnakoski/pyLibrary | mo_math/stats.py | https://github.com/klahnakoski/pyLibrary/blob/fa2dcbc48fda8d26999baef400e9a98149e0b982/mo_math/stats.py#L353-L370 | def percentile(values, percent):
"""
PERCENTILE WITH INTERPOLATION
RETURN VALUE AT, OR ABOVE, percentile OF THE VALUES
snagged from http://code.activestate.com/recipes/511478-finding-the-percentile-of-the-values/
"""
N = sorted(values)
if not N:
return None
k = (len(N) - 1) * pe... | [
"def",
"percentile",
"(",
"values",
",",
"percent",
")",
":",
"N",
"=",
"sorted",
"(",
"values",
")",
"if",
"not",
"N",
":",
"return",
"None",
"k",
"=",
"(",
"len",
"(",
"N",
")",
"-",
"1",
")",
"*",
"percent",
"f",
"=",
"int",
"(",
"math",
"... | PERCENTILE WITH INTERPOLATION
RETURN VALUE AT, OR ABOVE, percentile OF THE VALUES
snagged from http://code.activestate.com/recipes/511478-finding-the-percentile-of-the-values/ | [
"PERCENTILE",
"WITH",
"INTERPOLATION",
"RETURN",
"VALUE",
"AT",
"OR",
"ABOVE",
"percentile",
"OF",
"THE",
"VALUES"
] | python | train | 26 |
praekeltfoundation/molo.commenting | molo/commenting/templatetags/molo_commenting_tags.py | https://github.com/praekeltfoundation/molo.commenting/blob/94549bd75e4a5c5b3db43149e32d636330b3969c/molo/commenting/templatetags/molo_commenting_tags.py#L75-L95 | def get_comments_content_object(parser, token):
"""
Get a limited set of comments for a given object.
Defaults to a limit of 5. Setting the limit to -1 disables limiting.
usage:
{% get_comments_content_object for form_object as variable_name %}
"""
keywords = token.contents.split()
... | [
"def",
"get_comments_content_object",
"(",
"parser",
",",
"token",
")",
":",
"keywords",
"=",
"token",
".",
"contents",
".",
"split",
"(",
")",
"if",
"len",
"(",
"keywords",
")",
"!=",
"5",
":",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"\"'%s'... | Get a limited set of comments for a given object.
Defaults to a limit of 5. Setting the limit to -1 disables limiting.
usage:
{% get_comments_content_object for form_object as variable_name %} | [
"Get",
"a",
"limited",
"set",
"of",
"comments",
"for",
"a",
"given",
"object",
".",
"Defaults",
"to",
"a",
"limit",
"of",
"5",
".",
"Setting",
"the",
"limit",
"to",
"-",
"1",
"disables",
"limiting",
"."
] | python | train | 37.47619 |
ericsomdahl/python-bittrex | bittrex/bittrex.py | https://github.com/ericsomdahl/python-bittrex/blob/2dbc08e3221e07a9e618eaa025d98ed197d28e31/bittrex/bittrex.py#L401-L418 | def get_open_orders(self, market=None):
"""
Get all orders that you currently have opened.
A specific market can be requested.
Endpoint:
1.1 /market/getopenorders
2.0 /key/market/getopenorders
:param market: String literal for the market (ie. BTC-LTC)
:t... | [
"def",
"get_open_orders",
"(",
"self",
",",
"market",
"=",
"None",
")",
":",
"return",
"self",
".",
"_api_query",
"(",
"path_dict",
"=",
"{",
"API_V1_1",
":",
"'/market/getopenorders'",
",",
"API_V2_0",
":",
"'/key/market/getopenorders'",
"}",
",",
"options",
... | Get all orders that you currently have opened.
A specific market can be requested.
Endpoint:
1.1 /market/getopenorders
2.0 /key/market/getopenorders
:param market: String literal for the market (ie. BTC-LTC)
:type market: str
:return: Open orders info in JSON
... | [
"Get",
"all",
"orders",
"that",
"you",
"currently",
"have",
"opened",
".",
"A",
"specific",
"market",
"can",
"be",
"requested",
"."
] | python | train | 35.666667 |
aaren/notedown | notedown/notedown.py | https://github.com/aaren/notedown/blob/1e920c7e4ecbe47420c12eed3d5bcae735121222/notedown/notedown.py#L322-L327 | def create_markdown_cell(block):
"""Create a markdown cell from a block."""
kwargs = {'cell_type': block['type'],
'source': block['content']}
markdown_cell = nbbase.new_markdown_cell(**kwargs)
return markdown_cell | [
"def",
"create_markdown_cell",
"(",
"block",
")",
":",
"kwargs",
"=",
"{",
"'cell_type'",
":",
"block",
"[",
"'type'",
"]",
",",
"'source'",
":",
"block",
"[",
"'content'",
"]",
"}",
"markdown_cell",
"=",
"nbbase",
".",
"new_markdown_cell",
"(",
"*",
"*",
... | Create a markdown cell from a block. | [
"Create",
"a",
"markdown",
"cell",
"from",
"a",
"block",
"."
] | python | train | 43 |
calston/rhumba | rhumba/backends/redis.py | https://github.com/calston/rhumba/blob/05e3cbf4e531cc51b4777912eb98a4f006893f5e/rhumba/backends/redis.py#L207-L226 | def clusterQueues(self):
""" Return a dict of queues in cluster and servers running them
"""
servers = yield self.getClusterServers()
queues = {}
for sname in servers:
qs = yield self.get('rhumba.server.%s.queues' % sname)
uuid = yield self.get('rhumba.s... | [
"def",
"clusterQueues",
"(",
"self",
")",
":",
"servers",
"=",
"yield",
"self",
".",
"getClusterServers",
"(",
")",
"queues",
"=",
"{",
"}",
"for",
"sname",
"in",
"servers",
":",
"qs",
"=",
"yield",
"self",
".",
"get",
"(",
"'rhumba.server.%s.queues'",
"... | Return a dict of queues in cluster and servers running them | [
"Return",
"a",
"dict",
"of",
"queues",
"in",
"cluster",
"and",
"servers",
"running",
"them"
] | python | train | 28.05 |
OSLL/jabba | jabba/synonym_parser.py | https://github.com/OSLL/jabba/blob/71c1d008ab497020fba6ffa12a600721eb3f5ef7/jabba/synonym_parser.py#L79-L94 | def parse_from_array(arr):
"""
Parse 2d array into synonym set
Every array inside arr is considered a set of synonyms
"""
syn_set = SynonymSet()
for synonyms in arr:
_set = set()
for synonym in synonyms:
_set.add(synonym)
syn_set.add_set(_set)
return s... | [
"def",
"parse_from_array",
"(",
"arr",
")",
":",
"syn_set",
"=",
"SynonymSet",
"(",
")",
"for",
"synonyms",
"in",
"arr",
":",
"_set",
"=",
"set",
"(",
")",
"for",
"synonym",
"in",
"synonyms",
":",
"_set",
".",
"add",
"(",
"synonym",
")",
"syn_set",
"... | Parse 2d array into synonym set
Every array inside arr is considered a set of synonyms | [
"Parse",
"2d",
"array",
"into",
"synonym",
"set",
"Every",
"array",
"inside",
"arr",
"is",
"considered",
"a",
"set",
"of",
"synonyms"
] | python | train | 19.4375 |
coumbole/mailscanner | mailscanner/reader.py | https://github.com/coumbole/mailscanner/blob/ead19ac8c7dee27e507c1593032863232c13f636/mailscanner/reader.py#L75-L111 | def fetch_all_messages(self, conn, directory, readonly):
""" Fetches all messages at @conn from @directory.
Params:
conn IMAP4_SSL connection
directory The IMAP directory to look for
readonly readonly mode, true or false
Return... | [
"def",
"fetch_all_messages",
"(",
"self",
",",
"conn",
",",
"directory",
",",
"readonly",
")",
":",
"conn",
".",
"select",
"(",
"directory",
",",
"readonly",
")",
"message_data",
"=",
"[",
"]",
"typ",
",",
"data",
"=",
"conn",
".",
"search",
"(",
"None... | Fetches all messages at @conn from @directory.
Params:
conn IMAP4_SSL connection
directory The IMAP directory to look for
readonly readonly mode, true or false
Returns:
List of subject-body tuples | [
"Fetches",
"all",
"messages",
"at",
"@conn",
"from",
"@directory",
"."
] | python | train | 28.135135 |
KelSolaar/Umbra | umbra/components/factory/script_editor/search_in_files.py | https://github.com/KelSolaar/Umbra/blob/66f45f08d9d723787f1191989f8b0dda84b412ce/umbra/components/factory/script_editor/search_in_files.py#L1405-L1443 | def save_files(self, nodes):
"""
Saves user defined files using give nodes.
:param nodes: Nodes.
:type nodes: list
:return: Method success.
:rtype: bool
"""
metrics = {"Opened": 0, "Cached": 0}
for node in nodes:
file = node.file
... | [
"def",
"save_files",
"(",
"self",
",",
"nodes",
")",
":",
"metrics",
"=",
"{",
"\"Opened\"",
":",
"0",
",",
"\"Cached\"",
":",
"0",
"}",
"for",
"node",
"in",
"nodes",
":",
"file",
"=",
"node",
".",
"file",
"if",
"self",
".",
"__container",
".",
"ge... | Saves user defined files using give nodes.
:param nodes: Nodes.
:type nodes: list
:return: Method success.
:rtype: bool | [
"Saves",
"user",
"defined",
"files",
"using",
"give",
"nodes",
"."
] | python | train | 44.102564 |
saltstack/salt | salt/pillar/reclass_adapter.py | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/pillar/reclass_adapter.py#L93-L135 | def ext_pillar(minion_id, pillar, **kwargs):
'''
Obtain the Pillar data from **reclass** for the given ``minion_id``.
'''
# If reclass is installed, __virtual__ put it onto the search path, so we
# don't need to protect against ImportError:
# pylint: disable=3rd-party-module-not-gated
from ... | [
"def",
"ext_pillar",
"(",
"minion_id",
",",
"pillar",
",",
"*",
"*",
"kwargs",
")",
":",
"# If reclass is installed, __virtual__ put it onto the search path, so we",
"# don't need to protect against ImportError:",
"# pylint: disable=3rd-party-module-not-gated",
"from",
"reclass",
"... | Obtain the Pillar data from **reclass** for the given ``minion_id``. | [
"Obtain",
"the",
"Pillar",
"data",
"from",
"**",
"reclass",
"**",
"for",
"the",
"given",
"minion_id",
"."
] | python | train | 39.27907 |
raiden-network/raiden | raiden/encoding/format.py | https://github.com/raiden-network/raiden/blob/407ba15c72074e9de88771d6b9661ff4dc36bef5/raiden/encoding/format.py#L58-L189 | def namedbuffer(buffer_name, fields_spec): # noqa (ignore ciclomatic complexity)
""" Class factory, returns a class to wrap a buffer instance and expose the
data as fields.
The field spec specifies how many bytes should be used for a field and what
is the encoding / decoding function.
"""
# py... | [
"def",
"namedbuffer",
"(",
"buffer_name",
",",
"fields_spec",
")",
":",
"# noqa (ignore ciclomatic complexity)",
"# pylint: disable=protected-access,unused-argument",
"if",
"not",
"len",
"(",
"buffer_name",
")",
":",
"raise",
"ValueError",
"(",
"'buffer_name is empty'",
")"... | Class factory, returns a class to wrap a buffer instance and expose the
data as fields.
The field spec specifies how many bytes should be used for a field and what
is the encoding / decoding function. | [
"Class",
"factory",
"returns",
"a",
"class",
"to",
"wrap",
"a",
"buffer",
"instance",
"and",
"expose",
"the",
"data",
"as",
"fields",
"."
] | python | train | 30.280303 |
robertpeteuil/multi-cloud-control | mcc/uimode.py | https://github.com/robertpeteuil/multi-cloud-control/blob/f1565af1c0b6ed465ff312d3ccc592ba0609f4a2/mcc/uimode.py#L343-L350 | def input_yn(conf_mess):
"""Print Confirmation Message and Get Y/N response from user."""
ui_erase_ln()
ui_print(conf_mess)
with term.cbreak():
input_flush()
val = input_by_key()
return bool(val.lower() == 'y') | [
"def",
"input_yn",
"(",
"conf_mess",
")",
":",
"ui_erase_ln",
"(",
")",
"ui_print",
"(",
"conf_mess",
")",
"with",
"term",
".",
"cbreak",
"(",
")",
":",
"input_flush",
"(",
")",
"val",
"=",
"input_by_key",
"(",
")",
"return",
"bool",
"(",
"val",
".",
... | Print Confirmation Message and Get Y/N response from user. | [
"Print",
"Confirmation",
"Message",
"and",
"Get",
"Y",
"/",
"N",
"response",
"from",
"user",
"."
] | python | train | 29.875 |
JoseAntFer/pyny3d | pyny3d/geoms.py | https://github.com/JoseAntFer/pyny3d/blob/fb81684935a24f7e50c975cb4383c81a63ab56df/pyny3d/geoms.py#L415-L424 | def get_domain(self):
"""
:returns: opposite vertices of the bounding prism for this
object.
:rtype: ndarray([min], [max])
"""
if self.domain is None:
return np.array([self.points.min(axis=0),
self.points.max(axis=0)]... | [
"def",
"get_domain",
"(",
"self",
")",
":",
"if",
"self",
".",
"domain",
"is",
"None",
":",
"return",
"np",
".",
"array",
"(",
"[",
"self",
".",
"points",
".",
"min",
"(",
"axis",
"=",
"0",
")",
",",
"self",
".",
"points",
".",
"max",
"(",
"axi... | :returns: opposite vertices of the bounding prism for this
object.
:rtype: ndarray([min], [max]) | [
":",
"returns",
":",
"opposite",
"vertices",
"of",
"the",
"bounding",
"prism",
"for",
"this",
"object",
".",
":",
"rtype",
":",
"ndarray",
"(",
"[",
"min",
"]",
"[",
"max",
"]",
")"
] | python | train | 34 |
datacamp/sqlwhat | sqlwhat/checks/check_funcs.py | https://github.com/datacamp/sqlwhat/blob/9ae798c63124f994607a0e2c120b24ebbb2bdbe9/sqlwhat/checks/check_funcs.py#L471-L540 | def check_query(state, query, error_msg=None, expand_msg=None):
"""Run arbitrary queries against to the DB connection to verify the database state.
For queries that do not return any output (INSERTs, UPDATEs, ...),
you cannot use functions like ``check_col()`` and ``is_equal()`` to verify the query result.... | [
"def",
"check_query",
"(",
"state",
",",
"query",
",",
"error_msg",
"=",
"None",
",",
"expand_msg",
"=",
"None",
")",
":",
"if",
"error_msg",
"is",
"None",
":",
"error_msg",
"=",
"\"Running `{{query}}` after your submission generated an error.\"",
"if",
"expand_msg"... | Run arbitrary queries against to the DB connection to verify the database state.
For queries that do not return any output (INSERTs, UPDATEs, ...),
you cannot use functions like ``check_col()`` and ``is_equal()`` to verify the query result.
``check_query()`` will rerun the solution query in the transactio... | [
"Run",
"arbitrary",
"queries",
"against",
"to",
"the",
"DB",
"connection",
"to",
"verify",
"the",
"database",
"state",
"."
] | python | train | 39.057143 |
Yubico/python-yubico | yubico/yubikey_config.py | https://github.com/Yubico/python-yubico/blob/a72e8eddb90da6ee96e29f60912ca1f2872c9aea/yubico/yubikey_config.py#L188-L205 | def fixed_string(self, data=None):
"""
The fixed string is used to identify a particular Yubikey device.
The fixed string is referred to as the 'Token Identifier' in OATH-HOTP mode.
The length of the fixed string can be set between 0 and 16 bytes.
Tip: This can also be used to... | [
"def",
"fixed_string",
"(",
"self",
",",
"data",
"=",
"None",
")",
":",
"old",
"=",
"self",
".",
"fixed",
"if",
"data",
"!=",
"None",
":",
"new",
"=",
"self",
".",
"_decode_input_string",
"(",
"data",
")",
"if",
"len",
"(",
"new",
")",
"<=",
"16",
... | The fixed string is used to identify a particular Yubikey device.
The fixed string is referred to as the 'Token Identifier' in OATH-HOTP mode.
The length of the fixed string can be set between 0 and 16 bytes.
Tip: This can also be used to extend the length of a static password. | [
"The",
"fixed",
"string",
"is",
"used",
"to",
"identify",
"a",
"particular",
"Yubikey",
"device",
"."
] | python | train | 36 |
mozilla/mozdownload | mozdownload/scraper.py | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L552-L564 | def path_regex(self):
"""Return the regex for the path to the build folder."""
try:
path = '%s/' % urljoin(self.monthly_build_list_regex,
self.builds[self.build_index])
if self.application in APPLICATIONS_MULTI_LOCALE \
and s... | [
"def",
"path_regex",
"(",
"self",
")",
":",
"try",
":",
"path",
"=",
"'%s/'",
"%",
"urljoin",
"(",
"self",
".",
"monthly_build_list_regex",
",",
"self",
".",
"builds",
"[",
"self",
".",
"build_index",
"]",
")",
"if",
"self",
".",
"application",
"in",
"... | Return the regex for the path to the build folder. | [
"Return",
"the",
"regex",
"for",
"the",
"path",
"to",
"the",
"build",
"folder",
"."
] | python | train | 49.153846 |
INM-6/hybridLFPy | examples/example_microcircuit_params.py | https://github.com/INM-6/hybridLFPy/blob/c38bdf38982c4624c2f70caeb50c40f1d5980abd/examples/example_microcircuit_params.py#L44-L58 | def get_F_y(fname='binzegger_connectivity_table.json', y=['p23']):
'''
Extract frequency of occurrences of those cell types that are modeled.
The data set contains cell types that are not modeled (TCs etc.)
The returned percentages are renormalized onto modeled cell-types, i.e. they sum up to 1
''... | [
"def",
"get_F_y",
"(",
"fname",
"=",
"'binzegger_connectivity_table.json'",
",",
"y",
"=",
"[",
"'p23'",
"]",
")",
":",
"# Load data from json dictionary",
"f",
"=",
"open",
"(",
"fname",
",",
"'r'",
")",
"data",
"=",
"json",
".",
"load",
"(",
"f",
")",
... | Extract frequency of occurrences of those cell types that are modeled.
The data set contains cell types that are not modeled (TCs etc.)
The returned percentages are renormalized onto modeled cell-types, i.e. they sum up to 1 | [
"Extract",
"frequency",
"of",
"occurrences",
"of",
"those",
"cell",
"types",
"that",
"are",
"modeled",
".",
"The",
"data",
"set",
"contains",
"cell",
"types",
"that",
"are",
"not",
"modeled",
"(",
"TCs",
"etc",
".",
")",
"The",
"returned",
"percentages",
"... | python | train | 37.2 |
Metatab/metapack | metapack/jupyter/ipython.py | https://github.com/Metatab/metapack/blob/8365f221fbeaa3c0be9091f2eaf3447fd8e2e8d6/metapack/jupyter/ipython.py#L48-L103 | def open_package(locals=None, dr=None):
"""Try to open a package with the metatab_doc variable, which is set when a Notebook is run
as a resource. If that does not exist, try the local _packages directory"""
if locals is None:
locals = caller_locals()
try:
# Running in a package build... | [
"def",
"open_package",
"(",
"locals",
"=",
"None",
",",
"dr",
"=",
"None",
")",
":",
"if",
"locals",
"is",
"None",
":",
"locals",
"=",
"caller_locals",
"(",
")",
"try",
":",
"# Running in a package build",
"return",
"op",
"(",
"locals",
"[",
"'metatab_doc'... | Try to open a package with the metatab_doc variable, which is set when a Notebook is run
as a resource. If that does not exist, try the local _packages directory | [
"Try",
"to",
"open",
"a",
"package",
"with",
"the",
"metatab_doc",
"variable",
"which",
"is",
"set",
"when",
"a",
"Notebook",
"is",
"run",
"as",
"a",
"resource",
".",
"If",
"that",
"does",
"not",
"exist",
"try",
"the",
"local",
"_packages",
"directory"
] | python | train | 31.214286 |
puiterwijk/flask-oidc | flask_oidc/__init__.py | https://github.com/puiterwijk/flask-oidc/blob/7f16e27b926fc12953d6b2ae78a9b9cc9b8d1769/flask_oidc/__init__.py#L669-L683 | def custom_callback(self, view_func):
"""
Wrapper function to use a custom callback.
The custom OIDC callback will get the custom state field passed in with
redirect_to_auth_server.
"""
@wraps(view_func)
def decorated(*args, **kwargs):
plainreturn, dat... | [
"def",
"custom_callback",
"(",
"self",
",",
"view_func",
")",
":",
"@",
"wraps",
"(",
"view_func",
")",
"def",
"decorated",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"plainreturn",
",",
"data",
"=",
"self",
".",
"_process_callback",
"(",
"'c... | Wrapper function to use a custom callback.
The custom OIDC callback will get the custom state field passed in with
redirect_to_auth_server. | [
"Wrapper",
"function",
"to",
"use",
"a",
"custom",
"callback",
".",
"The",
"custom",
"OIDC",
"callback",
"will",
"get",
"the",
"custom",
"state",
"field",
"passed",
"in",
"with",
"redirect_to_auth_server",
"."
] | python | train | 35.933333 |
SMTG-UCL/sumo | sumo/cli/bandplot.py | https://github.com/SMTG-UCL/sumo/blob/47aec6bbfa033a624435a65bd4edabd18bfb437f/sumo/cli/bandplot.py#L306-L343 | def save_data_files(vr, bs, prefix=None, directory=None):
"""Write the band structure data files to disk.
Args:
vs (`Vasprun`): Pymatgen `Vasprun` object.
bs (`BandStructureSymmLine`): Calculated band structure.
prefix (`str`, optional): Prefix for data file.
directory (`str`, o... | [
"def",
"save_data_files",
"(",
"vr",
",",
"bs",
",",
"prefix",
"=",
"None",
",",
"directory",
"=",
"None",
")",
":",
"filename",
"=",
"'{}_band.dat'",
".",
"format",
"(",
"prefix",
")",
"if",
"prefix",
"else",
"'band.dat'",
"directory",
"=",
"directory",
... | Write the band structure data files to disk.
Args:
vs (`Vasprun`): Pymatgen `Vasprun` object.
bs (`BandStructureSymmLine`): Calculated band structure.
prefix (`str`, optional): Prefix for data file.
directory (`str`, optional): Directory in which to save the data.
Returns:
... | [
"Write",
"the",
"band",
"structure",
"data",
"files",
"to",
"disk",
"."
] | python | train | 34.526316 |
QInfer/python-qinfer | src/qinfer/tomography/plotting_tools.py | https://github.com/QInfer/python-qinfer/blob/8170c84a0be1723f8c6b09e0d3c7a40a886f1fe3/src/qinfer/tomography/plotting_tools.py#L156-L202 | def plot_rebit_prior(prior, rebit_axes=REBIT_AXES,
n_samples=2000, true_state=None, true_size=250,
force_mean=None,
legend=True,
mean_color_index=2
):
"""
Plots rebit states drawn from a given prior.
:param qinfer.tomography.DensityOperatorDistribution prior: Distributio... | [
"def",
"plot_rebit_prior",
"(",
"prior",
",",
"rebit_axes",
"=",
"REBIT_AXES",
",",
"n_samples",
"=",
"2000",
",",
"true_state",
"=",
"None",
",",
"true_size",
"=",
"250",
",",
"force_mean",
"=",
"None",
",",
"legend",
"=",
"True",
",",
"mean_color_index",
... | Plots rebit states drawn from a given prior.
:param qinfer.tomography.DensityOperatorDistribution prior: Distribution over
rebit states to plot.
:param list rebit_axes: List containing indices for the :math:`x`
and :math:`z` axes.
:param int n_samples: Number of samples to draw from the
... | [
"Plots",
"rebit",
"states",
"drawn",
"from",
"a",
"given",
"prior",
"."
] | python | train | 31.957447 |
xmikos/soapy_power | soapypower/power.py | https://github.com/xmikos/soapy_power/blob/46e12659b8d08af764dc09a1f31b0e85a68f808f/soapypower/power.py#L73-L81 | def nearest_overlap(self, overlap, bins):
"""Return nearest overlap/crop factor based on number of bins"""
bins_overlap = overlap * bins
if bins_overlap % 2 != 0:
bins_overlap = math.ceil(bins_overlap / 2) * 2
overlap = bins_overlap / bins
logger.warning('numb... | [
"def",
"nearest_overlap",
"(",
"self",
",",
"overlap",
",",
"bins",
")",
":",
"bins_overlap",
"=",
"overlap",
"*",
"bins",
"if",
"bins_overlap",
"%",
"2",
"!=",
"0",
":",
"bins_overlap",
"=",
"math",
".",
"ceil",
"(",
"bins_overlap",
"/",
"2",
")",
"*"... | Return nearest overlap/crop factor based on number of bins | [
"Return",
"nearest",
"overlap",
"/",
"crop",
"factor",
"based",
"on",
"number",
"of",
"bins"
] | python | test | 51.555556 |
fermiPy/fermipy | fermipy/diffuse/name_policy.py | https://github.com/fermiPy/fermipy/blob/9df5e7e3728307fd58c5bba36fd86783c39fbad4/fermipy/diffuse/name_policy.py#L311-L322 | def ft2file(self, **kwargs):
""" return the name of the input ft2 file list
"""
kwargs_copy = self.base_dict.copy()
kwargs_copy.update(**kwargs)
kwargs_copy['data_time'] = kwargs.get(
'data_time', self.dataset(**kwargs))
self._replace_none(kwargs_copy) ... | [
"def",
"ft2file",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"kwargs_copy",
"=",
"self",
".",
"base_dict",
".",
"copy",
"(",
")",
"kwargs_copy",
".",
"update",
"(",
"*",
"*",
"kwargs",
")",
"kwargs_copy",
"[",
"'data_time'",
"]",
"=",
"kwargs",
... | return the name of the input ft2 file list | [
"return",
"the",
"name",
"of",
"the",
"input",
"ft2",
"file",
"list"
] | python | train | 41.583333 |
bertrandvidal/parse_this | parse_this/core.py | https://github.com/bertrandvidal/parse_this/blob/aa2e3737f19642300ef1ca65cae21c90049718a2/parse_this/core.py#L118-L148 | def _get_parseable_methods(cls):
"""Return all methods of cls that are parseable i.e. have been decorated
by '@create_parser'.
Args:
cls: the class currently being decorated
Note:
classmethods will not be included as they can only be referenced once
the class has been defined
... | [
"def",
"_get_parseable_methods",
"(",
"cls",
")",
":",
"_LOG",
".",
"debug",
"(",
"\"Retrieving parseable methods for '%s'\"",
",",
"cls",
".",
"__name__",
")",
"init_parser",
"=",
"None",
"methods_to_parse",
"=",
"{",
"}",
"for",
"name",
",",
"obj",
"in",
"va... | Return all methods of cls that are parseable i.e. have been decorated
by '@create_parser'.
Args:
cls: the class currently being decorated
Note:
classmethods will not be included as they can only be referenced once
the class has been defined
Returns:
a 2-tuple with the p... | [
"Return",
"all",
"methods",
"of",
"cls",
"that",
"are",
"parseable",
"i",
".",
"e",
".",
"have",
"been",
"decorated",
"by",
"@create_parser",
"."
] | python | train | 40.451613 |
dropbox/stone | stone/frontend/parser.py | https://github.com/dropbox/stone/blob/2e95cbcd1c48e05cca68c919fd8d24adec6b0f58/stone/frontend/parser.py#L157-L167 | def p_namespace(self, p):
"""namespace : KEYWORD ID NL
| KEYWORD ID NL INDENT docsection DEDENT"""
if p[1] == 'namespace':
doc = None
if len(p) > 4:
doc = p[5]
p[0] = AstNamespace(
self.path, p.lineno(1), p.lexpos(1... | [
"def",
"p_namespace",
"(",
"self",
",",
"p",
")",
":",
"if",
"p",
"[",
"1",
"]",
"==",
"'namespace'",
":",
"doc",
"=",
"None",
"if",
"len",
"(",
"p",
")",
">",
"4",
":",
"doc",
"=",
"p",
"[",
"5",
"]",
"p",
"[",
"0",
"]",
"=",
"AstNamespace... | namespace : KEYWORD ID NL
| KEYWORD ID NL INDENT docsection DEDENT | [
"namespace",
":",
"KEYWORD",
"ID",
"NL",
"|",
"KEYWORD",
"ID",
"NL",
"INDENT",
"docsection",
"DEDENT"
] | python | train | 36 |
MLAB-project/pymlab | src/pymlab/sensors/iic.py | https://github.com/MLAB-project/pymlab/blob/d18d858ae83b203defcf2aead0dbd11b3c444658/src/pymlab/sensors/iic.py#L171-L184 | def write_block_data(self, address, register, value):
"""
SMBus Block Write: i2c_smbus_write_block_data()
================================================
The opposite of the Block Read command, this writes up to 32 bytes to
a device, to a designated register that is specified ... | [
"def",
"write_block_data",
"(",
"self",
",",
"address",
",",
"register",
",",
"value",
")",
":",
"return",
"self",
".",
"smbus",
".",
"write_block_data",
"(",
"address",
",",
"register",
",",
"value",
")"
] | SMBus Block Write: i2c_smbus_write_block_data()
================================================
The opposite of the Block Read command, this writes up to 32 bytes to
a device, to a designated register that is specified through the
Comm byte. The amount of data is specified in the Coun... | [
"SMBus",
"Block",
"Write",
":",
"i2c_smbus_write_block_data",
"()",
"================================================"
] | python | train | 43.5 |
woolfson-group/isambard | isambard/optimisation/optimizer.py | https://github.com/woolfson-group/isambard/blob/ebc33b48a28ad217e18f93b910dfba46e6e71e07/isambard/optimisation/optimizer.py#L674-L697 | def generate(self):
"""Generates a particle using the creator function.
Notes
-----
Position and speed are uniformly randomly seeded within
allowed bounds. The particle also has speed limit settings
taken from global values.
Returns
-------
partic... | [
"def",
"generate",
"(",
"self",
")",
":",
"part",
"=",
"creator",
".",
"Particle",
"(",
"[",
"random",
".",
"uniform",
"(",
"-",
"1",
",",
"1",
")",
"for",
"_",
"in",
"range",
"(",
"len",
"(",
"self",
".",
"_params",
"[",
"'value_means'",
"]",
")... | Generates a particle using the creator function.
Notes
-----
Position and speed are uniformly randomly seeded within
allowed bounds. The particle also has speed limit settings
taken from global values.
Returns
-------
particle object | [
"Generates",
"a",
"particle",
"using",
"the",
"creator",
"function",
".",
"Notes",
"-----",
"Position",
"and",
"speed",
"are",
"uniformly",
"randomly",
"seeded",
"within",
"allowed",
"bounds",
".",
"The",
"particle",
"also",
"has",
"speed",
"limit",
"settings",
... | python | train | 33.916667 |
nschloe/matplotlib2tikz | matplotlib2tikz/save.py | https://github.com/nschloe/matplotlib2tikz/blob/ac5daca6f38b834d757f6c6ae6cc34121956f46b/matplotlib2tikz/save.py#L337-L415 | def _recurse(data, obj):
"""Iterates over all children of the current object, gathers the contents
contributing to the resulting PGFPlots file, and returns those.
"""
content = _ContentManager()
for child in obj.get_children():
# Some patches are Spines, too; skip those entirely.
# S... | [
"def",
"_recurse",
"(",
"data",
",",
"obj",
")",
":",
"content",
"=",
"_ContentManager",
"(",
")",
"for",
"child",
"in",
"obj",
".",
"get_children",
"(",
")",
":",
"# Some patches are Spines, too; skip those entirely.",
"# See <https://github.com/nschloe/matplotlib2tikz... | Iterates over all children of the current object, gathers the contents
contributing to the resulting PGFPlots file, and returns those. | [
"Iterates",
"over",
"all",
"children",
"of",
"the",
"current",
"object",
"gathers",
"the",
"contents",
"contributing",
"to",
"the",
"resulting",
"PGFPlots",
"file",
"and",
"returns",
"those",
"."
] | python | train | 45.075949 |
ladybug-tools/ladybug | ladybug/datatype/base.py | https://github.com/ladybug-tools/ladybug/blob/c08b7308077a48d5612f644943f92d5b5dade583/ladybug/datatype/base.py#L139-L172 | def is_in_range(self, values, unit=None, raise_exception=True):
"""Check if a list of values is within physically/mathematically possible range.
Args:
values: A list of values.
unit: The unit of the values. If not specified, the default metric
unit will be assum... | [
"def",
"is_in_range",
"(",
"self",
",",
"values",
",",
"unit",
"=",
"None",
",",
"raise_exception",
"=",
"True",
")",
":",
"self",
".",
"_is_numeric",
"(",
"values",
")",
"if",
"unit",
"is",
"None",
"or",
"unit",
"==",
"self",
".",
"units",
"[",
"0",... | Check if a list of values is within physically/mathematically possible range.
Args:
values: A list of values.
unit: The unit of the values. If not specified, the default metric
unit will be assumed.
raise_exception: Set to True to raise an exception if not i... | [
"Check",
"if",
"a",
"list",
"of",
"values",
"is",
"within",
"physically",
"/",
"mathematically",
"possible",
"range",
"."
] | python | train | 41.882353 |
Fizzadar/pyinfra | pyinfra/modules/server.py | https://github.com/Fizzadar/pyinfra/blob/006f751f7db2e07d32522c0285160783de2feb79/pyinfra/modules/server.py#L316-L453 | def user(
state, host, name,
present=True, home=None, shell=None, group=None, groups=None,
public_keys=None, delete_keys=False, ensure_home=True,
system=False, uid=None,
):
'''
Add/remove/update system users & their ssh `authorized_keys`.
+ name: name of the user to ensure
+ present: wh... | [
"def",
"user",
"(",
"state",
",",
"host",
",",
"name",
",",
"present",
"=",
"True",
",",
"home",
"=",
"None",
",",
"shell",
"=",
"None",
",",
"group",
"=",
"None",
",",
"groups",
"=",
"None",
",",
"public_keys",
"=",
"None",
",",
"delete_keys",
"="... | Add/remove/update system users & their ssh `authorized_keys`.
+ name: name of the user to ensure
+ present: whether this user should exist
+ home: the users home directory
+ shell: the users shell
+ group: the users primary group
+ groups: the users secondary groups
+ public_keys: list of p... | [
"Add",
"/",
"remove",
"/",
"update",
"system",
"users",
"&",
"their",
"ssh",
"authorized_keys",
"."
] | python | train | 28.050725 |
prometheus/client_python | prometheus_client/exposition.py | https://github.com/prometheus/client_python/blob/31f5557e2e84ca4ffa9a03abf6e3f4d0c8b8c3eb/prometheus_client/exposition.py#L165-L177 | def factory(cls, registry):
"""Returns a dynamic MetricsHandler class tied
to the passed registry.
"""
# This implementation relies on MetricsHandler.registry
# (defined above and defaulted to REGISTRY).
# As we have unicode_literals, we need to create a str()
... | [
"def",
"factory",
"(",
"cls",
",",
"registry",
")",
":",
"# This implementation relies on MetricsHandler.registry",
"# (defined above and defaulted to REGISTRY).",
"# As we have unicode_literals, we need to create a str()",
"# object for type().",
"cls_name",
"=",
"str",
"(",
"cls"... | Returns a dynamic MetricsHandler class tied
to the passed registry. | [
"Returns",
"a",
"dynamic",
"MetricsHandler",
"class",
"tied",
"to",
"the",
"passed",
"registry",
"."
] | python | train | 39.461538 |
twisted/mantissa | xmantissa/webapp.py | https://github.com/twisted/mantissa/blob/53e5502aba23ce99be78b27f923a276593033fe8/xmantissa/webapp.py#L50-L60 | def _reorderForPreference(themeList, preferredThemeName):
"""
Re-order the input themeList according to the preferred theme.
Returns None.
"""
for theme in themeList:
if preferredThemeName == theme.themeName:
themeList.remove(theme)
themeList.insert(0, theme)
... | [
"def",
"_reorderForPreference",
"(",
"themeList",
",",
"preferredThemeName",
")",
":",
"for",
"theme",
"in",
"themeList",
":",
"if",
"preferredThemeName",
"==",
"theme",
".",
"themeName",
":",
"themeList",
".",
"remove",
"(",
"theme",
")",
"themeList",
".",
"i... | Re-order the input themeList according to the preferred theme.
Returns None. | [
"Re",
"-",
"order",
"the",
"input",
"themeList",
"according",
"to",
"the",
"preferred",
"theme",
"."
] | python | train | 29.181818 |
RudolfCardinal/pythonlib | cardinal_pythonlib/tools/pdf_to_booklet.py | https://github.com/RudolfCardinal/pythonlib/blob/0b84cb35f38bd7d8723958dae51b480a829b7227/cardinal_pythonlib/tools/pdf_to_booklet.py#L161-L185 | def page_sequence(n_sheets: int, one_based: bool = True) -> List[int]:
"""
Generates the final page sequence from the starting number of sheets.
"""
n_pages = calc_n_virtual_pages(n_sheets)
assert n_pages % 4 == 0
half_n_pages = n_pages // 2
firsthalf = list(range(half_n_pages))
secondha... | [
"def",
"page_sequence",
"(",
"n_sheets",
":",
"int",
",",
"one_based",
":",
"bool",
"=",
"True",
")",
"->",
"List",
"[",
"int",
"]",
":",
"n_pages",
"=",
"calc_n_virtual_pages",
"(",
"n_sheets",
")",
"assert",
"n_pages",
"%",
"4",
"==",
"0",
"half_n_page... | Generates the final page sequence from the starting number of sheets. | [
"Generates",
"the",
"final",
"page",
"sequence",
"from",
"the",
"starting",
"number",
"of",
"sheets",
"."
] | python | train | 42.36 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.