Unnamed: 0 int64 0 10k | repository_name stringlengths 7 54 | func_path_in_repository stringlengths 5 223 | func_name stringlengths 1 134 | whole_func_string stringlengths 100 30.3k | language stringclasses 1
value | func_code_string stringlengths 100 30.3k | func_code_tokens stringlengths 138 33.2k | func_documentation_string stringlengths 1 15k | func_documentation_tokens stringlengths 5 5.14k | split_name stringclasses 1
value | func_code_url stringlengths 91 315 |
|---|---|---|---|---|---|---|---|---|---|---|---|
0 | getsentry/libsourcemap | libsourcemap/highlevel.py | View.get_original_function_name | def get_original_function_name(self, line, col, minified_name,
minified_source):
"""Given a token location and a minified function name and the
minified source file this returns the original function name if it
can be found of the minified function in scope.
... | python | def get_original_function_name(self, line, col, minified_name,
minified_source):
"""Given a token location and a minified function name and the
minified source file this returns the original function name if it
can be found of the minified function in scope.
... | ['def', 'get_original_function_name', '(', 'self', ',', 'line', ',', 'col', ',', 'minified_name', ',', 'minified_source', ')', ':', '# Silently ignore underflows', 'if', 'line', '<', '0', 'or', 'col', '<', '0', ':', 'return', 'None', 'minified_name', '=', 'minified_name', '.', 'encode', '(', "'utf-8'", ')', 'sout', '='... | Given a token location and a minified function name and the
minified source file this returns the original function name if it
can be found of the minified function in scope. | ['Given', 'a', 'token', 'location', 'and', 'a', 'minified', 'function', 'name', 'and', 'the', 'minified', 'source', 'file', 'this', 'returns', 'the', 'original', 'function', 'name', 'if', 'it', 'can', 'be', 'found', 'of', 'the', 'minified', 'function', 'in', 'scope', '.'] | train | https://github.com/getsentry/libsourcemap/blob/94b5a34814fafee9dc23da8ec0ccca77f30e3370/libsourcemap/highlevel.py#L163-L185 |
1 | brocade/pynos | pynos/versions/ver_7/ver_7_1_0/yang/brocade_mac_address_table.py | brocade_mac_address_table.get_mac_address_table_input_request_type_get_interface_based_request_mac_type | def get_mac_address_table_input_request_type_get_interface_based_request_mac_type(self, **kwargs):
"""Auto Generated Code
"""
config = ET.Element("config")
get_mac_address_table = ET.Element("get_mac_address_table")
config = get_mac_address_table
input = ET.SubElement(get... | python | def get_mac_address_table_input_request_type_get_interface_based_request_mac_type(self, **kwargs):
"""Auto Generated Code
"""
config = ET.Element("config")
get_mac_address_table = ET.Element("get_mac_address_table")
config = get_mac_address_table
input = ET.SubElement(get... | ['def', 'get_mac_address_table_input_request_type_get_interface_based_request_mac_type', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'config', '=', 'ET', '.', 'Element', '(', '"config"', ')', 'get_mac_address_table', '=', 'ET', '.', 'Element', '(', '"get_mac_address_table"', ')', 'config', '=', 'get_mac_address_tab... | Auto Generated Code | ['Auto', 'Generated', 'Code'] | train | https://github.com/brocade/pynos/blob/bd8a34e98f322de3fc06750827d8bbc3a0c00380/pynos/versions/ver_7/ver_7_1_0/yang/brocade_mac_address_table.py#L297-L310 |
2 | nicolargo/glances | glances/plugins/glances_memswap.py | Plugin.update_views | def update_views(self):
"""Update stats views."""
# Call the father's method
super(Plugin, self).update_views()
# Add specifics informations
# Alert and log
self.views['used']['decoration'] = self.get_alert_log(self.stats['used'], maximum=self.stats['total']) | python | def update_views(self):
"""Update stats views."""
# Call the father's method
super(Plugin, self).update_views()
# Add specifics informations
# Alert and log
self.views['used']['decoration'] = self.get_alert_log(self.stats['used'], maximum=self.stats['total']) | ['def', 'update_views', '(', 'self', ')', ':', "# Call the father's method", 'super', '(', 'Plugin', ',', 'self', ')', '.', 'update_views', '(', ')', '# Add specifics informations', '# Alert and log', 'self', '.', 'views', '[', "'used'", ']', '[', "'decoration'", ']', '=', 'self', '.', 'get_alert_log', '(', 'self', '.'... | Update stats views. | ['Update', 'stats', 'views', '.'] | train | https://github.com/nicolargo/glances/blob/5bd4d587a736e0d2b03170b56926841d2a3eb7ee/glances/plugins/glances_memswap.py#L130-L137 |
3 | DLR-RM/RAFCON | source/rafcon/gui/helpers/meta_data.py | contains_geometric_info | def contains_geometric_info(var):
""" Check whether the passed variable is a tuple with two floats or integers """
return isinstance(var, tuple) and len(var) == 2 and all(isinstance(val, (int, float)) for val in var) | python | def contains_geometric_info(var):
""" Check whether the passed variable is a tuple with two floats or integers """
return isinstance(var, tuple) and len(var) == 2 and all(isinstance(val, (int, float)) for val in var) | ['def', 'contains_geometric_info', '(', 'var', ')', ':', 'return', 'isinstance', '(', 'var', ',', 'tuple', ')', 'and', 'len', '(', 'var', ')', '==', '2', 'and', 'all', '(', 'isinstance', '(', 'val', ',', '(', 'int', ',', 'float', ')', ')', 'for', 'val', 'in', 'var', ')'] | Check whether the passed variable is a tuple with two floats or integers | ['Check', 'whether', 'the', 'passed', 'variable', 'is', 'a', 'tuple', 'with', 'two', 'floats', 'or', 'integers'] | train | https://github.com/DLR-RM/RAFCON/blob/24942ef1a904531f49ab8830a1dbb604441be498/source/rafcon/gui/helpers/meta_data.py#L55-L57 |
4 | ejeschke/ginga | ginga/gtk3w/ImageViewGtk.py | ImageViewGtk.save_plain_image_as_file | def save_plain_image_as_file(self, filepath, format='png', quality=90):
"""Used for generating thumbnails. Does not include overlaid
graphics.
"""
pixbuf = self.get_plain_image_as_pixbuf()
options, values = [], []
if format == 'jpeg':
options.append('quality'... | python | def save_plain_image_as_file(self, filepath, format='png', quality=90):
"""Used for generating thumbnails. Does not include overlaid
graphics.
"""
pixbuf = self.get_plain_image_as_pixbuf()
options, values = [], []
if format == 'jpeg':
options.append('quality'... | ['def', 'save_plain_image_as_file', '(', 'self', ',', 'filepath', ',', 'format', '=', "'png'", ',', 'quality', '=', '90', ')', ':', 'pixbuf', '=', 'self', '.', 'get_plain_image_as_pixbuf', '(', ')', 'options', ',', 'values', '=', '[', ']', ',', '[', ']', 'if', 'format', '==', "'jpeg'", ':', 'options', '.', 'append', '(... | Used for generating thumbnails. Does not include overlaid
graphics. | ['Used', 'for', 'generating', 'thumbnails', '.', 'Does', 'not', 'include', 'overlaid', 'graphics', '.'] | train | https://github.com/ejeschke/ginga/blob/a78c893ec6f37a837de851947e9bb4625c597915/ginga/gtk3w/ImageViewGtk.py#L75-L84 |
5 | poppy-project/pypot | pypot/vrep/remoteApiBindings/vrep.py | simxClearFloatSignal | def simxClearFloatSignal(clientID, signalName, operationMode):
'''
Please have a look at the function description/documentation in the V-REP user manual
'''
if (sys.version_info[0] == 3) and (type(signalName) is str):
signalName=signalName.encode('utf-8')
return c_ClearFloatSignal(clientID,... | python | def simxClearFloatSignal(clientID, signalName, operationMode):
'''
Please have a look at the function description/documentation in the V-REP user manual
'''
if (sys.version_info[0] == 3) and (type(signalName) is str):
signalName=signalName.encode('utf-8')
return c_ClearFloatSignal(clientID,... | ['def', 'simxClearFloatSignal', '(', 'clientID', ',', 'signalName', ',', 'operationMode', ')', ':', 'if', '(', 'sys', '.', 'version_info', '[', '0', ']', '==', '3', ')', 'and', '(', 'type', '(', 'signalName', ')', 'is', 'str', ')', ':', 'signalName', '=', 'signalName', '.', 'encode', '(', "'utf-8'", ')', 'return', 'c_C... | Please have a look at the function description/documentation in the V-REP user manual | ['Please', 'have', 'a', 'look', 'at', 'the', 'function', 'description', '/', 'documentation', 'in', 'the', 'V', '-', 'REP', 'user', 'manual'] | train | https://github.com/poppy-project/pypot/blob/d9c6551bbc87d45d9d1f0bc15e35b616d0002afd/pypot/vrep/remoteApiBindings/vrep.py#L900-L907 |
6 | cloudify-cosmo/repex | repex.py | Repex.find_matches | def find_matches(self, content, file_to_handle):
"""Find all matches of an expression in a file
"""
# look for all match groups in the content
groups = [match.groupdict() for match in
self.match_expression.finditer(content)]
# filter out content not in the match... | python | def find_matches(self, content, file_to_handle):
"""Find all matches of an expression in a file
"""
# look for all match groups in the content
groups = [match.groupdict() for match in
self.match_expression.finditer(content)]
# filter out content not in the match... | ['def', 'find_matches', '(', 'self', ',', 'content', ',', 'file_to_handle', ')', ':', '# look for all match groups in the content', 'groups', '=', '[', 'match', '.', 'groupdict', '(', ')', 'for', 'match', 'in', 'self', '.', 'match_expression', '.', 'finditer', '(', 'content', ')', ']', '# filter out content not in the ... | Find all matches of an expression in a file | ['Find', 'all', 'matches', 'of', 'an', 'expression', 'in', 'a', 'file'] | train | https://github.com/cloudify-cosmo/repex/blob/589e442857fa4a99fa88670d7df1a72f983bbd28/repex.py#L605-L618 |
7 | mariocj89/github-token | github_token/__init__.py | TokenFactory.create | def create(self):
"""Creates a token
It uses the app_name as the notes and the scopes are
the permissions required by the application. See those
in github when configuring an app token
Raises a TFARequired if a two factor is required after
the atempt to create it withou... | python | def create(self):
"""Creates a token
It uses the app_name as the notes and the scopes are
the permissions required by the application. See those
in github when configuring an app token
Raises a TFARequired if a two factor is required after
the atempt to create it withou... | ['def', 'create', '(', 'self', ')', ':', 'headers', '=', 'dict', '(', ')', 'if', 'self', '.', 'tfa_token', ':', 'headers', '[', '"X-GitHub-OTP"', ']', '=', 'self', '.', 'tfa_token', 'token_name', '=', 'self', '.', 'app_name', '+', 'platform', '.', 'node', '(', ')', '# node specific in case the user has multiple hosts',... | Creates a token
It uses the app_name as the notes and the scopes are
the permissions required by the application. See those
in github when configuring an app token
Raises a TFARequired if a two factor is required after
the atempt to create it without having call tfa before | ['Creates', 'a', 'token'] | train | https://github.com/mariocj89/github-token/blob/8ca85fa51a52aef94cfb4f851eb229ee500bc28f/github_token/__init__.py#L81-L108 |
8 | brycedrennan/eulerian-magnification | eulerian_magnification/base.py | combine_pyramid_and_save | def combine_pyramid_and_save(g_video, orig_video, enlarge_multiple, fps, save_filename='media/output.avi'):
"""Combine a gaussian video representation with the original and save to file"""
width, height = get_frame_dimensions(orig_video[0])
fourcc = cv2.VideoWriter_fourcc(*'MJPG')
print("Outputting to %... | python | def combine_pyramid_and_save(g_video, orig_video, enlarge_multiple, fps, save_filename='media/output.avi'):
"""Combine a gaussian video representation with the original and save to file"""
width, height = get_frame_dimensions(orig_video[0])
fourcc = cv2.VideoWriter_fourcc(*'MJPG')
print("Outputting to %... | ['def', 'combine_pyramid_and_save', '(', 'g_video', ',', 'orig_video', ',', 'enlarge_multiple', ',', 'fps', ',', 'save_filename', '=', "'media/output.avi'", ')', ':', 'width', ',', 'height', '=', 'get_frame_dimensions', '(', 'orig_video', '[', '0', ']', ')', 'fourcc', '=', 'cv2', '.', 'VideoWriter_fourcc', '(', '*', "'... | Combine a gaussian video representation with the original and save to file | ['Combine', 'a', 'gaussian', 'video', 'representation', 'with', 'the', 'original', 'and', 'save', 'to', 'file'] | train | https://github.com/brycedrennan/eulerian-magnification/blob/9ae0651fe3334176300d183f8240ad36d77759a9/eulerian_magnification/base.py#L108-L122 |
9 | PmagPy/PmagPy | pmagpy/pmag.py | PintPars | def PintPars(datablock, araiblock, zijdblock, start, end, accept, **kwargs):
"""
calculate the paleointensity magic parameters make some definitions
"""
if 'version' in list(kwargs.keys()) and kwargs['version'] == 3:
meth_key = 'method_codes'
beta_key = 'int_b_beta'
temp_key, m... | python | def PintPars(datablock, araiblock, zijdblock, start, end, accept, **kwargs):
"""
calculate the paleointensity magic parameters make some definitions
"""
if 'version' in list(kwargs.keys()) and kwargs['version'] == 3:
meth_key = 'method_codes'
beta_key = 'int_b_beta'
temp_key, m... | ['def', 'PintPars', '(', 'datablock', ',', 'araiblock', ',', 'zijdblock', ',', 'start', ',', 'end', ',', 'accept', ',', '*', '*', 'kwargs', ')', ':', 'if', "'version'", 'in', 'list', '(', 'kwargs', '.', 'keys', '(', ')', ')', 'and', 'kwargs', '[', "'version'", ']', '==', '3', ':', 'meth_key', '=', "'method_codes'", 'be... | calculate the paleointensity magic parameters make some definitions | ['calculate', 'the', 'paleointensity', 'magic', 'parameters', 'make', 'some', 'definitions'] | train | https://github.com/PmagPy/PmagPy/blob/c7984f8809bf40fe112e53dcc311a33293b62d0b/pmagpy/pmag.py#L2827-L3374 |
10 | lowandrew/OLCTools | spadespipeline/vtyper.py | Vtyper.epcrparse | def epcrparse(self):
"""
Parse the ePCR text file outputs
"""
logging.info('Parsing ePCR results')
for sample in self.metadata:
if sample.general.bestassemblyfile != 'NA':
if 'stx' in sample.general.datastore:
# Initialise count - t... | python | def epcrparse(self):
"""
Parse the ePCR text file outputs
"""
logging.info('Parsing ePCR results')
for sample in self.metadata:
if sample.general.bestassemblyfile != 'NA':
if 'stx' in sample.general.datastore:
# Initialise count - t... | ['def', 'epcrparse', '(', 'self', ')', ':', 'logging', '.', 'info', '(', "'Parsing ePCR results'", ')', 'for', 'sample', 'in', 'self', '.', 'metadata', ':', 'if', 'sample', '.', 'general', '.', 'bestassemblyfile', '!=', "'NA'", ':', 'if', "'stx'", 'in', 'sample', '.', 'general', '.', 'datastore', ':', '# Initialise cou... | Parse the ePCR text file outputs | ['Parse', 'the', 'ePCR', 'text', 'file', 'outputs'] | train | https://github.com/lowandrew/OLCTools/blob/88aa90ac85f84d0bbeb03e43c29b0a9d36e4ce2a/spadespipeline/vtyper.py#L96-L131 |
11 | atztogo/phonopy | phonopy/structure/spglib.py | get_pointgroup | def get_pointgroup(rotations):
"""Return point group in international table symbol and number.
The symbols are mapped to the numbers as follows:
1 "1 "
2 "-1 "
3 "2 "
4 "m "
5 "2/m "
6 "222 "
7 "mm2 "
8 "mmm "
9 "4 "
10 "-4 "
11... | python | def get_pointgroup(rotations):
"""Return point group in international table symbol and number.
The symbols are mapped to the numbers as follows:
1 "1 "
2 "-1 "
3 "2 "
4 "m "
5 "2/m "
6 "222 "
7 "mm2 "
8 "mmm "
9 "4 "
10 "-4 "
11... | ['def', 'get_pointgroup', '(', 'rotations', ')', ':', '_set_no_error', '(', ')', '# (symbol, pointgroup_number, transformation_matrix)', 'pointgroup', '=', 'spg', '.', 'pointgroup', '(', 'np', '.', 'array', '(', 'rotations', ',', 'dtype', '=', "'intc'", ',', 'order', '=', "'C'", ')', ')', '_set_error_message', '(', ')'... | Return point group in international table symbol and number.
The symbols are mapped to the numbers as follows:
1 "1 "
2 "-1 "
3 "2 "
4 "m "
5 "2/m "
6 "222 "
7 "mm2 "
8 "mmm "
9 "4 "
10 "-4 "
11 "4/m "
12 "422 "
13 "4mm... | ['Return', 'point', 'group', 'in', 'international', 'table', 'symbol', 'and', 'number', '.'] | train | https://github.com/atztogo/phonopy/blob/869cc2ba9e7d495d5f4cf6942415ab3fc9e2a10f/phonopy/structure/spglib.py#L301-L343 |
12 | automl/HpBandSter | hpbandster/core/nameserver.py | NameServer.start | def start(self):
"""
starts a Pyro4 nameserver in a separate thread
Returns
-------
tuple (str, int):
the host name and the used port
"""
if self.host is None:
if self.nic_name is None:
self.host = 'localhost'
else:
self.host = nic_name_to_host(self.nic_name)
uri, self.pyro_ns,... | python | def start(self):
"""
starts a Pyro4 nameserver in a separate thread
Returns
-------
tuple (str, int):
the host name and the used port
"""
if self.host is None:
if self.nic_name is None:
self.host = 'localhost'
else:
self.host = nic_name_to_host(self.nic_name)
uri, self.pyro_ns,... | ['def', 'start', '(', 'self', ')', ':', 'if', 'self', '.', 'host', 'is', 'None', ':', 'if', 'self', '.', 'nic_name', 'is', 'None', ':', 'self', '.', 'host', '=', "'localhost'", 'else', ':', 'self', '.', 'host', '=', 'nic_name_to_host', '(', 'self', '.', 'nic_name', ')', 'uri', ',', 'self', '.', 'pyro_ns', ',', '_', '='... | starts a Pyro4 nameserver in a separate thread
Returns
-------
tuple (str, int):
the host name and the used port | ['starts', 'a', 'Pyro4', 'nameserver', 'in', 'a', 'separate', 'thread', 'Returns', '-------', 'tuple', '(', 'str', 'int', ')', ':', 'the', 'host', 'name', 'and', 'the', 'used', 'port'] | train | https://github.com/automl/HpBandSter/blob/841db4b827f342e5eb7f725723ea6461ac52d45a/hpbandster/core/nameserver.py#L48-L79 |
13 | PixelwarStudio/PyTree | Tree/core.py | Tree.grow | def grow(self, times=1):
"""Let the tree grow.
Args:
times (integer): Indicate how many times the tree will grow.
"""
self.nodes.append([])
for n, node in enumerate(self.nodes[self.age]):
if self.age == 0:
p_node = Node(self.pos[:2])
... | python | def grow(self, times=1):
"""Let the tree grow.
Args:
times (integer): Indicate how many times the tree will grow.
"""
self.nodes.append([])
for n, node in enumerate(self.nodes[self.age]):
if self.age == 0:
p_node = Node(self.pos[:2])
... | ['def', 'grow', '(', 'self', ',', 'times', '=', '1', ')', ':', 'self', '.', 'nodes', '.', 'append', '(', '[', ']', ')', 'for', 'n', ',', 'node', 'in', 'enumerate', '(', 'self', '.', 'nodes', '[', 'self', '.', 'age', ']', ')', ':', 'if', 'self', '.', 'age', '==', '0', ':', 'p_node', '=', 'Node', '(', 'self', '.', 'pos',... | Let the tree grow.
Args:
times (integer): Indicate how many times the tree will grow. | ['Let', 'the', 'tree', 'grow', '.'] | train | https://github.com/PixelwarStudio/PyTree/blob/f14b25ea145da6b00d836e34251d2a4c823766dc/Tree/core.py#L167-L189 |
14 | saltstack/salt | salt/modules/osquery.py | kernel_integrity | def kernel_integrity(attrs=None, where=None):
'''
Return kernel_integrity information from osquery
CLI Example:
.. code-block:: bash
salt '*' osquery.kernel_integrity
'''
if __grains__['os_family'] in ['RedHat', 'Debian']:
return _osquery_cmd(table='kernel_integrity', attrs=at... | python | def kernel_integrity(attrs=None, where=None):
'''
Return kernel_integrity information from osquery
CLI Example:
.. code-block:: bash
salt '*' osquery.kernel_integrity
'''
if __grains__['os_family'] in ['RedHat', 'Debian']:
return _osquery_cmd(table='kernel_integrity', attrs=at... | ['def', 'kernel_integrity', '(', 'attrs', '=', 'None', ',', 'where', '=', 'None', ')', ':', 'if', '__grains__', '[', "'os_family'", ']', 'in', '[', "'RedHat'", ',', "'Debian'", ']', ':', 'return', '_osquery_cmd', '(', 'table', '=', "'kernel_integrity'", ',', 'attrs', '=', 'attrs', ',', 'where', '=', 'where', ')', 'retu... | Return kernel_integrity information from osquery
CLI Example:
.. code-block:: bash
salt '*' osquery.kernel_integrity | ['Return', 'kernel_integrity', 'information', 'from', 'osquery'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/osquery.py#L149-L161 |
15 | rackerlabs/lambda-uploader | lambda_uploader/package.py | Package.virtualenv | def virtualenv(self, virtualenv):
'''
Sets the virtual environment for the lambda package
If this is not set then package_dependencies will create a new one.
Takes a path to a virtualenv or a boolean if the virtualenv creation
should be skipped.
'''
# If a boole... | python | def virtualenv(self, virtualenv):
'''
Sets the virtual environment for the lambda package
If this is not set then package_dependencies will create a new one.
Takes a path to a virtualenv or a boolean if the virtualenv creation
should be skipped.
'''
# If a boole... | ['def', 'virtualenv', '(', 'self', ',', 'virtualenv', ')', ':', '# If a boolean is passed then set the internal _skip_virtualenv flag', 'if', 'isinstance', '(', 'virtualenv', ',', 'bool', ')', ':', 'self', '.', '_skip_virtualenv', '=', 'virtualenv', 'else', ':', 'self', '.', '_virtualenv', '=', 'virtualenv', 'if', 'not... | Sets the virtual environment for the lambda package
If this is not set then package_dependencies will create a new one.
Takes a path to a virtualenv or a boolean if the virtualenv creation
should be skipped. | ['Sets', 'the', 'virtual', 'environment', 'for', 'the', 'lambda', 'package'] | train | https://github.com/rackerlabs/lambda-uploader/blob/a5036e60d45d1a4fdc07df071f5b6e3b113388d4/lambda_uploader/package.py#L114-L133 |
16 | agoragames/kairos | kairos/cassandra_backend.py | CassandraSet._insert_stmt | def _insert_stmt(self, name, value, timestamp, interval, config):
'''Helper to generate the insert statement.'''
# Calculate the TTL and abort if inserting into the past
expire, ttl = config['expire'], config['ttl'](timestamp)
if expire and not ttl:
return None
i_time = config['i_calc'].to_bu... | python | def _insert_stmt(self, name, value, timestamp, interval, config):
'''Helper to generate the insert statement.'''
# Calculate the TTL and abort if inserting into the past
expire, ttl = config['expire'], config['ttl'](timestamp)
if expire and not ttl:
return None
i_time = config['i_calc'].to_bu... | ['def', '_insert_stmt', '(', 'self', ',', 'name', ',', 'value', ',', 'timestamp', ',', 'interval', ',', 'config', ')', ':', '# Calculate the TTL and abort if inserting into the past', 'expire', ',', 'ttl', '=', 'config', '[', "'expire'", ']', ',', 'config', '[', "'ttl'", ']', '(', 'timestamp', ')', 'if', 'expire', 'and... | Helper to generate the insert statement. | ['Helper', 'to', 'generate', 'the', 'insert', 'statement', '.'] | train | https://github.com/agoragames/kairos/blob/0b062d543b0f4a46df460fa0eb6ec281232ab179/kairos/cassandra_backend.py#L646-L665 |
17 | pavelsof/ipalint | ipalint/read.py | Reader._determine_dialect | def _determine_dialect(self, lines):
"""
Expects a non-empty [] of strings; these would normally be the first
few lines of a csv file. Returns the most likely Dialect named tuple or
None if the data seems to form a single column.
Ensures that using the returned dialect, all the lines given will have
the sa... | python | def _determine_dialect(self, lines):
"""
Expects a non-empty [] of strings; these would normally be the first
few lines of a csv file. Returns the most likely Dialect named tuple or
None if the data seems to form a single column.
Ensures that using the returned dialect, all the lines given will have
the sa... | ['def', '_determine_dialect', '(', 'self', ',', 'lines', ')', ':', 'permuts', '=', '[', '(', 'quotechar', ',', 'escapechar', ')', 'for', 'quotechar', 'in', 'CSV_QUOTECHARS', 'for', 'escapechar', 'in', 'CSV_ESCAPECHARS', ']', 'for', 'delim', 'in', 'CSV_DELIMITERS', ':', 'counts', '=', '[', 'line', '.', 'count', '(', 'de... | Expects a non-empty [] of strings; these would normally be the first
few lines of a csv file. Returns the most likely Dialect named tuple or
None if the data seems to form a single column.
Ensures that using the returned dialect, all the lines given will have
the same number of columns.
Helper for the get_d... | ['Expects', 'a', 'non', '-', 'empty', '[]', 'of', 'strings', ';', 'these', 'would', 'normally', 'be', 'the', 'first', 'few', 'lines', 'of', 'a', 'csv', 'file', '.', 'Returns', 'the', 'most', 'likely', 'Dialect', 'named', 'tuple', 'or', 'None', 'if', 'the', 'data', 'seems', 'to', 'form', 'a', 'single', 'column', '.'] | train | https://github.com/pavelsof/ipalint/blob/763e5979ede6980cbfc746b06fd007b379762eeb/ipalint/read.py#L177-L218 |
18 | angr/angr | angr/utils/graph.py | compute_dominance_frontier | def compute_dominance_frontier(graph, domtree):
"""
Compute a dominance frontier based on the given post-dominator tree.
This implementation is based on figure 2 of paper An Efficient Method of Computing Static Single Assignment
Form by Ron Cytron, etc.
:param graph: The graph where we want to c... | python | def compute_dominance_frontier(graph, domtree):
"""
Compute a dominance frontier based on the given post-dominator tree.
This implementation is based on figure 2 of paper An Efficient Method of Computing Static Single Assignment
Form by Ron Cytron, etc.
:param graph: The graph where we want to c... | ['def', 'compute_dominance_frontier', '(', 'graph', ',', 'domtree', ')', ':', 'df', '=', '{', '}', '# Perform a post-order search on the dominator tree', 'for', 'x', 'in', 'networkx', '.', 'dfs_postorder_nodes', '(', 'domtree', ')', ':', 'if', 'x', 'not', 'in', 'graph', ':', '# Skip nodes that are not in the graph', 'c... | Compute a dominance frontier based on the given post-dominator tree.
This implementation is based on figure 2 of paper An Efficient Method of Computing Static Single Assignment
Form by Ron Cytron, etc.
:param graph: The graph where we want to compute the dominance frontier.
:param domtree: The domin... | ['Compute', 'a', 'dominance', 'frontier', 'based', 'on', 'the', 'given', 'post', '-', 'dominator', 'tree', '.'] | train | https://github.com/angr/angr/blob/4e2f97d56af5419ee73bdb30482c8dd8ff5f3e40/angr/utils/graph.py#L63-L104 |
19 | CI-WATER/mapkit | mapkit/RasterConverter.py | RasterConverter.isNumber | def isNumber(self, value):
"""
Validate whether a value is a number or not
"""
try:
str(value)
float(value)
return True
except ValueError:
return False | python | def isNumber(self, value):
"""
Validate whether a value is a number or not
"""
try:
str(value)
float(value)
return True
except ValueError:
return False | ['def', 'isNumber', '(', 'self', ',', 'value', ')', ':', 'try', ':', 'str', '(', 'value', ')', 'float', '(', 'value', ')', 'return', 'True', 'except', 'ValueError', ':', 'return', 'False'] | Validate whether a value is a number or not | ['Validate', 'whether', 'a', 'value', 'is', 'a', 'number', 'or', 'not'] | train | https://github.com/CI-WATER/mapkit/blob/ce5fbded6af7adabdf1eec85631c6811ef8ecc34/mapkit/RasterConverter.py#L1097-L1107 |
20 | CamDavidsonPilon/lifelines | lifelines/utils/__init__.py | survival_table_from_events | def survival_table_from_events(
death_times,
event_observed,
birth_times=None,
columns=["removed", "observed", "censored", "entrance", "at_risk"],
weights=None,
collapse=False,
intervals=None,
): # pylint: disable=dangerous-default-value,too-many-locals
"""
Parameters
----------... | python | def survival_table_from_events(
death_times,
event_observed,
birth_times=None,
columns=["removed", "observed", "censored", "entrance", "at_risk"],
weights=None,
collapse=False,
intervals=None,
): # pylint: disable=dangerous-default-value,too-many-locals
"""
Parameters
----------... | ['def', 'survival_table_from_events', '(', 'death_times', ',', 'event_observed', ',', 'birth_times', '=', 'None', ',', 'columns', '=', '[', '"removed"', ',', '"observed"', ',', '"censored"', ',', '"entrance"', ',', '"at_risk"', ']', ',', 'weights', '=', 'None', ',', 'collapse', '=', 'False', ',', 'intervals', '=', 'Non... | Parameters
----------
death_times: (n,) array
represent the event times
event_observed: (n,) array
1 if observed event, 0 is censored event.
birth_times: a (n,) array, optional
representing when the subject was first observed. A subject's death event is then at [birth times + duration ... | ['Parameters', '----------', 'death_times', ':', '(', 'n', ')', 'array', 'represent', 'the', 'event', 'times', 'event_observed', ':', '(', 'n', ')', 'array', '1', 'if', 'observed', 'event', '0', 'is', 'censored', 'event', '.', 'birth_times', ':', 'a', '(', 'n', ')', 'array', 'optional', 'representing', 'when', 'the', '... | train | https://github.com/CamDavidsonPilon/lifelines/blob/bdf6be6f1d10eea4c46365ee0ee6a47d8c30edf8/lifelines/utils/__init__.py#L262-L361 |
21 | django-blog-zinnia/cmsplugin-zinnia | cmsplugin_zinnia/cms_plugins.py | CMSRandomEntriesPlugin.render | def render(self, context, instance, placeholder):
"""
Update the context with plugin's data
"""
context = super(CMSRandomEntriesPlugin, self).render(
context, instance, placeholder)
context['template_to_render'] = (str(instance.template_to_render) or
... | python | def render(self, context, instance, placeholder):
"""
Update the context with plugin's data
"""
context = super(CMSRandomEntriesPlugin, self).render(
context, instance, placeholder)
context['template_to_render'] = (str(instance.template_to_render) or
... | ['def', 'render', '(', 'self', ',', 'context', ',', 'instance', ',', 'placeholder', ')', ':', 'context', '=', 'super', '(', 'CMSRandomEntriesPlugin', ',', 'self', ')', '.', 'render', '(', 'context', ',', 'instance', ',', 'placeholder', ')', 'context', '[', "'template_to_render'", ']', '=', '(', 'str', '(', 'instance', ... | Update the context with plugin's data | ['Update', 'the', 'context', 'with', 'plugin', 's', 'data'] | train | https://github.com/django-blog-zinnia/cmsplugin-zinnia/blob/7613c0d9ae29affe9ab97527e4b6d5bef124afdc/cmsplugin_zinnia/cms_plugins.py#L131-L139 |
22 | rjw57/throw | throw/minus/minus.py | Gallery.SaveGallery | def SaveGallery(self, name=None, items=None):
"""Use this to update the gallery name or change sort order.
Specify which attribute (name or items or both) you want to change."""
url = 'http://min.us/api/SaveGallery'
if not name:
if not self.name:
name = self... | python | def SaveGallery(self, name=None, items=None):
"""Use this to update the gallery name or change sort order.
Specify which attribute (name or items or both) you want to change."""
url = 'http://min.us/api/SaveGallery'
if not name:
if not self.name:
name = self... | ['def', 'SaveGallery', '(', 'self', ',', 'name', '=', 'None', ',', 'items', '=', 'None', ')', ':', 'url', '=', "'http://min.us/api/SaveGallery'", 'if', 'not', 'name', ':', 'if', 'not', 'self', '.', 'name', ':', 'name', '=', 'self', '.', 'GetItems', '(', ')', '[', '0', ']', 'if', 'self', '.', 'name', ':', 'name', '=', '... | Use this to update the gallery name or change sort order.
Specify which attribute (name or items or both) you want to change. | ['Use', 'this', 'to', 'update', 'the', 'gallery', 'name', 'or', 'change', 'sort', 'order', '.', 'Specify', 'which', 'attribute', '(', 'name', 'or', 'items', 'or', 'both', ')', 'you', 'want', 'to', 'change', '.'] | train | https://github.com/rjw57/throw/blob/74a7116362ba5b45635ab247472b25cfbdece4ee/throw/minus/minus.py#L62-L88 |
23 | letuananh/chirptext | chirptext/cli.py | setup_logging | def setup_logging(filename, log_dir=None, force_setup=False):
''' Try to load logging configuration from a file. Set level to INFO if failed.
'''
if not force_setup and ChirpCLI.SETUP_COMPLETED:
logging.debug("Master logging has been setup. This call will be ignored.")
return
if log_dir ... | python | def setup_logging(filename, log_dir=None, force_setup=False):
''' Try to load logging configuration from a file. Set level to INFO if failed.
'''
if not force_setup and ChirpCLI.SETUP_COMPLETED:
logging.debug("Master logging has been setup. This call will be ignored.")
return
if log_dir ... | ['def', 'setup_logging', '(', 'filename', ',', 'log_dir', '=', 'None', ',', 'force_setup', '=', 'False', ')', ':', 'if', 'not', 'force_setup', 'and', 'ChirpCLI', '.', 'SETUP_COMPLETED', ':', 'logging', '.', 'debug', '(', '"Master logging has been setup. This call will be ignored."', ')', 'return', 'if', 'log_dir', 'and... | Try to load logging configuration from a file. Set level to INFO if failed. | ['Try', 'to', 'load', 'logging', 'configuration', 'from', 'a', 'file', '.', 'Set', 'level', 'to', 'INFO', 'if', 'failed', '.'] | train | https://github.com/letuananh/chirptext/blob/ce60b47257b272a587c8703ea1f86cd1a45553a7/chirptext/cli.py#L35-L55 |
24 | doconix/django-mako-plus | django_mako_plus/template/adapter.py | MakoTemplateAdapter.name | def name(self):
'''Returns the name of this template (if created from a file) or "string" if not'''
if self.mako_template.filename:
return os.path.basename(self.mako_template.filename)
return 'string' | python | def name(self):
'''Returns the name of this template (if created from a file) or "string" if not'''
if self.mako_template.filename:
return os.path.basename(self.mako_template.filename)
return 'string' | ['def', 'name', '(', 'self', ')', ':', 'if', 'self', '.', 'mako_template', '.', 'filename', ':', 'return', 'os', '.', 'path', '.', 'basename', '(', 'self', '.', 'mako_template', '.', 'filename', ')', 'return', "'string'"] | Returns the name of this template (if created from a file) or "string" if not | ['Returns', 'the', 'name', 'of', 'this', 'template', '(', 'if', 'created', 'from', 'a', 'file', ')', 'or', 'string', 'if', 'not'] | train | https://github.com/doconix/django-mako-plus/blob/a90f9b4af19e5fa9f83452989cdcaed21569a181/django_mako_plus/template/adapter.py#L39-L43 |
25 | HazyResearch/fonduer | src/fonduer/learning/disc_models/sparse_lstm.py | SparseLSTM.forward | def forward(self, X):
"""Forward function.
:param X: The input (batch) of the model contains word sequences for lstm,
features and feature weights.
:type X: For word sequences: a list of torch.Tensor pair (word sequence
and word mask) of shape (batch_size, sequence_lengt... | python | def forward(self, X):
"""Forward function.
:param X: The input (batch) of the model contains word sequences for lstm,
features and feature weights.
:type X: For word sequences: a list of torch.Tensor pair (word sequence
and word mask) of shape (batch_size, sequence_lengt... | ['def', 'forward', '(', 'self', ',', 'X', ')', ':', 's', '=', 'X', '[', ':', '-', '2', ']', 'f', '=', 'X', '[', '-', '2', ']', 'w', '=', 'X', '[', '-', '1', ']', 'batch_size', '=', 'len', '(', 'f', ')', '# Generate lstm weight indices', 'x_idx', '=', 'self', '.', '_cuda', '(', 'torch', '.', 'as_tensor', '(', 'np', '.',... | Forward function.
:param X: The input (batch) of the model contains word sequences for lstm,
features and feature weights.
:type X: For word sequences: a list of torch.Tensor pair (word sequence
and word mask) of shape (batch_size, sequence_length).
For features: tor... | ['Forward', 'function', '.'] | train | https://github.com/HazyResearch/fonduer/blob/4520f86a716f03dcca458a9f4bddac75b4e7068f/src/fonduer/learning/disc_models/sparse_lstm.py#L25-L64 |
26 | google/grr | grr/server/grr_response_server/databases/mysql_flows.py | MySQLDBFlowMixin.CountFlowResultsByType | def CountFlowResultsByType(self, client_id, flow_id, cursor=None):
"""Returns counts of flow results grouped by result type."""
query = ("SELECT type, COUNT(*) FROM flow_results "
"FORCE INDEX (flow_results_by_client_id_flow_id_timestamp) "
"WHERE client_id = %s AND flow_id = %s "
... | python | def CountFlowResultsByType(self, client_id, flow_id, cursor=None):
"""Returns counts of flow results grouped by result type."""
query = ("SELECT type, COUNT(*) FROM flow_results "
"FORCE INDEX (flow_results_by_client_id_flow_id_timestamp) "
"WHERE client_id = %s AND flow_id = %s "
... | ['def', 'CountFlowResultsByType', '(', 'self', ',', 'client_id', ',', 'flow_id', ',', 'cursor', '=', 'None', ')', ':', 'query', '=', '(', '"SELECT type, COUNT(*) FROM flow_results "', '"FORCE INDEX (flow_results_by_client_id_flow_id_timestamp) "', '"WHERE client_id = %s AND flow_id = %s "', '"GROUP BY type"', ')', 'arg... | Returns counts of flow results grouped by result type. | ['Returns', 'counts', 'of', 'flow', 'results', 'grouped', 'by', 'result', 'type', '.'] | train | https://github.com/google/grr/blob/5cef4e8e2f0d5df43ea4877e9c798e0bf60bfe74/grr/server/grr_response_server/databases/mysql_flows.py#L1350-L1360 |
27 | google/grr | grr/client/grr_response_client/client_actions/artifact_collector.py | ArtifactCollector._ProcessGrepSource | def _ProcessGrepSource(self, source):
"""Find files fulfilling regex conditions."""
attributes = source.base_source.attributes
paths = artifact_utils.InterpolateListKbAttributes(
attributes["paths"], self.knowledge_base,
self.ignore_interpolation_errors)
regex = utils.RegexListDisjunctio... | python | def _ProcessGrepSource(self, source):
"""Find files fulfilling regex conditions."""
attributes = source.base_source.attributes
paths = artifact_utils.InterpolateListKbAttributes(
attributes["paths"], self.knowledge_base,
self.ignore_interpolation_errors)
regex = utils.RegexListDisjunctio... | ['def', '_ProcessGrepSource', '(', 'self', ',', 'source', ')', ':', 'attributes', '=', 'source', '.', 'base_source', '.', 'attributes', 'paths', '=', 'artifact_utils', '.', 'InterpolateListKbAttributes', '(', 'attributes', '[', '"paths"', ']', ',', 'self', '.', 'knowledge_base', ',', 'self', '.', 'ignore_interpolation_... | Find files fulfilling regex conditions. | ['Find', 'files', 'fulfilling', 'regex', 'conditions', '.'] | train | https://github.com/google/grr/blob/5cef4e8e2f0d5df43ea4877e9c798e0bf60bfe74/grr/client/grr_response_client/client_actions/artifact_collector.py#L208-L225 |
28 | F5Networks/f5-common-python | f5/multi_device/cluster/__init__.py | ClusterManager.manage_extant | def manage_extant(self, **kwargs):
'''Manage an existing cluster
:param kwargs: dict -- keyword args in dict
'''
self._check_device_number(kwargs['devices'])
self.trust_domain = TrustDomain(
devices=kwargs['devices'],
partition=kwargs['device_group_parti... | python | def manage_extant(self, **kwargs):
'''Manage an existing cluster
:param kwargs: dict -- keyword args in dict
'''
self._check_device_number(kwargs['devices'])
self.trust_domain = TrustDomain(
devices=kwargs['devices'],
partition=kwargs['device_group_parti... | ['def', 'manage_extant', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'self', '.', '_check_device_number', '(', 'kwargs', '[', "'devices'", ']', ')', 'self', '.', 'trust_domain', '=', 'TrustDomain', '(', 'devices', '=', 'kwargs', '[', "'devices'", ']', ',', 'partition', '=', 'kwargs', '[', "'device_group_partition'"... | Manage an existing cluster
:param kwargs: dict -- keyword args in dict | ['Manage', 'an', 'existing', 'cluster'] | train | https://github.com/F5Networks/f5-common-python/blob/7e67d5acd757a60e3d5f8c88c534bd72208f5494/f5/multi_device/cluster/__init__.py#L136-L148 |
29 | google/python-gflags | gflags/flagvalues.py | FlagValues._GetFlagsDefinedByModule | def _GetFlagsDefinedByModule(self, module):
"""Returns the list of flags defined by a module.
Args:
module: A module object or a module name (a string).
Returns:
A new list of Flag objects. Caller may update this list as he
wishes: none of those changes will affect the internals of this... | python | def _GetFlagsDefinedByModule(self, module):
"""Returns the list of flags defined by a module.
Args:
module: A module object or a module name (a string).
Returns:
A new list of Flag objects. Caller may update this list as he
wishes: none of those changes will affect the internals of this... | ['def', '_GetFlagsDefinedByModule', '(', 'self', ',', 'module', ')', ':', 'if', 'not', 'isinstance', '(', 'module', ',', 'str', ')', ':', 'module', '=', 'module', '.', '__name__', 'return', 'list', '(', 'self', '.', 'FlagsByModuleDict', '(', ')', '.', 'get', '(', 'module', ',', '[', ']', ')', ')'] | Returns the list of flags defined by a module.
Args:
module: A module object or a module name (a string).
Returns:
A new list of Flag objects. Caller may update this list as he
wishes: none of those changes will affect the internals of this
FlagValue object. | ['Returns', 'the', 'list', 'of', 'flags', 'defined', 'by', 'a', 'module', '.'] | train | https://github.com/google/python-gflags/blob/4f06c3d0d6cbe9b1fb90ee9fb1c082b3bf9285f6/gflags/flagvalues.py#L265-L279 |
30 | openp2pdesign/makerlabs | makerlabs/hackaday_io.py | get_labs | def get_labs(format):
"""Gets Hackerspaces data from hackaday.io."""
hackerspaces_json = data_from_hackaday_io(hackaday_io_labs_map_url)
hackerspaces = {}
# Load all the Hackerspaces
for i in hackerspaces_json:
current_lab = Hackerspace()
current_lab.id = i["id"]
current_la... | python | def get_labs(format):
"""Gets Hackerspaces data from hackaday.io."""
hackerspaces_json = data_from_hackaday_io(hackaday_io_labs_map_url)
hackerspaces = {}
# Load all the Hackerspaces
for i in hackerspaces_json:
current_lab = Hackerspace()
current_lab.id = i["id"]
current_la... | ['def', 'get_labs', '(', 'format', ')', ':', 'hackerspaces_json', '=', 'data_from_hackaday_io', '(', 'hackaday_io_labs_map_url', ')', 'hackerspaces', '=', '{', '}', '# Load all the Hackerspaces', 'for', 'i', 'in', 'hackerspaces_json', ':', 'current_lab', '=', 'Hackerspace', '(', ')', 'current_lab', '.', 'id', '=', 'i',... | Gets Hackerspaces data from hackaday.io. | ['Gets', 'Hackerspaces', 'data', 'from', 'hackaday', '.', 'io', '.'] | train | https://github.com/openp2pdesign/makerlabs/blob/b5838440174f10d370abb671358db9a99d7739fd/makerlabs/hackaday_io.py#L57-L137 |
31 | Azure/blobxfer | cli/cli.py | upload | def upload(ctx):
"""Upload files to Azure Storage"""
settings.add_cli_options(ctx.cli_options, settings.TransferAction.Upload)
ctx.initialize(settings.TransferAction.Upload)
specs = settings.create_upload_specifications(
ctx.cli_options, ctx.config)
del ctx.cli_options
for spec in specs:... | python | def upload(ctx):
"""Upload files to Azure Storage"""
settings.add_cli_options(ctx.cli_options, settings.TransferAction.Upload)
ctx.initialize(settings.TransferAction.Upload)
specs = settings.create_upload_specifications(
ctx.cli_options, ctx.config)
del ctx.cli_options
for spec in specs:... | ['def', 'upload', '(', 'ctx', ')', ':', 'settings', '.', 'add_cli_options', '(', 'ctx', '.', 'cli_options', ',', 'settings', '.', 'TransferAction', '.', 'Upload', ')', 'ctx', '.', 'initialize', '(', 'settings', '.', 'TransferAction', '.', 'Upload', ')', 'specs', '=', 'settings', '.', 'create_upload_specifications', '('... | Upload files to Azure Storage | ['Upload', 'files', 'to', 'Azure', 'Storage'] | train | https://github.com/Azure/blobxfer/blob/3eccbe7530cc6a20ab2d30f9e034b6f021817f34/cli/cli.py#L1106-L1116 |
32 | mitsei/dlkit | dlkit/json_/resource/sessions.py | ResourceBinAssignmentSession.get_assignable_bin_ids | def get_assignable_bin_ids(self, bin_id):
"""Gets a list of bins including and under the given bin node in which any resource can be assigned.
arg: bin_id (osid.id.Id): the ``Id`` of the ``Bin``
return: (osid.id.IdList) - list of assignable bin ``Ids``
raise: NullArgument - ``bin_id... | python | def get_assignable_bin_ids(self, bin_id):
"""Gets a list of bins including and under the given bin node in which any resource can be assigned.
arg: bin_id (osid.id.Id): the ``Id`` of the ``Bin``
return: (osid.id.IdList) - list of assignable bin ``Ids``
raise: NullArgument - ``bin_id... | ['def', 'get_assignable_bin_ids', '(', 'self', ',', 'bin_id', ')', ':', '# Implemented from template for', '# osid.resource.ResourceBinAssignmentSession.get_assignable_bin_ids', '# This will likely be overridden by an authorization adapter', 'mgr', '=', 'self', '.', '_get_provider_manager', '(', "'RESOURCE'", ',', 'loc... | Gets a list of bins including and under the given bin node in which any resource can be assigned.
arg: bin_id (osid.id.Id): the ``Id`` of the ``Bin``
return: (osid.id.IdList) - list of assignable bin ``Ids``
raise: NullArgument - ``bin_id`` is ``null``
raise: OperationFailed - unab... | ['Gets', 'a', 'list', 'of', 'bins', 'including', 'and', 'under', 'the', 'given', 'bin', 'node', 'in', 'which', 'any', 'resource', 'can', 'be', 'assigned', '.'] | train | https://github.com/mitsei/dlkit/blob/445f968a175d61c8d92c0f617a3c17dc1dc7c584/dlkit/json_/resource/sessions.py#L1562-L1581 |
33 | sbusard/wagoner | wagoner/table.py | Table.check | def check(self):
"""
Check that this table is complete, that is, every character of this
table can be followed by a new character.
:return: True if the table is complete, False otherwise.
"""
for character, followers in self.items():
for follower in followers... | python | def check(self):
"""
Check that this table is complete, that is, every character of this
table can be followed by a new character.
:return: True if the table is complete, False otherwise.
"""
for character, followers in self.items():
for follower in followers... | ['def', 'check', '(', 'self', ')', ':', 'for', 'character', ',', 'followers', 'in', 'self', '.', 'items', '(', ')', ':', 'for', 'follower', 'in', 'followers', ':', 'if', 'follower', 'not', 'in', 'self', ':', 'return', 'False', 'return', 'True'] | Check that this table is complete, that is, every character of this
table can be followed by a new character.
:return: True if the table is complete, False otherwise. | ['Check', 'that', 'this', 'table', 'is', 'complete', 'that', 'is', 'every', 'character', 'of', 'this', 'table', 'can', 'be', 'followed', 'by', 'a', 'new', 'character', '.'] | train | https://github.com/sbusard/wagoner/blob/7f83d66bbd0e009e4d4232ffdf319bd5a2a5683b/wagoner/table.py#L81-L92 |
34 | DataDog/integrations-core | openstack/datadog_checks/openstack/openstack.py | KeystoneCatalog.get_neutron_endpoint | def get_neutron_endpoint(cls, json_resp):
"""
Parse the service catalog returned by the Identity API for an endpoint matching the Neutron service
Sends a CRITICAL service check when none are found registered in the Catalog
"""
catalog = json_resp.get('token', {}).get('catalog', [... | python | def get_neutron_endpoint(cls, json_resp):
"""
Parse the service catalog returned by the Identity API for an endpoint matching the Neutron service
Sends a CRITICAL service check when none are found registered in the Catalog
"""
catalog = json_resp.get('token', {}).get('catalog', [... | ['def', 'get_neutron_endpoint', '(', 'cls', ',', 'json_resp', ')', ':', 'catalog', '=', 'json_resp', '.', 'get', '(', "'token'", ',', '{', '}', ')', '.', 'get', '(', "'catalog'", ',', '[', ']', ')', 'match', '=', "'neutron'", 'neutron_endpoint', '=', 'None', 'for', 'entry', 'in', 'catalog', ':', 'if', 'entry', '[', "'n... | Parse the service catalog returned by the Identity API for an endpoint matching the Neutron service
Sends a CRITICAL service check when none are found registered in the Catalog | ['Parse', 'the', 'service', 'catalog', 'returned', 'by', 'the', 'Identity', 'API', 'for', 'an', 'endpoint', 'matching', 'the', 'Neutron', 'service', 'Sends', 'a', 'CRITICAL', 'service', 'check', 'when', 'none', 'are', 'found', 'registered', 'in', 'the', 'Catalog'] | train | https://github.com/DataDog/integrations-core/blob/ebd41c873cf9f97a8c51bf9459bc6a7536af8acd/openstack/datadog_checks/openstack/openstack.py#L455-L479 |
35 | materialsproject/pymatgen | pymatgen/io/abinit/flows.py | Flow.validate_json_schema | def validate_json_schema(self):
"""Validate the JSON schema. Return list of errors."""
errors = []
for work in self:
for task in work:
if not task.get_results().validate_json_schema():
errors.append(task)
if not work.get_results().vali... | python | def validate_json_schema(self):
"""Validate the JSON schema. Return list of errors."""
errors = []
for work in self:
for task in work:
if not task.get_results().validate_json_schema():
errors.append(task)
if not work.get_results().vali... | ['def', 'validate_json_schema', '(', 'self', ')', ':', 'errors', '=', '[', ']', 'for', 'work', 'in', 'self', ':', 'for', 'task', 'in', 'work', ':', 'if', 'not', 'task', '.', 'get_results', '(', ')', '.', 'validate_json_schema', '(', ')', ':', 'errors', '.', 'append', '(', 'task', ')', 'if', 'not', 'work', '.', 'get_res... | Validate the JSON schema. Return list of errors. | ['Validate', 'the', 'JSON', 'schema', '.', 'Return', 'list', 'of', 'errors', '.'] | train | https://github.com/materialsproject/pymatgen/blob/4ca558cf72f8d5f8a1f21dfdfc0181a971c186da/pymatgen/io/abinit/flows.py#L426-L439 |
36 | ga4gh/ga4gh-server | ga4gh/server/datamodel/datasets.py | Dataset.getRnaQuantificationSetByName | def getRnaQuantificationSetByName(self, name):
"""
Returns the RnaQuantification set with the specified name, or raises
an exception otherwise.
"""
if name not in self._rnaQuantificationSetNameMap:
raise exceptions.RnaQuantificationSetNameNotFoundException(name)
... | python | def getRnaQuantificationSetByName(self, name):
"""
Returns the RnaQuantification set with the specified name, or raises
an exception otherwise.
"""
if name not in self._rnaQuantificationSetNameMap:
raise exceptions.RnaQuantificationSetNameNotFoundException(name)
... | ['def', 'getRnaQuantificationSetByName', '(', 'self', ',', 'name', ')', ':', 'if', 'name', 'not', 'in', 'self', '.', '_rnaQuantificationSetNameMap', ':', 'raise', 'exceptions', '.', 'RnaQuantificationSetNameNotFoundException', '(', 'name', ')', 'return', 'self', '.', '_rnaQuantificationSetNameMap', '[', 'name', ']'] | Returns the RnaQuantification set with the specified name, or raises
an exception otherwise. | ['Returns', 'the', 'RnaQuantification', 'set', 'with', 'the', 'specified', 'name', 'or', 'raises', 'an', 'exception', 'otherwise', '.'] | train | https://github.com/ga4gh/ga4gh-server/blob/1aa18922ef136db8604f6f098cb1732cba6f2a76/ga4gh/server/datamodel/datasets.py#L402-L409 |
37 | dpnova/python-xprintidle | xprintidle.py | _create_modulename | def _create_modulename(cdef_sources, source, sys_version):
"""
This is the same as CFFI's create modulename except we don't include the
CFFI version.
"""
key = '\x00'.join([sys_version[:3], source, cdef_sources])
key = key.encode('utf-8')
k1 = hex(binascii.crc32(key[0::2]) & 0xffffffff)
... | python | def _create_modulename(cdef_sources, source, sys_version):
"""
This is the same as CFFI's create modulename except we don't include the
CFFI version.
"""
key = '\x00'.join([sys_version[:3], source, cdef_sources])
key = key.encode('utf-8')
k1 = hex(binascii.crc32(key[0::2]) & 0xffffffff)
... | ['def', '_create_modulename', '(', 'cdef_sources', ',', 'source', ',', 'sys_version', ')', ':', 'key', '=', "'\\x00'", '.', 'join', '(', '[', 'sys_version', '[', ':', '3', ']', ',', 'source', ',', 'cdef_sources', ']', ')', 'key', '=', 'key', '.', 'encode', '(', "'utf-8'", ')', 'k1', '=', 'hex', '(', 'binascii', '.', 'c... | This is the same as CFFI's create modulename except we don't include the
CFFI version. | ['This', 'is', 'the', 'same', 'as', 'CFFI', 's', 'create', 'modulename', 'except', 'we', 'don', 't', 'include', 'the', 'CFFI', 'version', '.'] | train | https://github.com/dpnova/python-xprintidle/blob/cc8f3c13a5dd578073d20f3d42208fcb8e1983b8/xprintidle.py#L10-L21 |
38 | inasafe/inasafe | safe/gui/tools/batch/batch_dialog.py | BatchDialog.populate_table | def populate_table(self, scenario_directory):
""" Populate table with files from scenario_directory directory.
:param scenario_directory: Path where .txt & .py reside.
:type scenario_directory: QString
"""
parsed_files = []
unparsed_files = []
self.table.clearCon... | python | def populate_table(self, scenario_directory):
""" Populate table with files from scenario_directory directory.
:param scenario_directory: Path where .txt & .py reside.
:type scenario_directory: QString
"""
parsed_files = []
unparsed_files = []
self.table.clearCon... | ['def', 'populate_table', '(', 'self', ',', 'scenario_directory', ')', ':', 'parsed_files', '=', '[', ']', 'unparsed_files', '=', '[', ']', 'self', '.', 'table', '.', 'clearContents', '(', ')', '# Block signal to allow update checking only when the table is ready', 'self', '.', 'table', '.', 'blockSignals', '(', 'True'... | Populate table with files from scenario_directory directory.
:param scenario_directory: Path where .txt & .py reside.
:type scenario_directory: QString | ['Populate', 'table', 'with', 'files', 'from', 'scenario_directory', 'directory', '.'] | train | https://github.com/inasafe/inasafe/blob/831d60abba919f6d481dc94a8d988cc205130724/safe/gui/tools/batch/batch_dialog.py#L213-L251 |
39 | sibirrer/lenstronomy | lenstronomy/Sampling/likelihood.py | LikelihoodModule.logL | def logL(self, args):
"""
routine to compute X2 given variable parameters for a MCMC/PSO chain
"""
#extract parameters
kwargs_lens, kwargs_source, kwargs_lens_light, kwargs_ps, kwargs_cosmo = self.param.args2kwargs(args)
#generate image and computes likelihood
sel... | python | def logL(self, args):
"""
routine to compute X2 given variable parameters for a MCMC/PSO chain
"""
#extract parameters
kwargs_lens, kwargs_source, kwargs_lens_light, kwargs_ps, kwargs_cosmo = self.param.args2kwargs(args)
#generate image and computes likelihood
sel... | ['def', 'logL', '(', 'self', ',', 'args', ')', ':', '#extract parameters', 'kwargs_lens', ',', 'kwargs_source', ',', 'kwargs_lens_light', ',', 'kwargs_ps', ',', 'kwargs_cosmo', '=', 'self', '.', 'param', '.', 'args2kwargs', '(', 'args', ')', '#generate image and computes likelihood', 'self', '.', '_reset_point_source_c... | routine to compute X2 given variable parameters for a MCMC/PSO chain | ['routine', 'to', 'compute', 'X2', 'given', 'variable', 'parameters', 'for', 'a', 'MCMC', '/', 'PSO', 'chain'] | train | https://github.com/sibirrer/lenstronomy/blob/4edb100a4f3f4fdc4fac9b0032d2b0283d0aa1d6/lenstronomy/Sampling/likelihood.py#L96-L123 |
40 | wandb/client | wandb/__init__.py | log | def log(row=None, commit=True, *args, **kargs):
"""Log a dict to the global run's history. If commit is false, enables multiple calls before commiting.
Eg.
wandb.log({'train-loss': 0.5, 'accuracy': 0.9})
"""
if run is None:
raise ValueError(
"You must call `wandb.init` in the ... | python | def log(row=None, commit=True, *args, **kargs):
"""Log a dict to the global run's history. If commit is false, enables multiple calls before commiting.
Eg.
wandb.log({'train-loss': 0.5, 'accuracy': 0.9})
"""
if run is None:
raise ValueError(
"You must call `wandb.init` in the ... | ['def', 'log', '(', 'row', '=', 'None', ',', 'commit', '=', 'True', ',', '*', 'args', ',', '*', '*', 'kargs', ')', ':', 'if', 'run', 'is', 'None', ':', 'raise', 'ValueError', '(', '"You must call `wandb.init` in the same process before calling log"', ')', 'if', 'row', 'is', 'None', ':', 'row', '=', '{', '}', 'if', 'com... | Log a dict to the global run's history. If commit is false, enables multiple calls before commiting.
Eg.
wandb.log({'train-loss': 0.5, 'accuracy': 0.9}) | ['Log', 'a', 'dict', 'to', 'the', 'global', 'run', 's', 'history', '.', 'If', 'commit', 'is', 'false', 'enables', 'multiple', 'calls', 'before', 'commiting', '.'] | train | https://github.com/wandb/client/blob/7d08954ed5674fee223cd85ed0d8518fe47266b2/wandb/__init__.py#L465-L481 |
41 | aras7/deployr-python-client | deployr_connection.py | DeployRConnection.login | def login(self, username, password, disableautosave=True, print_response=True):
"""
:param username:
:param password:
:param disableautosave: boolean
:param print_response: print log if required
:return: status code, response data
"""
if type(username) != ... | python | def login(self, username, password, disableautosave=True, print_response=True):
"""
:param username:
:param password:
:param disableautosave: boolean
:param print_response: print log if required
:return: status code, response data
"""
if type(username) != ... | ['def', 'login', '(', 'self', ',', 'username', ',', 'password', ',', 'disableautosave', '=', 'True', ',', 'print_response', '=', 'True', ')', ':', 'if', 'type', '(', 'username', ')', '!=', 'str', ':', 'return', 'False', ',', '"Username must be string"', 'if', 'type', '(', 'password', ')', '!=', 'str', ':', 'return', 'F... | :param username:
:param password:
:param disableautosave: boolean
:param print_response: print log if required
:return: status code, response data | [':', 'param', 'username', ':', ':', 'param', 'password', ':', ':', 'param', 'disableautosave', ':', 'boolean', ':', 'param', 'print_response', ':', 'print', 'log', 'if', 'required', ':', 'return', ':', 'status', 'code', 'response', 'data'] | train | https://github.com/aras7/deployr-python-client/blob/3ca517ff38e9a7dd1e21fcc88d54537546b9e7e5/deployr_connection.py#L28-L55 |
42 | saltstack/salt | salt/states/zone.py | detached | def detached(name):
'''
Ensure zone is detached
name : string
name of the zone
'''
ret = {'name': name,
'changes': {},
'result': None,
'comment': ''}
zones = __salt__['zoneadm.list'](installed=True, configured=True)
if name in zones:
if zon... | python | def detached(name):
'''
Ensure zone is detached
name : string
name of the zone
'''
ret = {'name': name,
'changes': {},
'result': None,
'comment': ''}
zones = __salt__['zoneadm.list'](installed=True, configured=True)
if name in zones:
if zon... | ['def', 'detached', '(', 'name', ')', ':', 'ret', '=', '{', "'name'", ':', 'name', ',', "'changes'", ':', '{', '}', ',', "'result'", ':', 'None', ',', "'comment'", ':', "''", '}', 'zones', '=', '__salt__', '[', "'zoneadm.list'", ']', '(', 'installed', '=', 'True', ',', 'configured', '=', 'True', ')', 'if', 'name', 'in'... | Ensure zone is detached
name : string
name of the zone | ['Ensure', 'zone', 'is', 'detached'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/states/zone.py#L1100-L1138 |
43 | onelogin/python3-saml | src/onelogin/saml2/response.py | OneLogin_Saml2_Response.__query | def __query(self, query, tagid=None):
"""
Extracts nodes that match the query from the Response
:param query: Xpath Expresion
:type query: String
:param tagid: Tag ID
:type query: String
:returns: The queried nodes
:rtype: list
"""
if se... | python | def __query(self, query, tagid=None):
"""
Extracts nodes that match the query from the Response
:param query: Xpath Expresion
:type query: String
:param tagid: Tag ID
:type query: String
:returns: The queried nodes
:rtype: list
"""
if se... | ['def', '__query', '(', 'self', ',', 'query', ',', 'tagid', '=', 'None', ')', ':', 'if', 'self', '.', 'encrypted', ':', 'document', '=', 'self', '.', 'decrypted_document', 'else', ':', 'document', '=', 'self', '.', 'document', 'return', 'OneLogin_Saml2_XML', '.', 'query', '(', 'document', ',', 'query', ',', 'None', ','... | Extracts nodes that match the query from the Response
:param query: Xpath Expresion
:type query: String
:param tagid: Tag ID
:type query: String
:returns: The queried nodes
:rtype: list | ['Extracts', 'nodes', 'that', 'match', 'the', 'query', 'from', 'the', 'Response'] | train | https://github.com/onelogin/python3-saml/blob/064b7275fba1e5f39a9116ba1cdcc5d01fc34daa/src/onelogin/saml2/response.py#L760-L777 |
44 | ray-project/ray | python/ray/rllib/env/atari_wrappers.py | wrap_deepmind | def wrap_deepmind(env, dim=84, framestack=True):
"""Configure environment for DeepMind-style Atari.
Note that we assume reward clipping is done outside the wrapper.
Args:
dim (int): Dimension to resize observations to (dim x dim).
framestack (bool): Whether to framestack observations.
... | python | def wrap_deepmind(env, dim=84, framestack=True):
"""Configure environment for DeepMind-style Atari.
Note that we assume reward clipping is done outside the wrapper.
Args:
dim (int): Dimension to resize observations to (dim x dim).
framestack (bool): Whether to framestack observations.
... | ['def', 'wrap_deepmind', '(', 'env', ',', 'dim', '=', '84', ',', 'framestack', '=', 'True', ')', ':', 'env', '=', 'MonitorEnv', '(', 'env', ')', 'env', '=', 'NoopResetEnv', '(', 'env', ',', 'noop_max', '=', '30', ')', 'if', '"NoFrameskip"', 'in', 'env', '.', 'spec', '.', 'id', ':', 'env', '=', 'MaxAndSkipEnv', '(', 'en... | Configure environment for DeepMind-style Atari.
Note that we assume reward clipping is done outside the wrapper.
Args:
dim (int): Dimension to resize observations to (dim x dim).
framestack (bool): Whether to framestack observations. | ['Configure', 'environment', 'for', 'DeepMind', '-', 'style', 'Atari', '.'] | train | https://github.com/ray-project/ray/blob/4eade036a0505e244c976f36aaa2d64386b5129b/python/ray/rllib/env/atari_wrappers.py#L270-L291 |
45 | monarch-initiative/dipper | dipper/sources/Reactome.py | Reactome._parse_reactome_association_file | def _parse_reactome_association_file(
self, file, limit=None, subject_prefix=None, object_prefix=None):
"""
Parse ensembl gene to reactome pathway file
:param file: file path (not handle)
:param limit: limit (int, optional) limit the number of rows processed
:return: ... | python | def _parse_reactome_association_file(
self, file, limit=None, subject_prefix=None, object_prefix=None):
"""
Parse ensembl gene to reactome pathway file
:param file: file path (not handle)
:param limit: limit (int, optional) limit the number of rows processed
:return: ... | ['def', '_parse_reactome_association_file', '(', 'self', ',', 'file', ',', 'limit', '=', 'None', ',', 'subject_prefix', '=', 'None', ',', 'object_prefix', '=', 'None', ')', ':', 'eco_map', '=', 'Reactome', '.', 'get_eco_map', '(', 'Reactome', '.', 'map_files', '[', "'eco_map'", ']', ')', 'count', '=', '0', 'with', 'ope... | Parse ensembl gene to reactome pathway file
:param file: file path (not handle)
:param limit: limit (int, optional) limit the number of rows processed
:return: None | ['Parse', 'ensembl', 'gene', 'to', 'reactome', 'pathway', 'file', ':', 'param', 'file', ':', 'file', 'path', '(', 'not', 'handle', ')', ':', 'param', 'limit', ':', 'limit', '(', 'int', 'optional', ')', 'limit', 'the', 'number', 'of', 'rows', 'processed', ':', 'return', ':', 'None'] | train | https://github.com/monarch-initiative/dipper/blob/24cc80db355bbe15776edc5c7b41e0886959ba41/dipper/sources/Reactome.py#L75-L98 |
46 | jtwhite79/pyemu | pyemu/utils/helpers.py | PstFromFlopyModel.build_prior | def build_prior(self, fmt="ascii",filename=None,droptol=None, chunk=None, sparse=False,
sigma_range=6):
""" build a prior parameter covariance matrix.
Parameters
----------
fmt : str
the format to save the cov matrix. Options are "ascii","binary"... | python | def build_prior(self, fmt="ascii",filename=None,droptol=None, chunk=None, sparse=False,
sigma_range=6):
""" build a prior parameter covariance matrix.
Parameters
----------
fmt : str
the format to save the cov matrix. Options are "ascii","binary"... | ['def', 'build_prior', '(', 'self', ',', 'fmt', '=', '"ascii"', ',', 'filename', '=', 'None', ',', 'droptol', '=', 'None', ',', 'chunk', '=', 'None', ',', 'sparse', '=', 'False', ',', 'sigma_range', '=', '6', ')', ':', 'fmt', '=', 'fmt', '.', 'lower', '(', ')', 'acc_fmts', '=', '[', '"ascii"', ',', '"binary"', ',', '"u... | build a prior parameter covariance matrix.
Parameters
----------
fmt : str
the format to save the cov matrix. Options are "ascii","binary","uncfile", "coo".
default is "ascii"
filename : str
the filename to save the prior cov matr... | ['build', 'a', 'prior', 'parameter', 'covariance', 'matrix', '.'] | train | https://github.com/jtwhite79/pyemu/blob/c504d8e7a4097cec07655a6318d275739bd8148a/pyemu/utils/helpers.py#L2632-L2744 |
47 | wummel/linkchecker | linkcheck/checker/__init__.py | get_url_from | def get_url_from (base_url, recursion_level, aggregate,
parent_url=None, base_ref=None, line=0, column=0, page=0,
name=u"", parent_content_type=None, extern=None):
"""
Get url data from given base data.
@param base_url: base url from a link tag
@type base_url: string... | python | def get_url_from (base_url, recursion_level, aggregate,
parent_url=None, base_ref=None, line=0, column=0, page=0,
name=u"", parent_content_type=None, extern=None):
"""
Get url data from given base data.
@param base_url: base url from a link tag
@type base_url: string... | ['def', 'get_url_from', '(', 'base_url', ',', 'recursion_level', ',', 'aggregate', ',', 'parent_url', '=', 'None', ',', 'base_ref', '=', 'None', ',', 'line', '=', '0', ',', 'column', '=', '0', ',', 'page', '=', '0', ',', 'name', '=', 'u""', ',', 'parent_content_type', '=', 'None', ',', 'extern', '=', 'None', ')', ':', ... | Get url data from given base data.
@param base_url: base url from a link tag
@type base_url: string or None
@param recursion_level: current recursion level
@type recursion_level: number
@param aggregate: aggregate object
@type aggregate: aggregate.Consumer
@param parent_url: parent url
... | ['Get', 'url', 'data', 'from', 'given', 'base', 'data', '.'] | train | https://github.com/wummel/linkchecker/blob/c2ce810c3fb00b895a841a7be6b2e78c64e7b042/linkcheck/checker/__init__.py#L67-L125 |
48 | peercoin/peercoin_rpc | peercoin_rpc/peercoin_rpc.py | Client.sendtoaddress | def sendtoaddress(self, recv_addr, amount, comment=""):
"""send ammount to address, with optional comment. Returns txid.
sendtoaddress(ADDRESS, AMMOUNT, COMMENT)"""
return self.req("sendtoaddress", [recv_addr, amount, comment]) | python | def sendtoaddress(self, recv_addr, amount, comment=""):
"""send ammount to address, with optional comment. Returns txid.
sendtoaddress(ADDRESS, AMMOUNT, COMMENT)"""
return self.req("sendtoaddress", [recv_addr, amount, comment]) | ['def', 'sendtoaddress', '(', 'self', ',', 'recv_addr', ',', 'amount', ',', 'comment', '=', '""', ')', ':', 'return', 'self', '.', 'req', '(', '"sendtoaddress"', ',', '[', 'recv_addr', ',', 'amount', ',', 'comment', ']', ')'] | send ammount to address, with optional comment. Returns txid.
sendtoaddress(ADDRESS, AMMOUNT, COMMENT) | ['send', 'ammount', 'to', 'address', 'with', 'optional', 'comment', '.', 'Returns', 'txid', '.', 'sendtoaddress', '(', 'ADDRESS', 'AMMOUNT', 'COMMENT', ')'] | train | https://github.com/peercoin/peercoin_rpc/blob/6edd854c7fd607ad9f6f4d5eb8b8b7c7fd8c16cc/peercoin_rpc/peercoin_rpc.py#L182-L185 |
49 | tensorflow/hub | examples/image_retraining/retrain.py | variable_summaries | def variable_summaries(var):
"""Attach a lot of summaries to a Tensor (for TensorBoard visualization)."""
with tf.name_scope('summaries'):
mean = tf.reduce_mean(var)
tf.summary.scalar('mean', mean)
with tf.name_scope('stddev'):
stddev = tf.sqrt(tf.reduce_mean(tf.square(var - mean)))
tf.summary... | python | def variable_summaries(var):
"""Attach a lot of summaries to a Tensor (for TensorBoard visualization)."""
with tf.name_scope('summaries'):
mean = tf.reduce_mean(var)
tf.summary.scalar('mean', mean)
with tf.name_scope('stddev'):
stddev = tf.sqrt(tf.reduce_mean(tf.square(var - mean)))
tf.summary... | ['def', 'variable_summaries', '(', 'var', ')', ':', 'with', 'tf', '.', 'name_scope', '(', "'summaries'", ')', ':', 'mean', '=', 'tf', '.', 'reduce_mean', '(', 'var', ')', 'tf', '.', 'summary', '.', 'scalar', '(', "'mean'", ',', 'mean', ')', 'with', 'tf', '.', 'name_scope', '(', "'stddev'", ')', ':', 'stddev', '=', 'tf'... | Attach a lot of summaries to a Tensor (for TensorBoard visualization). | ['Attach', 'a', 'lot', 'of', 'summaries', 'to', 'a', 'Tensor', '(', 'for', 'TensorBoard', 'visualization', ')', '.'] | train | https://github.com/tensorflow/hub/blob/09f45963f6787322967b6fec61459f3ac56fbb27/examples/image_retraining/retrain.py#L709-L719 |
50 | henocdz/workon | workon/script.py | WorkOn._is_unique | def _is_unique(self, name, path):
"""verify if there is a project with given name or path
on the database
"""
project = None
try:
project = Project.select().where(
(Project.name == name) |
(Project.path == path)
)[0]
... | python | def _is_unique(self, name, path):
"""verify if there is a project with given name or path
on the database
"""
project = None
try:
project = Project.select().where(
(Project.name == name) |
(Project.path == path)
)[0]
... | ['def', '_is_unique', '(', 'self', ',', 'name', ',', 'path', ')', ':', 'project', '=', 'None', 'try', ':', 'project', '=', 'Project', '.', 'select', '(', ')', '.', 'where', '(', '(', 'Project', '.', 'name', '==', 'name', ')', '|', '(', 'Project', '.', 'path', '==', 'path', ')', ')', '[', '0', ']', 'except', ':', 'pass'... | verify if there is a project with given name or path
on the database | ['verify', 'if', 'there', 'is', 'a', 'project', 'with', 'given', 'name', 'or', 'path', 'on', 'the', 'database'] | train | https://github.com/henocdz/workon/blob/46f1f6dc4ea95d8efd10adf93a06737237a6874d/workon/script.py#L34-L47 |
51 | tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_cifar_tpu_range | def imagetransformer_cifar_tpu_range(rhp):
"""Range of hyperparameters for vizier."""
# After starting from base, set intervals for some parameters.
rhp.set_float("learning_rate", 0.01, 1.0, scale=rhp.LOG_SCALE)
rhp.set_discrete("num_decoder_layers", [8, 10, 12, 14, 16])
rhp.set_discrete("hidden_size", [256, ... | python | def imagetransformer_cifar_tpu_range(rhp):
"""Range of hyperparameters for vizier."""
# After starting from base, set intervals for some parameters.
rhp.set_float("learning_rate", 0.01, 1.0, scale=rhp.LOG_SCALE)
rhp.set_discrete("num_decoder_layers", [8, 10, 12, 14, 16])
rhp.set_discrete("hidden_size", [256, ... | ['def', 'imagetransformer_cifar_tpu_range', '(', 'rhp', ')', ':', '# After starting from base, set intervals for some parameters.', 'rhp', '.', 'set_float', '(', '"learning_rate"', ',', '0.01', ',', '1.0', ',', 'scale', '=', 'rhp', '.', 'LOG_SCALE', ')', 'rhp', '.', 'set_discrete', '(', '"num_decoder_layers"', ',', '['... | Range of hyperparameters for vizier. | ['Range', 'of', 'hyperparameters', 'for', 'vizier', '.'] | train | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1054-L1062 |
52 | sethmlarson/trytravis | trytravis.py | _wait_for_travis_build | def _wait_for_travis_build(url, commit, committed_at):
""" Waits for a Travis build to appear with the given commit SHA """
print('Waiting for a Travis build to appear '
'for `%s` after `%s`...' % (commit, committed_at))
import requests
slug = _slug_from_url(url)
start_time = time.time()
... | python | def _wait_for_travis_build(url, commit, committed_at):
""" Waits for a Travis build to appear with the given commit SHA """
print('Waiting for a Travis build to appear '
'for `%s` after `%s`...' % (commit, committed_at))
import requests
slug = _slug_from_url(url)
start_time = time.time()
... | ['def', '_wait_for_travis_build', '(', 'url', ',', 'commit', ',', 'committed_at', ')', ':', 'print', '(', "'Waiting for a Travis build to appear '", "'for `%s` after `%s`...'", '%', '(', 'commit', ',', 'committed_at', ')', ')', 'import', 'requests', 'slug', '=', '_slug_from_url', '(', 'url', ')', 'start_time', '=', 'ti... | Waits for a Travis build to appear with the given commit SHA | ['Waits', 'for', 'a', 'Travis', 'build', 'to', 'appear', 'with', 'the', 'given', 'commit', 'SHA'] | train | https://github.com/sethmlarson/trytravis/blob/d92ed708fe71d8db93a6df8077d23ee39ec0364e/trytravis.py#L188-L234 |
53 | nameko/nameko | nameko/cli/shell.py | make_nameko_helper | def make_nameko_helper(config):
"""Create a fake module that provides some convenient access to nameko
standalone functionality for interactive shell usage.
"""
module = ModuleType('nameko')
module.__doc__ = """Nameko shell helper for making rpc calls and dispatching
events.
Usage:
>>> n.rpc.se... | python | def make_nameko_helper(config):
"""Create a fake module that provides some convenient access to nameko
standalone functionality for interactive shell usage.
"""
module = ModuleType('nameko')
module.__doc__ = """Nameko shell helper for making rpc calls and dispatching
events.
Usage:
>>> n.rpc.se... | ['def', 'make_nameko_helper', '(', 'config', ')', ':', 'module', '=', 'ModuleType', '(', "'nameko'", ')', 'module', '.', '__doc__', '=', '"""Nameko shell helper for making rpc calls and dispatching\nevents.\n\nUsage:\n >>> n.rpc.service.method()\n "reply"\n\n >>> n.dispatch_event(\'service\', \'event_type\', \... | Create a fake module that provides some convenient access to nameko
standalone functionality for interactive shell usage. | ['Create', 'a', 'fake', 'module', 'that', 'provides', 'some', 'convenient', 'access', 'to', 'nameko', 'standalone', 'functionality', 'for', 'interactive', 'shell', 'usage', '.'] | train | https://github.com/nameko/nameko/blob/88d7e5211de4fcc1c34cd7f84d7c77f0619c5f5d/nameko/cli/shell.py#L58-L77 |
54 | earwig/mwparserfromhell | mwparserfromhell/nodes/template.py | Template.add | def add(self, name, value, showkey=None, before=None,
preserve_spacing=True):
"""Add a parameter to the template with a given *name* and *value*.
*name* and *value* can be anything parsable by
:func:`.utils.parse_anything`; pipes and equal signs are automatically
escaped fro... | python | def add(self, name, value, showkey=None, before=None,
preserve_spacing=True):
"""Add a parameter to the template with a given *name* and *value*.
*name* and *value* can be anything parsable by
:func:`.utils.parse_anything`; pipes and equal signs are automatically
escaped fro... | ['def', 'add', '(', 'self', ',', 'name', ',', 'value', ',', 'showkey', '=', 'None', ',', 'before', '=', 'None', ',', 'preserve_spacing', '=', 'True', ')', ':', 'name', ',', 'value', '=', 'parse_anything', '(', 'name', ')', ',', 'parse_anything', '(', 'value', ')', 'self', '.', '_surface_escape', '(', 'value', ',', '"|"... | Add a parameter to the template with a given *name* and *value*.
*name* and *value* can be anything parsable by
:func:`.utils.parse_anything`; pipes and equal signs are automatically
escaped from *value* when appropriate.
If *name* is already a parameter in the template, we'll replace ... | ['Add', 'a', 'parameter', 'to', 'the', 'template', 'with', 'a', 'given', '*', 'name', '*', 'and', '*', 'value', '*', '.'] | train | https://github.com/earwig/mwparserfromhell/blob/98dc30902d35c714a70aca8e6616f49d71cb24cc/mwparserfromhell/nodes/template.py#L232-L310 |
55 | liamw9534/bt-manager | bt_manager/audio.py | SBCAudioCodec.register_transport_ready_event | def register_transport_ready_event(self, user_cb, user_arg):
"""
Register for transport ready events. The `transport ready`
event is raised via a user callback. If the endpoint
is configured as a source, then the user may then
call :py:meth:`write_transport` in order to send da... | python | def register_transport_ready_event(self, user_cb, user_arg):
"""
Register for transport ready events. The `transport ready`
event is raised via a user callback. If the endpoint
is configured as a source, then the user may then
call :py:meth:`write_transport` in order to send da... | ['def', 'register_transport_ready_event', '(', 'self', ',', 'user_cb', ',', 'user_arg', ')', ':', 'self', '.', 'user_cb', '=', 'user_cb', 'self', '.', 'user_arg', '=', 'user_arg'] | Register for transport ready events. The `transport ready`
event is raised via a user callback. If the endpoint
is configured as a source, then the user may then
call :py:meth:`write_transport` in order to send data to
the associated sink.
Otherwise, if the endpoint is configur... | ['Register', 'for', 'transport', 'ready', 'events', '.', 'The', 'transport', 'ready', 'event', 'is', 'raised', 'via', 'a', 'user', 'callback', '.', 'If', 'the', 'endpoint', 'is', 'configured', 'as', 'a', 'source', 'then', 'the', 'user', 'may', 'then', 'call', ':', 'py', ':', 'meth', ':', 'write_transport', 'in', 'order... | train | https://github.com/liamw9534/bt-manager/blob/51be2919394ce8134c698359649bfad09eedf4ec/bt_manager/audio.py#L207-L227 |
56 | LesPatamechanix/patalib | src/patalib/anomaly.py | Anomaly.generate_anomaly | def generate_anomaly(self, input_word, list_of_dict_words, num):
""" Generate an anomaly. This is done
via a Psuedo-random number generator.
"""
results = []
for i in range(0,num):
index = randint(0,len(list_of_dict_words)-1)
name = list_of_dict_words[ind... | python | def generate_anomaly(self, input_word, list_of_dict_words, num):
""" Generate an anomaly. This is done
via a Psuedo-random number generator.
"""
results = []
for i in range(0,num):
index = randint(0,len(list_of_dict_words)-1)
name = list_of_dict_words[ind... | ['def', 'generate_anomaly', '(', 'self', ',', 'input_word', ',', 'list_of_dict_words', ',', 'num', ')', ':', 'results', '=', '[', ']', 'for', 'i', 'in', 'range', '(', '0', ',', 'num', ')', ':', 'index', '=', 'randint', '(', '0', ',', 'len', '(', 'list_of_dict_words', ')', '-', '1', ')', 'name', '=', 'list_of_dict_words... | Generate an anomaly. This is done
via a Psuedo-random number generator. | ['Generate', 'an', 'anomaly', '.', 'This', 'is', 'done', 'via', 'a', 'Psuedo', '-', 'random', 'number', 'generator', '.'] | train | https://github.com/LesPatamechanix/patalib/blob/d88cca409b1750fdeb88cece048b308f2a710955/src/patalib/anomaly.py#L9-L23 |
57 | cqlengine/cqlengine | cqlengine/query.py | AbstractQuerySet.count | def count(self):
""" Returns the number of rows matched by this query """
if self._batch:
raise CQLEngineException("Only inserts, updates, and deletes are available in batch mode")
if self._result_cache is None:
query = self._select_query()
query.count = True... | python | def count(self):
""" Returns the number of rows matched by this query """
if self._batch:
raise CQLEngineException("Only inserts, updates, and deletes are available in batch mode")
if self._result_cache is None:
query = self._select_query()
query.count = True... | ['def', 'count', '(', 'self', ')', ':', 'if', 'self', '.', '_batch', ':', 'raise', 'CQLEngineException', '(', '"Only inserts, updates, and deletes are available in batch mode"', ')', 'if', 'self', '.', '_result_cache', 'is', 'None', ':', 'query', '=', 'self', '.', '_select_query', '(', ')', 'query', '.', 'count', '=', ... | Returns the number of rows matched by this query | ['Returns', 'the', 'number', 'of', 'rows', 'matched', 'by', 'this', 'query'] | train | https://github.com/cqlengine/cqlengine/blob/7079eaf7071cbf5a045e1d1ab57f6d1b5ba3f9dc/cqlengine/query.py#L568-L579 |
58 | chaimleib/intervaltree | intervaltree/intervaltree.py | IntervalTree.remove | def remove(self, interval):
"""
Removes an interval from the tree, if present. If not, raises
ValueError.
Completes in O(log n) time.
"""
#self.verify()
if interval not in self:
#print(self.all_intervals)
raise ValueError
self.top_... | python | def remove(self, interval):
"""
Removes an interval from the tree, if present. If not, raises
ValueError.
Completes in O(log n) time.
"""
#self.verify()
if interval not in self:
#print(self.all_intervals)
raise ValueError
self.top_... | ['def', 'remove', '(', 'self', ',', 'interval', ')', ':', '#self.verify()', 'if', 'interval', 'not', 'in', 'self', ':', '#print(self.all_intervals)', 'raise', 'ValueError', 'self', '.', 'top_node', '=', 'self', '.', 'top_node', '.', 'remove', '(', 'interval', ')', 'self', '.', 'all_intervals', '.', 'remove', '(', 'inte... | Removes an interval from the tree, if present. If not, raises
ValueError.
Completes in O(log n) time. | ['Removes', 'an', 'interval', 'from', 'the', 'tree', 'if', 'present', '.', 'If', 'not', 'raises', 'ValueError', '.'] | train | https://github.com/chaimleib/intervaltree/blob/ffb2b1667f8b832e89324a75a175be8440504c9d/intervaltree/intervaltree.py#L356-L369 |
59 | aws/sagemaker-python-sdk | src/sagemaker/predictor.py | RealTimePredictor.delete_model | def delete_model(self):
"""Deletes the Amazon SageMaker models backing this predictor.
"""
request_failed = False
failed_models = []
for model_name in self._model_names:
try:
self.sagemaker_session.delete_model(model_name)
except Exception... | python | def delete_model(self):
"""Deletes the Amazon SageMaker models backing this predictor.
"""
request_failed = False
failed_models = []
for model_name in self._model_names:
try:
self.sagemaker_session.delete_model(model_name)
except Exception... | ['def', 'delete_model', '(', 'self', ')', ':', 'request_failed', '=', 'False', 'failed_models', '=', '[', ']', 'for', 'model_name', 'in', 'self', '.', '_model_names', ':', 'try', ':', 'self', '.', 'sagemaker_session', '.', 'delete_model', '(', 'model_name', ')', 'except', 'Exception', ':', '# pylint: disable=broad-exce... | Deletes the Amazon SageMaker models backing this predictor. | ['Deletes', 'the', 'Amazon', 'SageMaker', 'models', 'backing', 'this', 'predictor', '.'] | train | https://github.com/aws/sagemaker-python-sdk/blob/a9e724c7d3f5572b68c3903548c792a59d99799a/src/sagemaker/predictor.py#L131-L146 |
60 | google/mobly | mobly/asserts.py | assert_raises | def assert_raises(expected_exception, extras=None, *args, **kwargs):
"""Assert that an exception is raised when a function is called.
If no exception is raised, test fail. If an exception is raised but not
of the expected type, the exception is let through.
This should only be used as a context manage... | python | def assert_raises(expected_exception, extras=None, *args, **kwargs):
"""Assert that an exception is raised when a function is called.
If no exception is raised, test fail. If an exception is raised but not
of the expected type, the exception is let through.
This should only be used as a context manage... | ['def', 'assert_raises', '(', 'expected_exception', ',', 'extras', '=', 'None', ',', '*', 'args', ',', '*', '*', 'kwargs', ')', ':', 'context', '=', '_AssertRaisesContext', '(', 'expected_exception', ',', 'extras', '=', 'extras', ')', 'return', 'context'] | Assert that an exception is raised when a function is called.
If no exception is raised, test fail. If an exception is raised but not
of the expected type, the exception is let through.
This should only be used as a context manager:
with assert_raises(Exception):
func()
Args:
... | ['Assert', 'that', 'an', 'exception', 'is', 'raised', 'when', 'a', 'function', 'is', 'called', '.'] | train | https://github.com/google/mobly/blob/38ba2cf7d29a20e6a2fca1718eecb337df38db26/mobly/asserts.py#L60-L77 |
61 | saltstack/salt | salt/modules/mandrill.py | _http_request | def _http_request(url,
headers=None,
data=None):
'''
Make the HTTP request and return the body as python object.
'''
if not headers:
headers = _get_headers()
session = requests.session()
log.debug('Querying %s', url)
req = session.post(url,
... | python | def _http_request(url,
headers=None,
data=None):
'''
Make the HTTP request and return the body as python object.
'''
if not headers:
headers = _get_headers()
session = requests.session()
log.debug('Querying %s', url)
req = session.post(url,
... | ['def', '_http_request', '(', 'url', ',', 'headers', '=', 'None', ',', 'data', '=', 'None', ')', ':', 'if', 'not', 'headers', ':', 'headers', '=', '_get_headers', '(', ')', 'session', '=', 'requests', '.', 'session', '(', ')', 'log', '.', 'debug', '(', "'Querying %s'", ',', 'url', ')', 'req', '=', 'session', '.', 'post... | Make the HTTP request and return the body as python object. | ['Make', 'the', 'HTTP', 'request', 'and', 'return', 'the', 'body', 'as', 'python', 'object', '.'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/mandrill.py#L101-L132 |
62 | saltstack/salt | salt/returners/sqlite3_return.py | save_load | def save_load(jid, load, minions=None):
'''
Save the load to the specified jid
'''
log.debug('sqlite3 returner <save_load> called jid: %s load: %s', jid, load)
conn = _get_conn(ret=None)
cur = conn.cursor()
sql = '''INSERT INTO jids (jid, load) VALUES (:jid, :load)'''
cur.execute(sql,
... | python | def save_load(jid, load, minions=None):
'''
Save the load to the specified jid
'''
log.debug('sqlite3 returner <save_load> called jid: %s load: %s', jid, load)
conn = _get_conn(ret=None)
cur = conn.cursor()
sql = '''INSERT INTO jids (jid, load) VALUES (:jid, :load)'''
cur.execute(sql,
... | ['def', 'save_load', '(', 'jid', ',', 'load', ',', 'minions', '=', 'None', ')', ':', 'log', '.', 'debug', '(', "'sqlite3 returner <save_load> called jid: %s load: %s'", ',', 'jid', ',', 'load', ')', 'conn', '=', '_get_conn', '(', 'ret', '=', 'None', ')', 'cur', '=', 'conn', '.', 'cursor', '(', ')', 'sql', '=', "'''INSE... | Save the load to the specified jid | ['Save', 'the', 'load', 'to', 'the', 'specified', 'jid'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/returners/sqlite3_return.py#L182-L193 |
63 | iotile/coretools | iotilebuild/iotile/build/config/scons-local-3.0.1/SCons/Action.py | CommandGeneratorAction.get_presig | def get_presig(self, target, source, env, executor=None):
"""Return the signature contents of this action's command line.
This strips $(-$) and everything in between the string,
since those parts don't affect signatures.
"""
return self._generate(target, source, env, 1, executor... | python | def get_presig(self, target, source, env, executor=None):
"""Return the signature contents of this action's command line.
This strips $(-$) and everything in between the string,
since those parts don't affect signatures.
"""
return self._generate(target, source, env, 1, executor... | ['def', 'get_presig', '(', 'self', ',', 'target', ',', 'source', ',', 'env', ',', 'executor', '=', 'None', ')', ':', 'return', 'self', '.', '_generate', '(', 'target', ',', 'source', ',', 'env', ',', '1', ',', 'executor', ')', '.', 'get_presig', '(', 'target', ',', 'source', ',', 'env', ')'] | Return the signature contents of this action's command line.
This strips $(-$) and everything in between the string,
since those parts don't affect signatures. | ['Return', 'the', 'signature', 'contents', 'of', 'this', 'action', 's', 'command', 'line', '.'] | train | https://github.com/iotile/coretools/blob/2d794f5f1346b841b0dcd16c9d284e9bf2f3c6ec/iotilebuild/iotile/build/config/scons-local-3.0.1/SCons/Action.py#L1043-L1049 |
64 | cmbruns/pyopenvr | src/openvr/__init__.py | IVROverlay.getOverlayTransformAbsolute | def getOverlayTransformAbsolute(self, ulOverlayHandle):
"""Gets the transform if it is absolute. Returns an error if the transform is some other type."""
fn = self.function_table.getOverlayTransformAbsolute
peTrackingOrigin = ETrackingUniverseOrigin()
pmatTrackingOriginToOverlayTransfor... | python | def getOverlayTransformAbsolute(self, ulOverlayHandle):
"""Gets the transform if it is absolute. Returns an error if the transform is some other type."""
fn = self.function_table.getOverlayTransformAbsolute
peTrackingOrigin = ETrackingUniverseOrigin()
pmatTrackingOriginToOverlayTransfor... | ['def', 'getOverlayTransformAbsolute', '(', 'self', ',', 'ulOverlayHandle', ')', ':', 'fn', '=', 'self', '.', 'function_table', '.', 'getOverlayTransformAbsolute', 'peTrackingOrigin', '=', 'ETrackingUniverseOrigin', '(', ')', 'pmatTrackingOriginToOverlayTransform', '=', 'HmdMatrix34_t', '(', ')', 'result', '=', 'fn', '... | Gets the transform if it is absolute. Returns an error if the transform is some other type. | ['Gets', 'the', 'transform', 'if', 'it', 'is', 'absolute', '.', 'Returns', 'an', 'error', 'if', 'the', 'transform', 'is', 'some', 'other', 'type', '.'] | train | https://github.com/cmbruns/pyopenvr/blob/68395d26bb3df6ab1f0f059c38d441f962938be6/src/openvr/__init__.py#L4852-L4859 |
65 | gem/oq-engine | openquake/hazardlib/geo/geodetic.py | spherical_to_cartesian | def spherical_to_cartesian(lons, lats, depths=None):
"""
Return the position vectors (in Cartesian coordinates) of list of spherical
coordinates.
For equations see: http://mathworld.wolfram.com/SphericalCoordinates.html.
Parameters are components of spherical coordinates in a form of scalars,
... | python | def spherical_to_cartesian(lons, lats, depths=None):
"""
Return the position vectors (in Cartesian coordinates) of list of spherical
coordinates.
For equations see: http://mathworld.wolfram.com/SphericalCoordinates.html.
Parameters are components of spherical coordinates in a form of scalars,
... | ['def', 'spherical_to_cartesian', '(', 'lons', ',', 'lats', ',', 'depths', '=', 'None', ')', ':', 'phi', '=', 'numpy', '.', 'radians', '(', 'lons', ')', 'theta', '=', 'numpy', '.', 'radians', '(', 'lats', ')', 'if', 'depths', 'is', 'None', ':', 'rr', '=', 'EARTH_RADIUS', 'else', ':', 'rr', '=', 'EARTH_RADIUS', '-', 'nu... | Return the position vectors (in Cartesian coordinates) of list of spherical
coordinates.
For equations see: http://mathworld.wolfram.com/SphericalCoordinates.html.
Parameters are components of spherical coordinates in a form of scalars,
lists or numpy arrays. ``depths`` can be ``None`` in which case i... | ['Return', 'the', 'position', 'vectors', '(', 'in', 'Cartesian', 'coordinates', ')', 'of', 'list', 'of', 'spherical', 'coordinates', '.'] | train | https://github.com/gem/oq-engine/blob/8294553a0b8aba33fd96437a35065d03547d0040/openquake/hazardlib/geo/geodetic.py#L183-L221 |
66 | iterative/dvc | dvc/command/metrics.py | show_metrics | def show_metrics(metrics, all_branches=False, all_tags=False):
"""
Args:
metrics (list): Where each element is either a `list`
if an xpath was specified, otherwise a `str`
"""
for branch, val in metrics.items():
if all_branches or all_tags:
logger.info("{branch}:"... | python | def show_metrics(metrics, all_branches=False, all_tags=False):
"""
Args:
metrics (list): Where each element is either a `list`
if an xpath was specified, otherwise a `str`
"""
for branch, val in metrics.items():
if all_branches or all_tags:
logger.info("{branch}:"... | ['def', 'show_metrics', '(', 'metrics', ',', 'all_branches', '=', 'False', ',', 'all_tags', '=', 'False', ')', ':', 'for', 'branch', ',', 'val', 'in', 'metrics', '.', 'items', '(', ')', ':', 'if', 'all_branches', 'or', 'all_tags', ':', 'logger', '.', 'info', '(', '"{branch}:"', '.', 'format', '(', 'branch', '=', 'branc... | Args:
metrics (list): Where each element is either a `list`
if an xpath was specified, otherwise a `str` | ['Args', ':', 'metrics', '(', 'list', ')', ':', 'Where', 'each', 'element', 'is', 'either', 'a', 'list', 'if', 'an', 'xpath', 'was', 'specified', 'otherwise', 'a', 'str'] | train | https://github.com/iterative/dvc/blob/8bb21261e34c9632453e09090de7ebe50e38d341/dvc/command/metrics.py#L13-L33 |
67 | redbridge/molnctrl | molnctrl/cachemaker.py | monkeycache | def monkeycache(apis):
"""
Feed this a dictionary of api bananas, it spits out processed cache
"""
if isinstance(type(apis), type(None)) or apis is None:
return {}
verbs = set()
cache = {}
cache['count'] = apis['count']
cache['asyncapis'] = []
apilist = apis['api']
if a... | python | def monkeycache(apis):
"""
Feed this a dictionary of api bananas, it spits out processed cache
"""
if isinstance(type(apis), type(None)) or apis is None:
return {}
verbs = set()
cache = {}
cache['count'] = apis['count']
cache['asyncapis'] = []
apilist = apis['api']
if a... | ['def', 'monkeycache', '(', 'apis', ')', ':', 'if', 'isinstance', '(', 'type', '(', 'apis', ')', ',', 'type', '(', 'None', ')', ')', 'or', 'apis', 'is', 'None', ':', 'return', '{', '}', 'verbs', '=', 'set', '(', ')', 'cache', '=', '{', '}', 'cache', '[', "'count'", ']', '=', 'apis', '[', "'count'", ']', 'cache', '[', "... | Feed this a dictionary of api bananas, it spits out processed cache | ['Feed', 'this', 'a', 'dictionary', 'of', 'api', 'bananas', 'it', 'spits', 'out', 'processed', 'cache'] | train | https://github.com/redbridge/molnctrl/blob/9990ae7e522ce364bb61a735f774dc28de5f8e60/molnctrl/cachemaker.py#L82-L132 |
68 | portfoliome/foil | foil/ftp.py | download_ftp_url | def download_ftp_url(source_url, target_uri, buffer_size=8192):
"""Uses urllib. thread safe?"""
ensure_file_directory(target_uri)
with urllib.request.urlopen(source_url) as source_file:
with open(target_uri, 'wb') as target_file:
shutil.copyfileobj(source_file, target_file, buffer_size... | python | def download_ftp_url(source_url, target_uri, buffer_size=8192):
"""Uses urllib. thread safe?"""
ensure_file_directory(target_uri)
with urllib.request.urlopen(source_url) as source_file:
with open(target_uri, 'wb') as target_file:
shutil.copyfileobj(source_file, target_file, buffer_size... | ['def', 'download_ftp_url', '(', 'source_url', ',', 'target_uri', ',', 'buffer_size', '=', '8192', ')', ':', 'ensure_file_directory', '(', 'target_uri', ')', 'with', 'urllib', '.', 'request', '.', 'urlopen', '(', 'source_url', ')', 'as', 'source_file', ':', 'with', 'open', '(', 'target_uri', ',', "'wb'", ')', 'as', 'ta... | Uses urllib. thread safe? | ['Uses', 'urllib', '.', 'thread', 'safe?'] | train | https://github.com/portfoliome/foil/blob/b66d8cf4ab048a387d8c7a033b47e922ed6917d6/foil/ftp.py#L90-L97 |
69 | couchbase/couchbase-python-client | couchbase/bucket.py | Bucket.map_get | def map_get(self, key, mapkey):
"""
Retrieve a value from a map.
:param str key: The document ID
:param str mapkey: Key within the map to retrieve
:return: :class:`~.ValueResult`
:raise: :exc:`IndexError` if the mapkey does not exist
:raise: :cb_exc:`NotFoundErro... | python | def map_get(self, key, mapkey):
"""
Retrieve a value from a map.
:param str key: The document ID
:param str mapkey: Key within the map to retrieve
:return: :class:`~.ValueResult`
:raise: :exc:`IndexError` if the mapkey does not exist
:raise: :cb_exc:`NotFoundErro... | ['def', 'map_get', '(', 'self', ',', 'key', ',', 'mapkey', ')', ':', 'op', '=', 'SD', '.', 'get', '(', 'mapkey', ')', 'sdres', '=', 'self', '.', 'lookup_in', '(', 'key', ',', 'op', ')', 'return', 'self', '.', '_wrap_dsop', '(', 'sdres', ',', 'True', ')'] | Retrieve a value from a map.
:param str key: The document ID
:param str mapkey: Key within the map to retrieve
:return: :class:`~.ValueResult`
:raise: :exc:`IndexError` if the mapkey does not exist
:raise: :cb_exc:`NotFoundError` if the document does not exist.
.. seeal... | ['Retrieve', 'a', 'value', 'from', 'a', 'map', '.'] | train | https://github.com/couchbase/couchbase-python-client/blob/a7bada167785bf79a29c39f820d932a433a6a535/couchbase/bucket.py#L2147-L2161 |
70 | persandstrom/python-verisure | verisure/session.py | Session.set_lock_state | def set_lock_state(self, code, device_label, state):
""" Lock or unlock
Args:
code (str): Lock code
device_label (str): device label of lock
state (str): 'lock' or 'unlock'
"""
response = None
try:
response = requests.put(
... | python | def set_lock_state(self, code, device_label, state):
""" Lock or unlock
Args:
code (str): Lock code
device_label (str): device label of lock
state (str): 'lock' or 'unlock'
"""
response = None
try:
response = requests.put(
... | ['def', 'set_lock_state', '(', 'self', ',', 'code', ',', 'device_label', ',', 'state', ')', ':', 'response', '=', 'None', 'try', ':', 'response', '=', 'requests', '.', 'put', '(', 'urls', '.', 'set_lockstate', '(', 'self', '.', '_giid', ',', 'device_label', ',', 'state', ')', ',', 'headers', '=', '{', "'Accept'", ':', ... | Lock or unlock
Args:
code (str): Lock code
device_label (str): device label of lock
state (str): 'lock' or 'unlock' | ['Lock', 'or', 'unlock'] | train | https://github.com/persandstrom/python-verisure/blob/babd25e7f8fb2b24f12e4109dfa8a04041e8dcb8/verisure/session.py#L309-L329 |
71 | Sanji-IO/sanji | sanji/message.py | Message.to_event | def to_event(self):
"""
get rid of id, sign, tunnel and update message type
Notice: this method will return a deepcopy
"""
msg = copy.deepcopy(self)
for _ in ["id", "sign", "tunnel", "query", "param"]:
if not hasattr(msg, _):
continue
... | python | def to_event(self):
"""
get rid of id, sign, tunnel and update message type
Notice: this method will return a deepcopy
"""
msg = copy.deepcopy(self)
for _ in ["id", "sign", "tunnel", "query", "param"]:
if not hasattr(msg, _):
continue
... | ['def', 'to_event', '(', 'self', ')', ':', 'msg', '=', 'copy', '.', 'deepcopy', '(', 'self', ')', 'for', '_', 'in', '[', '"id"', ',', '"sign"', ',', '"tunnel"', ',', '"query"', ',', '"param"', ']', ':', 'if', 'not', 'hasattr', '(', 'msg', ',', '_', ')', ':', 'continue', 'delattr', '(', 'msg', ',', '_', ')', 'msg', '.',... | get rid of id, sign, tunnel and update message type
Notice: this method will return a deepcopy | ['get', 'rid', 'of', 'id', 'sign', 'tunnel', 'and', 'update', 'message', 'type', 'Notice', ':', 'this', 'method', 'will', 'return', 'a', 'deepcopy'] | train | https://github.com/Sanji-IO/sanji/blob/5c54cc2772bdfeae3337f785de1957237b828b34/sanji/message.py#L209-L222 |
72 | evhub/coconut | coconut/compiler/compiler.py | Compiler.destructuring_stmt_handle | def destructuring_stmt_handle(self, original, loc, tokens):
"""Process match assign blocks."""
internal_assert(len(tokens) == 2, "invalid destructuring assignment tokens", tokens)
matches, item = tokens
out = match_handle(loc, [matches, "in", item, None])
out += self.pattern_erro... | python | def destructuring_stmt_handle(self, original, loc, tokens):
"""Process match assign blocks."""
internal_assert(len(tokens) == 2, "invalid destructuring assignment tokens", tokens)
matches, item = tokens
out = match_handle(loc, [matches, "in", item, None])
out += self.pattern_erro... | ['def', 'destructuring_stmt_handle', '(', 'self', ',', 'original', ',', 'loc', ',', 'tokens', ')', ':', 'internal_assert', '(', 'len', '(', 'tokens', ')', '==', '2', ',', '"invalid destructuring assignment tokens"', ',', 'tokens', ')', 'matches', ',', 'item', '=', 'tokens', 'out', '=', 'match_handle', '(', 'loc', ',', ... | Process match assign blocks. | ['Process', 'match', 'assign', 'blocks', '.'] | train | https://github.com/evhub/coconut/blob/ff97177344e7604e89a0a98a977a87ed2a56fc6d/coconut/compiler/compiler.py#L1424-L1430 |
73 | Erotemic/utool | utool/util_alg.py | knapsack_iterative_int | def knapsack_iterative_int(items, maxweight):
r"""
Iterative knapsack method
Math:
maximize \sum_{i \in T} v_i
subject to \sum_{i \in T} w_i \leq W
Notes:
dpmat is the dynamic programming memoization matrix.
dpmat[i, w] is the total value of the items with weight at mos... | python | def knapsack_iterative_int(items, maxweight):
r"""
Iterative knapsack method
Math:
maximize \sum_{i \in T} v_i
subject to \sum_{i \in T} w_i \leq W
Notes:
dpmat is the dynamic programming memoization matrix.
dpmat[i, w] is the total value of the items with weight at mos... | ['def', 'knapsack_iterative_int', '(', 'items', ',', 'maxweight', ')', ':', 'values', '=', '[', 't', '[', '0', ']', 'for', 't', 'in', 'items', ']', 'weights', '=', '[', 't', '[', '1', ']', 'for', 't', 'in', 'items', ']', 'maxsize', '=', 'maxweight', '+', '1', '# Sparse representation seems better', 'dpmat', '=', 'defau... | r"""
Iterative knapsack method
Math:
maximize \sum_{i \in T} v_i
subject to \sum_{i \in T} w_i \leq W
Notes:
dpmat is the dynamic programming memoization matrix.
dpmat[i, w] is the total value of the items with weight at most W
T is idx_subset, the set of indicies i... | ['r', 'Iterative', 'knapsack', 'method'] | train | https://github.com/Erotemic/utool/blob/3b27e1f4e6e6fb23cd8744af7b7195b57d99e03a/utool/util_alg.py#L1400-L1477 |
74 | jorgenschaefer/elpy | elpy/server.py | ElpyRPCServer.rpc_get_names | def rpc_get_names(self, filename, source, offset):
"""Get all possible names
"""
source = get_source(source)
if hasattr(self.backend, "rpc_get_names"):
return self.backend.rpc_get_names(filename, source, offset)
else:
raise Fault("get_names not implemente... | python | def rpc_get_names(self, filename, source, offset):
"""Get all possible names
"""
source = get_source(source)
if hasattr(self.backend, "rpc_get_names"):
return self.backend.rpc_get_names(filename, source, offset)
else:
raise Fault("get_names not implemente... | ['def', 'rpc_get_names', '(', 'self', ',', 'filename', ',', 'source', ',', 'offset', ')', ':', 'source', '=', 'get_source', '(', 'source', ')', 'if', 'hasattr', '(', 'self', '.', 'backend', ',', '"rpc_get_names"', ')', ':', 'return', 'self', '.', 'backend', '.', 'rpc_get_names', '(', 'filename', ',', 'source', ',', 'of... | Get all possible names | ['Get', 'all', 'possible', 'names'] | train | https://github.com/jorgenschaefer/elpy/blob/ffd982f829b11e53f2be187c7b770423341f29bc/elpy/server.py#L199-L208 |
75 | praekeltfoundation/seaworthy | seaworthy/helpers.py | fetch_image | def fetch_image(client, name):
"""
Fetch an image if it isn't already present.
This works like ``docker pull`` and will pull the tag ``latest`` if no tag
is specified in the image name.
"""
try:
image = client.images.get(name)
except docker.errors.ImageNotFound:
name, tag = ... | python | def fetch_image(client, name):
"""
Fetch an image if it isn't already present.
This works like ``docker pull`` and will pull the tag ``latest`` if no tag
is specified in the image name.
"""
try:
image = client.images.get(name)
except docker.errors.ImageNotFound:
name, tag = ... | ['def', 'fetch_image', '(', 'client', ',', 'name', ')', ':', 'try', ':', 'image', '=', 'client', '.', 'images', '.', 'get', '(', 'name', ')', 'except', 'docker', '.', 'errors', '.', 'ImageNotFound', ':', 'name', ',', 'tag', '=', '_parse_image_tag', '(', 'name', ')', 'tag', '=', "'latest'", 'if', 'tag', 'is', 'None', 'e... | Fetch an image if it isn't already present.
This works like ``docker pull`` and will pull the tag ``latest`` if no tag
is specified in the image name. | ['Fetch', 'an', 'image', 'if', 'it', 'isn', 't', 'already', 'present', '.'] | train | https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L27-L44 |
76 | F5Networks/f5-common-python | f5/bigip/tm/ltm/pool.py | Members.update | def update(self, **kwargs):
"""Call this to change the configuration of the service on the device.
This method uses HTTP PUT to alter the service state on the device.
The attributes of the instance will be packaged as a dictionary. That
dictionary will be updated with kwargs. It is t... | python | def update(self, **kwargs):
"""Call this to change the configuration of the service on the device.
This method uses HTTP PUT to alter the service state on the device.
The attributes of the instance will be packaged as a dictionary. That
dictionary will be updated with kwargs. It is t... | ['def', 'update', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'checked', '=', 'self', '.', '_check_member_parameters', '(', '*', '*', 'kwargs', ')', 'return', 'super', '(', 'Members', ',', 'self', ')', '.', '_update', '(', '*', '*', 'checked', ')'] | Call this to change the configuration of the service on the device.
This method uses HTTP PUT to alter the service state on the device.
The attributes of the instance will be packaged as a dictionary. That
dictionary will be updated with kwargs. It is then submitted as JSON
to the de... | ['Call', 'this', 'to', 'change', 'the', 'configuration', 'of', 'the', 'service', 'on', 'the', 'device', '.'] | train | https://github.com/F5Networks/f5-common-python/blob/7e67d5acd757a60e3d5f8c88c534bd72208f5494/f5/bigip/tm/ltm/pool.py#L144-L160 |
77 | ethereum/py-evm | eth/estimators/gas.py | binary_gas_search | def binary_gas_search(state: BaseState, transaction: BaseTransaction, tolerance: int=1) -> int:
"""
Run the transaction with various gas limits, progressively
approaching the minimum needed to succeed without an OutOfGas exception.
The starting range of possible estimates is:
[transaction.intrinsic... | python | def binary_gas_search(state: BaseState, transaction: BaseTransaction, tolerance: int=1) -> int:
"""
Run the transaction with various gas limits, progressively
approaching the minimum needed to succeed without an OutOfGas exception.
The starting range of possible estimates is:
[transaction.intrinsic... | ['def', 'binary_gas_search', '(', 'state', ':', 'BaseState', ',', 'transaction', ':', 'BaseTransaction', ',', 'tolerance', ':', 'int', '=', '1', ')', '->', 'int', ':', 'if', 'not', 'hasattr', '(', 'transaction', ',', "'sender'", ')', ':', 'raise', 'TypeError', '(', '"Transaction is missing attribute sender."', ',', '"I... | Run the transaction with various gas limits, progressively
approaching the minimum needed to succeed without an OutOfGas exception.
The starting range of possible estimates is:
[transaction.intrinsic_gas, state.gas_limit].
After the first OutOfGas exception, the range is: (largest_limit_out_of_gas, sta... | ['Run', 'the', 'transaction', 'with', 'various', 'gas', 'limits', 'progressively', 'approaching', 'the', 'minimum', 'needed', 'to', 'succeed', 'without', 'an', 'OutOfGas', 'exception', '.'] | train | https://github.com/ethereum/py-evm/blob/58346848f076116381d3274bbcea96b9e2cfcbdf/eth/estimators/gas.py#L28-L78 |
78 | SatelliteQE/nailgun | nailgun/entities.py | ContentViewPuppetModule.read | def read(self, entity=None, attrs=None, ignore=None, params=None):
"""Provide a default value for ``entity``.
By default, ``nailgun.entity_mixins.EntityReadMixin.read provides a
default value for ``entity`` like so::
entity = type(self)()
However, :class:`ContentViewPuppet... | python | def read(self, entity=None, attrs=None, ignore=None, params=None):
"""Provide a default value for ``entity``.
By default, ``nailgun.entity_mixins.EntityReadMixin.read provides a
default value for ``entity`` like so::
entity = type(self)()
However, :class:`ContentViewPuppet... | ['def', 'read', '(', 'self', ',', 'entity', '=', 'None', ',', 'attrs', '=', 'None', ',', 'ignore', '=', 'None', ',', 'params', '=', 'None', ')', ':', "# read() should not change the state of the object it's called on, but", '# super() alters the attributes of any entity passed in. Creating a new', '# object and passing... | Provide a default value for ``entity``.
By default, ``nailgun.entity_mixins.EntityReadMixin.read provides a
default value for ``entity`` like so::
entity = type(self)()
However, :class:`ContentViewPuppetModule` requires that an
``content_view`` be provided, so this techniq... | ['Provide', 'a', 'default', 'value', 'for', 'entity', '.'] | train | https://github.com/SatelliteQE/nailgun/blob/c36d8c20862e87bf6975bd48ac1ca40a9e634eaa/nailgun/entities.py#L2453-L2480 |
79 | jordanh/neurio-python | neurio/__init__.py | Client.get_samples_live_last | def get_samples_live_last(self, sensor_id):
"""Get the last sample recorded by the sensor.
Args:
sensor_id (string): hexadecimal id of the sensor to query, e.g.
``0x0013A20040B65FAD``
Returns:
list: dictionary objects containing sample data
"""
url = "https://api.neur.io/v1/sam... | python | def get_samples_live_last(self, sensor_id):
"""Get the last sample recorded by the sensor.
Args:
sensor_id (string): hexadecimal id of the sensor to query, e.g.
``0x0013A20040B65FAD``
Returns:
list: dictionary objects containing sample data
"""
url = "https://api.neur.io/v1/sam... | ['def', 'get_samples_live_last', '(', 'self', ',', 'sensor_id', ')', ':', 'url', '=', '"https://api.neur.io/v1/samples/live/last"', 'headers', '=', 'self', '.', '__gen_headers', '(', ')', 'headers', '[', '"Content-Type"', ']', '=', '"application/json"', 'params', '=', '{', '"sensorId"', ':', 'sensor_id', '}', 'url', '=... | Get the last sample recorded by the sensor.
Args:
sensor_id (string): hexadecimal id of the sensor to query, e.g.
``0x0013A20040B65FAD``
Returns:
list: dictionary objects containing sample data | ['Get', 'the', 'last', 'sample', 'recorded', 'by', 'the', 'sensor', '.'] | train | https://github.com/jordanh/neurio-python/blob/3a1bcadadb3bb3ad48f2df41c039d8b828ffd9c8/neurio/__init__.py#L439-L458 |
80 | markovmodel/PyEMMA | pyemma/plots/plots2d.py | plot_contour | def plot_contour(
xall, yall, zall, ax=None, cmap=None,
ncontours=100, vmin=None, vmax=None, levels=None,
cbar=True, cax=None, cbar_label=None,
cbar_orientation='vertical', norm=None, nbins=100,
method='nearest', mask=False, **kwargs):
"""Plot a two-dimensional contour map by... | python | def plot_contour(
xall, yall, zall, ax=None, cmap=None,
ncontours=100, vmin=None, vmax=None, levels=None,
cbar=True, cax=None, cbar_label=None,
cbar_orientation='vertical', norm=None, nbins=100,
method='nearest', mask=False, **kwargs):
"""Plot a two-dimensional contour map by... | ['def', 'plot_contour', '(', 'xall', ',', 'yall', ',', 'zall', ',', 'ax', '=', 'None', ',', 'cmap', '=', 'None', ',', 'ncontours', '=', '100', ',', 'vmin', '=', 'None', ',', 'vmax', '=', 'None', ',', 'levels', '=', 'None', ',', 'cbar', '=', 'True', ',', 'cax', '=', 'None', ',', 'cbar_label', '=', 'None', ',', 'cbar_ori... | Plot a two-dimensional contour map by interpolating
scattered data on a grid.
Parameters
----------
xall : ndarray(T)
Sample x-coordinates.
yall : ndarray(T)
Sample y-coordinates.
zall : ndarray(T)
Sample z-coordinates.
ax : matplotlib.Axes object, optional, default=... | ['Plot', 'a', 'two', '-', 'dimensional', 'contour', 'map', 'by', 'interpolating', 'scattered', 'data', 'on', 'a', 'grid', '.'] | train | https://github.com/markovmodel/PyEMMA/blob/5c3124398217de05ba5ce9c8fb01519222481ab8/pyemma/plots/plots2d.py#L686-L815 |
81 | tomturner/django-tenants | django_tenants/clone.py | CloneSchema.clone_schema | def clone_schema(self, base_schema_name, new_schema_name):
"""
Creates a new schema `new_schema_name` as a clone of an existing schema
`old_schema_name`.
"""
connection.set_schema_to_public()
cursor = connection.cursor()
# check if the clone_schema function alrea... | python | def clone_schema(self, base_schema_name, new_schema_name):
"""
Creates a new schema `new_schema_name` as a clone of an existing schema
`old_schema_name`.
"""
connection.set_schema_to_public()
cursor = connection.cursor()
# check if the clone_schema function alrea... | ['def', 'clone_schema', '(', 'self', ',', 'base_schema_name', ',', 'new_schema_name', ')', ':', 'connection', '.', 'set_schema_to_public', '(', ')', 'cursor', '=', 'connection', '.', 'cursor', '(', ')', '# check if the clone_schema function already exists in the db', 'try', ':', 'cursor', '.', 'execute', '(', '"SELECT ... | Creates a new schema `new_schema_name` as a clone of an existing schema
`old_schema_name`. | ['Creates', 'a', 'new', 'schema', 'new_schema_name', 'as', 'a', 'clone', 'of', 'an', 'existing', 'schema', 'old_schema_name', '.'] | train | https://github.com/tomturner/django-tenants/blob/f3e06e2b0facee7ed797e5694bcac433df3e5315/django_tenants/clone.py#L213-L233 |
82 | toumorokoshi/jenks | jenks/utils.py | get_configuration_file | def get_configuration_file():
""" return jenks configuration file """
path = os.path.abspath(os.curdir)
while path != os.sep:
config_path = os.path.join(path, CONFIG_FILE_NAME)
if os.path.exists(config_path):
return config_path
path = os.path.dirname(path)
return None | python | def get_configuration_file():
""" return jenks configuration file """
path = os.path.abspath(os.curdir)
while path != os.sep:
config_path = os.path.join(path, CONFIG_FILE_NAME)
if os.path.exists(config_path):
return config_path
path = os.path.dirname(path)
return None | ['def', 'get_configuration_file', '(', ')', ':', 'path', '=', 'os', '.', 'path', '.', 'abspath', '(', 'os', '.', 'curdir', ')', 'while', 'path', '!=', 'os', '.', 'sep', ':', 'config_path', '=', 'os', '.', 'path', '.', 'join', '(', 'path', ',', 'CONFIG_FILE_NAME', ')', 'if', 'os', '.', 'path', '.', 'exists', '(', 'confi... | return jenks configuration file | ['return', 'jenks', 'configuration', 'file'] | train | https://github.com/toumorokoshi/jenks/blob/d3333a7b86ba290b7185aa5b8da75e76a28124f5/jenks/utils.py#L36-L44 |
83 | mitsei/dlkit | dlkit/json_/repository/sessions.py | AssetAdminSession.get_asset_form_for_create | def get_asset_form_for_create(self, asset_record_types):
"""Gets the asset form for creating new assets.
A new form should be requested for each create transaction.
arg: asset_record_types (osid.type.Type[]): array of asset
record types
return: (osid.repository.Asset... | python | def get_asset_form_for_create(self, asset_record_types):
"""Gets the asset form for creating new assets.
A new form should be requested for each create transaction.
arg: asset_record_types (osid.type.Type[]): array of asset
record types
return: (osid.repository.Asset... | ['def', 'get_asset_form_for_create', '(', 'self', ',', 'asset_record_types', ')', ':', '# Implemented from template for', '# osid.resource.ResourceAdminSession.get_resource_form_for_create_template', 'for', 'arg', 'in', 'asset_record_types', ':', 'if', 'not', 'isinstance', '(', 'arg', ',', 'ABCType', ')', ':', 'raise',... | Gets the asset form for creating new assets.
A new form should be requested for each create transaction.
arg: asset_record_types (osid.type.Type[]): array of asset
record types
return: (osid.repository.AssetForm) - the asset form
raise: NullArgument - ``asset_record... | ['Gets', 'the', 'asset', 'form', 'for', 'creating', 'new', 'assets', '.'] | train | https://github.com/mitsei/dlkit/blob/445f968a175d61c8d92c0f617a3c17dc1dc7c584/dlkit/json_/repository/sessions.py#L1279-L1314 |
84 | martinkosir/neverbounce-python | neverbounce/client.py | NeverBounce.verify | def verify(self, email):
"""
Verify a single email address.
:param str email: Email address to verify.
:return: A VerifiedEmail object.
"""
resp = self._call(endpoint='single', data={'email': email})
return VerifiedEmail(email, resp['result']) | python | def verify(self, email):
"""
Verify a single email address.
:param str email: Email address to verify.
:return: A VerifiedEmail object.
"""
resp = self._call(endpoint='single', data={'email': email})
return VerifiedEmail(email, resp['result']) | ['def', 'verify', '(', 'self', ',', 'email', ')', ':', 'resp', '=', 'self', '.', '_call', '(', 'endpoint', '=', "'single'", ',', 'data', '=', '{', "'email'", ':', 'email', '}', ')', 'return', 'VerifiedEmail', '(', 'email', ',', 'resp', '[', "'result'", ']', ')'] | Verify a single email address.
:param str email: Email address to verify.
:return: A VerifiedEmail object. | ['Verify', 'a', 'single', 'email', 'address', '.', ':', 'param', 'str', 'email', ':', 'Email', 'address', 'to', 'verify', '.', ':', 'return', ':', 'A', 'VerifiedEmail', 'object', '.'] | train | https://github.com/martinkosir/neverbounce-python/blob/8d8b3f381dbff2a753a8770fac0d2bfab80d5bec/neverbounce/client.py#L18-L25 |
85 | JarryShaw/PyPCAPKit | src/protocols/internet/hip.py | HIP._read_para_hip_signature_2 | def _read_para_hip_signature_2(self, code, cbit, clen, *, desc, length, version):
"""Read HIP HIP_SIGNATURE_2 parameter.
Structure of HIP HIP_SIGNATURE_2 parameter [RFC 7401]:
0 1 2 3
0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9... | python | def _read_para_hip_signature_2(self, code, cbit, clen, *, desc, length, version):
"""Read HIP HIP_SIGNATURE_2 parameter.
Structure of HIP HIP_SIGNATURE_2 parameter [RFC 7401]:
0 1 2 3
0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9... | ['def', '_read_para_hip_signature_2', '(', 'self', ',', 'code', ',', 'cbit', ',', 'clen', ',', '*', ',', 'desc', ',', 'length', ',', 'version', ')', ':', '_algo', '=', 'self', '.', '_read_unpack', '(', '2', ')', '_sign', '=', 'self', '.', '_read_fileng', '(', 'clen', '-', '2', ')', 'hip_signature_2', '=', 'dict', '(', ... | Read HIP HIP_SIGNATURE_2 parameter.
Structure of HIP HIP_SIGNATURE_2 parameter [RFC 7401]:
0 1 2 3
0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-... | ['Read', 'HIP', 'HIP_SIGNATURE_2', 'parameter', '.'] | train | https://github.com/JarryShaw/PyPCAPKit/blob/c7f0da9aebc2cf210bf8f8b912f7d3cbb98ca10e/src/protocols/internet/hip.py#L2023-L2061 |
86 | Preston-Landers/concurrent-log-handler | src/concurrent_log_handler/__init__.py | ConcurrentRotatingFileHandler.emit | def emit(self, record):
"""
Emit a record.
Override from parent class to handle file locking for the duration of rollover and write.
This also does the formatting *before* locks are obtained, in case the format itself does
logging calls from within. Rollover also occurs while th... | python | def emit(self, record):
"""
Emit a record.
Override from parent class to handle file locking for the duration of rollover and write.
This also does the formatting *before* locks are obtained, in case the format itself does
logging calls from within. Rollover also occurs while th... | ['def', 'emit', '(', 'self', ',', 'record', ')', ':', '# noinspection PyBroadException', 'try', ':', 'msg', '=', 'self', '.', 'format', '(', 'record', ')', 'try', ':', 'self', '.', '_do_lock', '(', ')', 'try', ':', 'if', 'self', '.', 'shouldRollover', '(', 'record', ')', ':', 'self', '.', 'doRollover', '(', ')', 'excep... | Emit a record.
Override from parent class to handle file locking for the duration of rollover and write.
This also does the formatting *before* locks are obtained, in case the format itself does
logging calls from within. Rollover also occurs while the lock is held. | ['Emit', 'a', 'record', '.'] | train | https://github.com/Preston-Landers/concurrent-log-handler/blob/8e0b8e28c2b12e854853d723b3c28346a3218914/src/concurrent_log_handler/__init__.py#L298-L324 |
87 | Autodesk/aomi | aomi/vault.py | Client.server_version | def server_version(self):
"""Attempts to determine the version of Vault that a
server is running. Some actions will change on older
Vault deployments."""
health_url = "%s/v1/sys/health" % self.vault_addr
resp = self.session.request('get', health_url, **self._kwargs)
if re... | python | def server_version(self):
"""Attempts to determine the version of Vault that a
server is running. Some actions will change on older
Vault deployments."""
health_url = "%s/v1/sys/health" % self.vault_addr
resp = self.session.request('get', health_url, **self._kwargs)
if re... | ['def', 'server_version', '(', 'self', ')', ':', 'health_url', '=', '"%s/v1/sys/health"', '%', 'self', '.', 'vault_addr', 'resp', '=', 'self', '.', 'session', '.', 'request', '(', "'get'", ',', 'health_url', ',', '*', '*', 'self', '.', '_kwargs', ')', 'if', 'resp', '.', 'status_code', '==', '200', 'or', 'resp', '.', 's... | Attempts to determine the version of Vault that a
server is running. Some actions will change on older
Vault deployments. | ['Attempts', 'to', 'determine', 'the', 'version', 'of', 'Vault', 'that', 'a', 'server', 'is', 'running', '.', 'Some', 'actions', 'will', 'change', 'on', 'older', 'Vault', 'deployments', '.'] | train | https://github.com/Autodesk/aomi/blob/84da2dfb0424837adf9c4ddc1aa352e942bb7a4a/aomi/vault.py#L191-L204 |
88 | ryanmcgrath/twython | twython/streaming/types.py | TwythonStreamerTypes.site | def site(self, **params):
"""Stream site
Accepted params found at:
https://dev.twitter.com/docs/api/1.1/get/site
"""
url = 'https://sitestream.twitter.com/%s/site.json' \
% self.streamer.api_version
self.streamer._request(url, params=params) | python | def site(self, **params):
"""Stream site
Accepted params found at:
https://dev.twitter.com/docs/api/1.1/get/site
"""
url = 'https://sitestream.twitter.com/%s/site.json' \
% self.streamer.api_version
self.streamer._request(url, params=params) | ['def', 'site', '(', 'self', ',', '*', '*', 'params', ')', ':', 'url', '=', "'https://sitestream.twitter.com/%s/site.json'", '%', 'self', '.', 'streamer', '.', 'api_version', 'self', '.', 'streamer', '.', '_request', '(', 'url', ',', 'params', '=', 'params', ')'] | Stream site
Accepted params found at:
https://dev.twitter.com/docs/api/1.1/get/site | ['Stream', 'site'] | train | https://github.com/ryanmcgrath/twython/blob/7366de80efcbbdfaf615d3f1fea72546196916fc/twython/streaming/types.py#L33-L41 |
89 | rsinger86/drf-flex-fields | rest_flex_fields/serializers.py | FlexFieldsSerializerMixin._get_expand_input | def _get_expand_input(self, passed_settings):
"""
If expand value is explicitliy passed, just return it.
If parsing from request, ensure that the value complies with
the "permitted_expands" list passed into the context from the
FlexFieldsMixin.
"""
... | python | def _get_expand_input(self, passed_settings):
"""
If expand value is explicitliy passed, just return it.
If parsing from request, ensure that the value complies with
the "permitted_expands" list passed into the context from the
FlexFieldsMixin.
"""
... | ['def', '_get_expand_input', '(', 'self', ',', 'passed_settings', ')', ':', 'value', '=', 'passed_settings', '.', 'get', '(', '"expand"', ')', 'if', 'len', '(', 'value', ')', '>', '0', ':', 'return', 'value', 'if', 'not', 'self', '.', '_can_access_request', ':', 'return', '[', ']', 'expand', '=', 'self', '.', '_parse_r... | If expand value is explicitliy passed, just return it.
If parsing from request, ensure that the value complies with
the "permitted_expands" list passed into the context from the
FlexFieldsMixin. | ['If', 'expand', 'value', 'is', 'explicitliy', 'passed', 'just', 'return', 'it', '.', 'If', 'parsing', 'from', 'request', 'ensure', 'that', 'the', 'value', 'complies', 'with', 'the', 'permitted_expands', 'list', 'passed', 'into', 'the', 'context', 'from', 'the', 'FlexFieldsMixin', '.'] | train | https://github.com/rsinger86/drf-flex-fields/blob/56495f15977d76697972acac571792e8fd67003d/rest_flex_fields/serializers.py#L196-L221 |
90 | eyurtsev/FlowCytometryTools | FlowCytometryTools/gui/fc_widget.py | BaseVertex.update_coordinates | def update_coordinates(self, new_coordinates):
"""
new_coordinates : dict
"""
for k, v in new_coordinates.items():
if k in self.coordinates:
self.coordinates[k] = v
for svertex in self.spawn_list:
verts = tuple([self.coordinates.get(ch, No... | python | def update_coordinates(self, new_coordinates):
"""
new_coordinates : dict
"""
for k, v in new_coordinates.items():
if k in self.coordinates:
self.coordinates[k] = v
for svertex in self.spawn_list:
verts = tuple([self.coordinates.get(ch, No... | ['def', 'update_coordinates', '(', 'self', ',', 'new_coordinates', ')', ':', 'for', 'k', ',', 'v', 'in', 'new_coordinates', '.', 'items', '(', ')', ':', 'if', 'k', 'in', 'self', '.', 'coordinates', ':', 'self', '.', 'coordinates', '[', 'k', ']', '=', 'v', 'for', 'svertex', 'in', 'self', '.', 'spawn_list', ':', 'verts',... | new_coordinates : dict | ['new_coordinates', ':', 'dict'] | train | https://github.com/eyurtsev/FlowCytometryTools/blob/4355632508b875273d68c7e2972c17668bcf7b40/FlowCytometryTools/gui/fc_widget.py#L150-L165 |
91 | abelcarreras/DynaPhoPy | dynaphopy/analysis/fitting/fitting_functions.py | Gaussian_function._function | def _function(self, x, a, b, c, d):
"""Gaussian PDF function
x: coordinate
a: peak position
b: deviation (sigma)
c: area proportional parameter
d: base line
"""
return c/b*np.sqrt(2*np.pi)*np.exp(-(x-a)**2/(2*b**2))+d | python | def _function(self, x, a, b, c, d):
"""Gaussian PDF function
x: coordinate
a: peak position
b: deviation (sigma)
c: area proportional parameter
d: base line
"""
return c/b*np.sqrt(2*np.pi)*np.exp(-(x-a)**2/(2*b**2))+d | ['def', '_function', '(', 'self', ',', 'x', ',', 'a', ',', 'b', ',', 'c', ',', 'd', ')', ':', 'return', 'c', '/', 'b', '*', 'np', '.', 'sqrt', '(', '2', '*', 'np', '.', 'pi', ')', '*', 'np', '.', 'exp', '(', '-', '(', 'x', '-', 'a', ')', '**', '2', '/', '(', '2', '*', 'b', '**', '2', ')', ')', '+', 'd'] | Gaussian PDF function
x: coordinate
a: peak position
b: deviation (sigma)
c: area proportional parameter
d: base line | ['Gaussian', 'PDF', 'function', 'x', ':', 'coordinate', 'a', ':', 'peak', 'position', 'b', ':', 'deviation', '(', 'sigma', ')', 'c', ':', 'area', 'proportional', 'parameter', 'd', ':', 'base', 'line'] | train | https://github.com/abelcarreras/DynaPhoPy/blob/51e99422228e6be84830d659b88a0ca904d9136f/dynaphopy/analysis/fitting/fitting_functions.py#L302-L310 |
92 | woolfson-group/isambard | isambard/optimisation/evo_optimizers.py | CMAES._initial_individual | def _initial_individual(self):
"""Generates an individual with random parameters within bounds."""
ind = creator.Individual(
[random.uniform(-1, 1)
for _ in range(len(self.value_means))])
return ind | python | def _initial_individual(self):
"""Generates an individual with random parameters within bounds."""
ind = creator.Individual(
[random.uniform(-1, 1)
for _ in range(len(self.value_means))])
return ind | ['def', '_initial_individual', '(', 'self', ')', ':', 'ind', '=', 'creator', '.', 'Individual', '(', '[', 'random', '.', 'uniform', '(', '-', '1', ',', '1', ')', 'for', '_', 'in', 'range', '(', 'len', '(', 'self', '.', 'value_means', ')', ')', ']', ')', 'return', 'ind'] | Generates an individual with random parameters within bounds. | ['Generates', 'an', 'individual', 'with', 'random', 'parameters', 'within', 'bounds', '.'] | train | https://github.com/woolfson-group/isambard/blob/ebc33b48a28ad217e18f93b910dfba46e6e71e07/isambard/optimisation/evo_optimizers.py#L472-L477 |
93 | yfpeng/bioc | bioc/biocxml/encoder.py | BioCXMLDocumentWriter.write_collection_info | def write_collection_info(self, collection: BioCCollection):
"""
Writes the collection information: encoding, version, DTD, source, date, key, infons, etc.
"""
elem = etree.Element('source')
elem.text = collection.source
self.__writer.send(elem)
elem = et... | python | def write_collection_info(self, collection: BioCCollection):
"""
Writes the collection information: encoding, version, DTD, source, date, key, infons, etc.
"""
elem = etree.Element('source')
elem.text = collection.source
self.__writer.send(elem)
elem = et... | ['def', 'write_collection_info', '(', 'self', ',', 'collection', ':', 'BioCCollection', ')', ':', 'elem', '=', 'etree', '.', 'Element', '(', "'source'", ')', 'elem', '.', 'text', '=', 'collection', '.', 'source', 'self', '.', '__writer', '.', 'send', '(', 'elem', ')', 'elem', '=', 'etree', '.', 'Element', '(', "'date'"... | Writes the collection information: encoding, version, DTD, source, date, key, infons, etc. | ['Writes', 'the', 'collection', 'information', ':', 'encoding', 'version', 'DTD', 'source', 'date', 'key', 'infons', 'etc', '.'] | train | https://github.com/yfpeng/bioc/blob/47ddaa010960d9ba673aefe068e7bbaf39f0fff4/bioc/biocxml/encoder.py#L185-L204 |
94 | pymc-devs/pymc | pymc/Matplot.py | autocorrelation | def autocorrelation(
data, name, maxlags=100, format='png', reflected=False, suffix='-acf', path='./',
fontmap=None, new=True, last=True, rows=1, columns=1, num=1, verbose=1):
"""
Generate bar plot of the autocorrelation function for a series (usually an MCMC trace).
:Arguments:
data: P... | python | def autocorrelation(
data, name, maxlags=100, format='png', reflected=False, suffix='-acf', path='./',
fontmap=None, new=True, last=True, rows=1, columns=1, num=1, verbose=1):
"""
Generate bar plot of the autocorrelation function for a series (usually an MCMC trace).
:Arguments:
data: P... | ['def', 'autocorrelation', '(', 'data', ',', 'name', ',', 'maxlags', '=', '100', ',', 'format', '=', "'png'", ',', 'reflected', '=', 'False', ',', 'suffix', '=', "'-acf'", ',', 'path', '=', "'./'", ',', 'fontmap', '=', 'None', ',', 'new', '=', 'True', ',', 'last', '=', 'True', ',', 'rows', '=', '1', ',', 'columns', '='... | Generate bar plot of the autocorrelation function for a series (usually an MCMC trace).
:Arguments:
data: PyMC object, trace or array
A trace from an MCMC sample or a PyMC object with one or more traces.
name: string
The name of the object.
maxlags (optional): int
... | ['Generate', 'bar', 'plot', 'of', 'the', 'autocorrelation', 'function', 'for', 'a', 'series', '(', 'usually', 'an', 'MCMC', 'trace', ')', '.'] | train | https://github.com/pymc-devs/pymc/blob/c6e530210bff4c0d7189b35b2c971bc53f93f7cd/pymc/Matplot.py#L899-L997 |
95 | pywbem/pywbem | pywbem/cim_obj.py | CIMProperty.tocimxml | def tocimxml(self):
"""
Return the CIM-XML representation of this CIM property,
as an object of an appropriate subclass of :term:`Element`.
The returned CIM-XML representation is a `PROPERTY`,
`PROPERTY.REFERENCE`, or `PROPERTY.ARRAY` element dependent on the
property ty... | python | def tocimxml(self):
"""
Return the CIM-XML representation of this CIM property,
as an object of an appropriate subclass of :term:`Element`.
The returned CIM-XML representation is a `PROPERTY`,
`PROPERTY.REFERENCE`, or `PROPERTY.ARRAY` element dependent on the
property ty... | ['def', 'tocimxml', '(', 'self', ')', ':', 'qualifiers', '=', '[', 'q', '.', 'tocimxml', '(', ')', 'for', 'q', 'in', 'self', '.', 'qualifiers', '.', 'values', '(', ')', ']', 'if', 'self', '.', 'is_array', ':', '# pylint: disable=no-else-return', 'assert', 'self', '.', 'type', '!=', "'reference'", 'if', 'self', '.', 'va... | Return the CIM-XML representation of this CIM property,
as an object of an appropriate subclass of :term:`Element`.
The returned CIM-XML representation is a `PROPERTY`,
`PROPERTY.REFERENCE`, or `PROPERTY.ARRAY` element dependent on the
property type, and consistent with :term:`DSP0201`.... | ['Return', 'the', 'CIM', '-', 'XML', 'representation', 'of', 'this', 'CIM', 'property', 'as', 'an', 'object', 'of', 'an', 'appropriate', 'subclass', 'of', ':', 'term', ':', 'Element', '.'] | train | https://github.com/pywbem/pywbem/blob/e54ecb82c2211e289a268567443d60fdd489f1e4/pywbem/cim_obj.py#L4883-L4967 |
96 | rocky/python-spark | spark_parser/spark.py | GenericParser.get_profile_info | def get_profile_info(self):
"""Show the accumulated results of how many times each rule was used"""
return sorted(self.profile_info.items(),
key=lambda kv: kv[1],
reverse=False)
return | python | def get_profile_info(self):
"""Show the accumulated results of how many times each rule was used"""
return sorted(self.profile_info.items(),
key=lambda kv: kv[1],
reverse=False)
return | ['def', 'get_profile_info', '(', 'self', ')', ':', 'return', 'sorted', '(', 'self', '.', 'profile_info', '.', 'items', '(', ')', ',', 'key', '=', 'lambda', 'kv', ':', 'kv', '[', '1', ']', ',', 'reverse', '=', 'False', ')', 'return'] | Show the accumulated results of how many times each rule was used | ['Show', 'the', 'accumulated', 'results', 'of', 'how', 'many', 'times', 'each', 'rule', 'was', 'used'] | train | https://github.com/rocky/python-spark/blob/8899954bcf0e166726841a43e87c23790eb3441f/spark_parser/spark.py#L983-L988 |
97 | scivision/pymap3d | pymap3d/sidereal.py | juliandate | def juliandate(time: datetime) -> float:
"""
Python datetime to Julian time
from D.Vallado Fundamentals of Astrodynamics and Applications p.187
and J. Meeus Astronomical Algorithms 1991 Eqn. 7.1 pg. 61
Parameters
----------
time : datetime.datetime
time to convert
Results
... | python | def juliandate(time: datetime) -> float:
"""
Python datetime to Julian time
from D.Vallado Fundamentals of Astrodynamics and Applications p.187
and J. Meeus Astronomical Algorithms 1991 Eqn. 7.1 pg. 61
Parameters
----------
time : datetime.datetime
time to convert
Results
... | ['def', 'juliandate', '(', 'time', ':', 'datetime', ')', '->', 'float', ':', 'times', '=', 'np', '.', 'atleast_1d', '(', 'time', ')', 'assert', 'times', '.', 'ndim', '==', '1', 'jd', '=', 'np', '.', 'empty', '(', 'times', '.', 'size', ')', 'for', 'i', ',', 't', 'in', 'enumerate', '(', 'times', ')', ':', 'if', 't', '.',... | Python datetime to Julian time
from D.Vallado Fundamentals of Astrodynamics and Applications p.187
and J. Meeus Astronomical Algorithms 1991 Eqn. 7.1 pg. 61
Parameters
----------
time : datetime.datetime
time to convert
Results
-------
jd : float
Julian date | ['Python', 'datetime', 'to', 'Julian', 'time'] | train | https://github.com/scivision/pymap3d/blob/c9cf676594611cdb52ff7e0eca6388c80ed4f63f/pymap3d/sidereal.py#L58-L97 |
98 | Phylliade/ikpy | contrib/transformations.py | pose_to_list | def pose_to_list(pose):
"""
Convert a Pose or PoseStamped in Python list ((position), (quaternion))
:param pose: geometry_msgs.msg.PoseStamped or geometry_msgs.msg.Pose
:return: the equivalent in list ((position), (quaternion))
"""
if type(pose) == geometry_msgs.msg.PoseStamped:
return [... | python | def pose_to_list(pose):
"""
Convert a Pose or PoseStamped in Python list ((position), (quaternion))
:param pose: geometry_msgs.msg.PoseStamped or geometry_msgs.msg.Pose
:return: the equivalent in list ((position), (quaternion))
"""
if type(pose) == geometry_msgs.msg.PoseStamped:
return [... | ['def', 'pose_to_list', '(', 'pose', ')', ':', 'if', 'type', '(', 'pose', ')', '==', 'geometry_msgs', '.', 'msg', '.', 'PoseStamped', ':', 'return', '[', '[', 'pose', '.', 'pose', '.', 'position', '.', 'x', ',', 'pose', '.', 'pose', '.', 'position', '.', 'y', ',', 'pose', '.', 'pose', '.', 'position', '.', 'z', ']', ',... | Convert a Pose or PoseStamped in Python list ((position), (quaternion))
:param pose: geometry_msgs.msg.PoseStamped or geometry_msgs.msg.Pose
:return: the equivalent in list ((position), (quaternion)) | ['Convert', 'a', 'Pose', 'or', 'PoseStamped', 'in', 'Python', 'list', '((', 'position', ')', '(', 'quaternion', '))', ':', 'param', 'pose', ':', 'geometry_msgs', '.', 'msg', '.', 'PoseStamped', 'or', 'geometry_msgs', '.', 'msg', '.', 'Pose', ':', 'return', ':', 'the', 'equivalent', 'in', 'list', '((', 'position', ')', ... | train | https://github.com/Phylliade/ikpy/blob/60e36d6163136942bf520d952db17123c658d0b6/contrib/transformations.py#L48-L68 |
99 | ARMmbed/mbed-cloud-sdk-python | src/mbed_cloud/device_directory/device_directory.py | DeviceDirectoryAPI.list_device_events | def list_device_events(self, **kwargs):
"""List all device logs.
:param int limit: The number of logs to retrieve.
:param str order: The ordering direction, ascending (asc) or
descending (desc)
:param str after: Get logs after/starting at given `device_event_id`
:par... | python | def list_device_events(self, **kwargs):
"""List all device logs.
:param int limit: The number of logs to retrieve.
:param str order: The ordering direction, ascending (asc) or
descending (desc)
:param str after: Get logs after/starting at given `device_event_id`
:par... | ['def', 'list_device_events', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'kwargs', '=', 'self', '.', '_verify_sort_options', '(', 'kwargs', ')', 'kwargs', '=', 'self', '.', '_verify_filters', '(', 'kwargs', ',', 'DeviceEvent', ',', 'True', ')', 'api', '=', 'self', '.', '_get_api', '(', 'device_directory', '.', 'De... | List all device logs.
:param int limit: The number of logs to retrieve.
:param str order: The ordering direction, ascending (asc) or
descending (desc)
:param str after: Get logs after/starting at given `device_event_id`
:param dict filters: Dictionary of filters to apply.
... | ['List', 'all', 'device', 'logs', '.'] | train | https://github.com/ARMmbed/mbed-cloud-sdk-python/blob/c0af86fb2cdd4dc7ed26f236139241067d293509/src/mbed_cloud/device_directory/device_directory.py#L290-L305 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.