Unnamed: 0
int64
0
10k
repository_name
stringlengths
7
54
func_path_in_repository
stringlengths
5
223
func_name
stringlengths
1
134
whole_func_string
stringlengths
100
30.3k
language
stringclasses
1 value
func_code_string
stringlengths
100
30.3k
func_code_tokens
stringlengths
138
33.2k
func_documentation_string
stringlengths
1
15k
func_documentation_tokens
stringlengths
5
5.14k
split_name
stringclasses
1 value
func_code_url
stringlengths
91
315
0
getsentry/libsourcemap
libsourcemap/highlevel.py
View.get_original_function_name
def get_original_function_name(self, line, col, minified_name, minified_source): """Given a token location and a minified function name and the minified source file this returns the original function name if it can be found of the minified function in scope. ...
python
def get_original_function_name(self, line, col, minified_name, minified_source): """Given a token location and a minified function name and the minified source file this returns the original function name if it can be found of the minified function in scope. ...
['def', 'get_original_function_name', '(', 'self', ',', 'line', ',', 'col', ',', 'minified_name', ',', 'minified_source', ')', ':', '# Silently ignore underflows', 'if', 'line', '<', '0', 'or', 'col', '<', '0', ':', 'return', 'None', 'minified_name', '=', 'minified_name', '.', 'encode', '(', "'utf-8'", ')', 'sout', '='...
Given a token location and a minified function name and the minified source file this returns the original function name if it can be found of the minified function in scope.
['Given', 'a', 'token', 'location', 'and', 'a', 'minified', 'function', 'name', 'and', 'the', 'minified', 'source', 'file', 'this', 'returns', 'the', 'original', 'function', 'name', 'if', 'it', 'can', 'be', 'found', 'of', 'the', 'minified', 'function', 'in', 'scope', '.']
train
https://github.com/getsentry/libsourcemap/blob/94b5a34814fafee9dc23da8ec0ccca77f30e3370/libsourcemap/highlevel.py#L163-L185
1
brocade/pynos
pynos/versions/ver_7/ver_7_1_0/yang/brocade_mac_address_table.py
brocade_mac_address_table.get_mac_address_table_input_request_type_get_interface_based_request_mac_type
def get_mac_address_table_input_request_type_get_interface_based_request_mac_type(self, **kwargs): """Auto Generated Code """ config = ET.Element("config") get_mac_address_table = ET.Element("get_mac_address_table") config = get_mac_address_table input = ET.SubElement(get...
python
def get_mac_address_table_input_request_type_get_interface_based_request_mac_type(self, **kwargs): """Auto Generated Code """ config = ET.Element("config") get_mac_address_table = ET.Element("get_mac_address_table") config = get_mac_address_table input = ET.SubElement(get...
['def', 'get_mac_address_table_input_request_type_get_interface_based_request_mac_type', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'config', '=', 'ET', '.', 'Element', '(', '"config"', ')', 'get_mac_address_table', '=', 'ET', '.', 'Element', '(', '"get_mac_address_table"', ')', 'config', '=', 'get_mac_address_tab...
Auto Generated Code
['Auto', 'Generated', 'Code']
train
https://github.com/brocade/pynos/blob/bd8a34e98f322de3fc06750827d8bbc3a0c00380/pynos/versions/ver_7/ver_7_1_0/yang/brocade_mac_address_table.py#L297-L310
2
nicolargo/glances
glances/plugins/glances_memswap.py
Plugin.update_views
def update_views(self): """Update stats views.""" # Call the father's method super(Plugin, self).update_views() # Add specifics informations # Alert and log self.views['used']['decoration'] = self.get_alert_log(self.stats['used'], maximum=self.stats['total'])
python
def update_views(self): """Update stats views.""" # Call the father's method super(Plugin, self).update_views() # Add specifics informations # Alert and log self.views['used']['decoration'] = self.get_alert_log(self.stats['used'], maximum=self.stats['total'])
['def', 'update_views', '(', 'self', ')', ':', "# Call the father's method", 'super', '(', 'Plugin', ',', 'self', ')', '.', 'update_views', '(', ')', '# Add specifics informations', '# Alert and log', 'self', '.', 'views', '[', "'used'", ']', '[', "'decoration'", ']', '=', 'self', '.', 'get_alert_log', '(', 'self', '.'...
Update stats views.
['Update', 'stats', 'views', '.']
train
https://github.com/nicolargo/glances/blob/5bd4d587a736e0d2b03170b56926841d2a3eb7ee/glances/plugins/glances_memswap.py#L130-L137
3
DLR-RM/RAFCON
source/rafcon/gui/helpers/meta_data.py
contains_geometric_info
def contains_geometric_info(var): """ Check whether the passed variable is a tuple with two floats or integers """ return isinstance(var, tuple) and len(var) == 2 and all(isinstance(val, (int, float)) for val in var)
python
def contains_geometric_info(var): """ Check whether the passed variable is a tuple with two floats or integers """ return isinstance(var, tuple) and len(var) == 2 and all(isinstance(val, (int, float)) for val in var)
['def', 'contains_geometric_info', '(', 'var', ')', ':', 'return', 'isinstance', '(', 'var', ',', 'tuple', ')', 'and', 'len', '(', 'var', ')', '==', '2', 'and', 'all', '(', 'isinstance', '(', 'val', ',', '(', 'int', ',', 'float', ')', ')', 'for', 'val', 'in', 'var', ')']
Check whether the passed variable is a tuple with two floats or integers
['Check', 'whether', 'the', 'passed', 'variable', 'is', 'a', 'tuple', 'with', 'two', 'floats', 'or', 'integers']
train
https://github.com/DLR-RM/RAFCON/blob/24942ef1a904531f49ab8830a1dbb604441be498/source/rafcon/gui/helpers/meta_data.py#L55-L57
4
ejeschke/ginga
ginga/gtk3w/ImageViewGtk.py
ImageViewGtk.save_plain_image_as_file
def save_plain_image_as_file(self, filepath, format='png', quality=90): """Used for generating thumbnails. Does not include overlaid graphics. """ pixbuf = self.get_plain_image_as_pixbuf() options, values = [], [] if format == 'jpeg': options.append('quality'...
python
def save_plain_image_as_file(self, filepath, format='png', quality=90): """Used for generating thumbnails. Does not include overlaid graphics. """ pixbuf = self.get_plain_image_as_pixbuf() options, values = [], [] if format == 'jpeg': options.append('quality'...
['def', 'save_plain_image_as_file', '(', 'self', ',', 'filepath', ',', 'format', '=', "'png'", ',', 'quality', '=', '90', ')', ':', 'pixbuf', '=', 'self', '.', 'get_plain_image_as_pixbuf', '(', ')', 'options', ',', 'values', '=', '[', ']', ',', '[', ']', 'if', 'format', '==', "'jpeg'", ':', 'options', '.', 'append', '(...
Used for generating thumbnails. Does not include overlaid graphics.
['Used', 'for', 'generating', 'thumbnails', '.', 'Does', 'not', 'include', 'overlaid', 'graphics', '.']
train
https://github.com/ejeschke/ginga/blob/a78c893ec6f37a837de851947e9bb4625c597915/ginga/gtk3w/ImageViewGtk.py#L75-L84
5
poppy-project/pypot
pypot/vrep/remoteApiBindings/vrep.py
simxClearFloatSignal
def simxClearFloatSignal(clientID, signalName, operationMode): ''' Please have a look at the function description/documentation in the V-REP user manual ''' if (sys.version_info[0] == 3) and (type(signalName) is str): signalName=signalName.encode('utf-8') return c_ClearFloatSignal(clientID,...
python
def simxClearFloatSignal(clientID, signalName, operationMode): ''' Please have a look at the function description/documentation in the V-REP user manual ''' if (sys.version_info[0] == 3) and (type(signalName) is str): signalName=signalName.encode('utf-8') return c_ClearFloatSignal(clientID,...
['def', 'simxClearFloatSignal', '(', 'clientID', ',', 'signalName', ',', 'operationMode', ')', ':', 'if', '(', 'sys', '.', 'version_info', '[', '0', ']', '==', '3', ')', 'and', '(', 'type', '(', 'signalName', ')', 'is', 'str', ')', ':', 'signalName', '=', 'signalName', '.', 'encode', '(', "'utf-8'", ')', 'return', 'c_C...
Please have a look at the function description/documentation in the V-REP user manual
['Please', 'have', 'a', 'look', 'at', 'the', 'function', 'description', '/', 'documentation', 'in', 'the', 'V', '-', 'REP', 'user', 'manual']
train
https://github.com/poppy-project/pypot/blob/d9c6551bbc87d45d9d1f0bc15e35b616d0002afd/pypot/vrep/remoteApiBindings/vrep.py#L900-L907
6
cloudify-cosmo/repex
repex.py
Repex.find_matches
def find_matches(self, content, file_to_handle): """Find all matches of an expression in a file """ # look for all match groups in the content groups = [match.groupdict() for match in self.match_expression.finditer(content)] # filter out content not in the match...
python
def find_matches(self, content, file_to_handle): """Find all matches of an expression in a file """ # look for all match groups in the content groups = [match.groupdict() for match in self.match_expression.finditer(content)] # filter out content not in the match...
['def', 'find_matches', '(', 'self', ',', 'content', ',', 'file_to_handle', ')', ':', '# look for all match groups in the content', 'groups', '=', '[', 'match', '.', 'groupdict', '(', ')', 'for', 'match', 'in', 'self', '.', 'match_expression', '.', 'finditer', '(', 'content', ')', ']', '# filter out content not in the ...
Find all matches of an expression in a file
['Find', 'all', 'matches', 'of', 'an', 'expression', 'in', 'a', 'file']
train
https://github.com/cloudify-cosmo/repex/blob/589e442857fa4a99fa88670d7df1a72f983bbd28/repex.py#L605-L618
7
mariocj89/github-token
github_token/__init__.py
TokenFactory.create
def create(self): """Creates a token It uses the app_name as the notes and the scopes are the permissions required by the application. See those in github when configuring an app token Raises a TFARequired if a two factor is required after the atempt to create it withou...
python
def create(self): """Creates a token It uses the app_name as the notes and the scopes are the permissions required by the application. See those in github when configuring an app token Raises a TFARequired if a two factor is required after the atempt to create it withou...
['def', 'create', '(', 'self', ')', ':', 'headers', '=', 'dict', '(', ')', 'if', 'self', '.', 'tfa_token', ':', 'headers', '[', '"X-GitHub-OTP"', ']', '=', 'self', '.', 'tfa_token', 'token_name', '=', 'self', '.', 'app_name', '+', 'platform', '.', 'node', '(', ')', '# node specific in case the user has multiple hosts',...
Creates a token It uses the app_name as the notes and the scopes are the permissions required by the application. See those in github when configuring an app token Raises a TFARequired if a two factor is required after the atempt to create it without having call tfa before
['Creates', 'a', 'token']
train
https://github.com/mariocj89/github-token/blob/8ca85fa51a52aef94cfb4f851eb229ee500bc28f/github_token/__init__.py#L81-L108
8
brycedrennan/eulerian-magnification
eulerian_magnification/base.py
combine_pyramid_and_save
def combine_pyramid_and_save(g_video, orig_video, enlarge_multiple, fps, save_filename='media/output.avi'): """Combine a gaussian video representation with the original and save to file""" width, height = get_frame_dimensions(orig_video[0]) fourcc = cv2.VideoWriter_fourcc(*'MJPG') print("Outputting to %...
python
def combine_pyramid_and_save(g_video, orig_video, enlarge_multiple, fps, save_filename='media/output.avi'): """Combine a gaussian video representation with the original and save to file""" width, height = get_frame_dimensions(orig_video[0]) fourcc = cv2.VideoWriter_fourcc(*'MJPG') print("Outputting to %...
['def', 'combine_pyramid_and_save', '(', 'g_video', ',', 'orig_video', ',', 'enlarge_multiple', ',', 'fps', ',', 'save_filename', '=', "'media/output.avi'", ')', ':', 'width', ',', 'height', '=', 'get_frame_dimensions', '(', 'orig_video', '[', '0', ']', ')', 'fourcc', '=', 'cv2', '.', 'VideoWriter_fourcc', '(', '*', "'...
Combine a gaussian video representation with the original and save to file
['Combine', 'a', 'gaussian', 'video', 'representation', 'with', 'the', 'original', 'and', 'save', 'to', 'file']
train
https://github.com/brycedrennan/eulerian-magnification/blob/9ae0651fe3334176300d183f8240ad36d77759a9/eulerian_magnification/base.py#L108-L122
9
PmagPy/PmagPy
pmagpy/pmag.py
PintPars
def PintPars(datablock, araiblock, zijdblock, start, end, accept, **kwargs): """ calculate the paleointensity magic parameters make some definitions """ if 'version' in list(kwargs.keys()) and kwargs['version'] == 3: meth_key = 'method_codes' beta_key = 'int_b_beta' temp_key, m...
python
def PintPars(datablock, araiblock, zijdblock, start, end, accept, **kwargs): """ calculate the paleointensity magic parameters make some definitions """ if 'version' in list(kwargs.keys()) and kwargs['version'] == 3: meth_key = 'method_codes' beta_key = 'int_b_beta' temp_key, m...
['def', 'PintPars', '(', 'datablock', ',', 'araiblock', ',', 'zijdblock', ',', 'start', ',', 'end', ',', 'accept', ',', '*', '*', 'kwargs', ')', ':', 'if', "'version'", 'in', 'list', '(', 'kwargs', '.', 'keys', '(', ')', ')', 'and', 'kwargs', '[', "'version'", ']', '==', '3', ':', 'meth_key', '=', "'method_codes'", 'be...
calculate the paleointensity magic parameters make some definitions
['calculate', 'the', 'paleointensity', 'magic', 'parameters', 'make', 'some', 'definitions']
train
https://github.com/PmagPy/PmagPy/blob/c7984f8809bf40fe112e53dcc311a33293b62d0b/pmagpy/pmag.py#L2827-L3374
10
lowandrew/OLCTools
spadespipeline/vtyper.py
Vtyper.epcrparse
def epcrparse(self): """ Parse the ePCR text file outputs """ logging.info('Parsing ePCR results') for sample in self.metadata: if sample.general.bestassemblyfile != 'NA': if 'stx' in sample.general.datastore: # Initialise count - t...
python
def epcrparse(self): """ Parse the ePCR text file outputs """ logging.info('Parsing ePCR results') for sample in self.metadata: if sample.general.bestassemblyfile != 'NA': if 'stx' in sample.general.datastore: # Initialise count - t...
['def', 'epcrparse', '(', 'self', ')', ':', 'logging', '.', 'info', '(', "'Parsing ePCR results'", ')', 'for', 'sample', 'in', 'self', '.', 'metadata', ':', 'if', 'sample', '.', 'general', '.', 'bestassemblyfile', '!=', "'NA'", ':', 'if', "'stx'", 'in', 'sample', '.', 'general', '.', 'datastore', ':', '# Initialise cou...
Parse the ePCR text file outputs
['Parse', 'the', 'ePCR', 'text', 'file', 'outputs']
train
https://github.com/lowandrew/OLCTools/blob/88aa90ac85f84d0bbeb03e43c29b0a9d36e4ce2a/spadespipeline/vtyper.py#L96-L131
11
atztogo/phonopy
phonopy/structure/spglib.py
get_pointgroup
def get_pointgroup(rotations): """Return point group in international table symbol and number. The symbols are mapped to the numbers as follows: 1 "1 " 2 "-1 " 3 "2 " 4 "m " 5 "2/m " 6 "222 " 7 "mm2 " 8 "mmm " 9 "4 " 10 "-4 " 11...
python
def get_pointgroup(rotations): """Return point group in international table symbol and number. The symbols are mapped to the numbers as follows: 1 "1 " 2 "-1 " 3 "2 " 4 "m " 5 "2/m " 6 "222 " 7 "mm2 " 8 "mmm " 9 "4 " 10 "-4 " 11...
['def', 'get_pointgroup', '(', 'rotations', ')', ':', '_set_no_error', '(', ')', '# (symbol, pointgroup_number, transformation_matrix)', 'pointgroup', '=', 'spg', '.', 'pointgroup', '(', 'np', '.', 'array', '(', 'rotations', ',', 'dtype', '=', "'intc'", ',', 'order', '=', "'C'", ')', ')', '_set_error_message', '(', ')'...
Return point group in international table symbol and number. The symbols are mapped to the numbers as follows: 1 "1 " 2 "-1 " 3 "2 " 4 "m " 5 "2/m " 6 "222 " 7 "mm2 " 8 "mmm " 9 "4 " 10 "-4 " 11 "4/m " 12 "422 " 13 "4mm...
['Return', 'point', 'group', 'in', 'international', 'table', 'symbol', 'and', 'number', '.']
train
https://github.com/atztogo/phonopy/blob/869cc2ba9e7d495d5f4cf6942415ab3fc9e2a10f/phonopy/structure/spglib.py#L301-L343
12
automl/HpBandSter
hpbandster/core/nameserver.py
NameServer.start
def start(self): """ starts a Pyro4 nameserver in a separate thread Returns ------- tuple (str, int): the host name and the used port """ if self.host is None: if self.nic_name is None: self.host = 'localhost' else: self.host = nic_name_to_host(self.nic_name) uri, self.pyro_ns,...
python
def start(self): """ starts a Pyro4 nameserver in a separate thread Returns ------- tuple (str, int): the host name and the used port """ if self.host is None: if self.nic_name is None: self.host = 'localhost' else: self.host = nic_name_to_host(self.nic_name) uri, self.pyro_ns,...
['def', 'start', '(', 'self', ')', ':', 'if', 'self', '.', 'host', 'is', 'None', ':', 'if', 'self', '.', 'nic_name', 'is', 'None', ':', 'self', '.', 'host', '=', "'localhost'", 'else', ':', 'self', '.', 'host', '=', 'nic_name_to_host', '(', 'self', '.', 'nic_name', ')', 'uri', ',', 'self', '.', 'pyro_ns', ',', '_', '='...
starts a Pyro4 nameserver in a separate thread Returns ------- tuple (str, int): the host name and the used port
['starts', 'a', 'Pyro4', 'nameserver', 'in', 'a', 'separate', 'thread', 'Returns', '-------', 'tuple', '(', 'str', 'int', ')', ':', 'the', 'host', 'name', 'and', 'the', 'used', 'port']
train
https://github.com/automl/HpBandSter/blob/841db4b827f342e5eb7f725723ea6461ac52d45a/hpbandster/core/nameserver.py#L48-L79
13
PixelwarStudio/PyTree
Tree/core.py
Tree.grow
def grow(self, times=1): """Let the tree grow. Args: times (integer): Indicate how many times the tree will grow. """ self.nodes.append([]) for n, node in enumerate(self.nodes[self.age]): if self.age == 0: p_node = Node(self.pos[:2]) ...
python
def grow(self, times=1): """Let the tree grow. Args: times (integer): Indicate how many times the tree will grow. """ self.nodes.append([]) for n, node in enumerate(self.nodes[self.age]): if self.age == 0: p_node = Node(self.pos[:2]) ...
['def', 'grow', '(', 'self', ',', 'times', '=', '1', ')', ':', 'self', '.', 'nodes', '.', 'append', '(', '[', ']', ')', 'for', 'n', ',', 'node', 'in', 'enumerate', '(', 'self', '.', 'nodes', '[', 'self', '.', 'age', ']', ')', ':', 'if', 'self', '.', 'age', '==', '0', ':', 'p_node', '=', 'Node', '(', 'self', '.', 'pos',...
Let the tree grow. Args: times (integer): Indicate how many times the tree will grow.
['Let', 'the', 'tree', 'grow', '.']
train
https://github.com/PixelwarStudio/PyTree/blob/f14b25ea145da6b00d836e34251d2a4c823766dc/Tree/core.py#L167-L189
14
saltstack/salt
salt/modules/osquery.py
kernel_integrity
def kernel_integrity(attrs=None, where=None): ''' Return kernel_integrity information from osquery CLI Example: .. code-block:: bash salt '*' osquery.kernel_integrity ''' if __grains__['os_family'] in ['RedHat', 'Debian']: return _osquery_cmd(table='kernel_integrity', attrs=at...
python
def kernel_integrity(attrs=None, where=None): ''' Return kernel_integrity information from osquery CLI Example: .. code-block:: bash salt '*' osquery.kernel_integrity ''' if __grains__['os_family'] in ['RedHat', 'Debian']: return _osquery_cmd(table='kernel_integrity', attrs=at...
['def', 'kernel_integrity', '(', 'attrs', '=', 'None', ',', 'where', '=', 'None', ')', ':', 'if', '__grains__', '[', "'os_family'", ']', 'in', '[', "'RedHat'", ',', "'Debian'", ']', ':', 'return', '_osquery_cmd', '(', 'table', '=', "'kernel_integrity'", ',', 'attrs', '=', 'attrs', ',', 'where', '=', 'where', ')', 'retu...
Return kernel_integrity information from osquery CLI Example: .. code-block:: bash salt '*' osquery.kernel_integrity
['Return', 'kernel_integrity', 'information', 'from', 'osquery']
train
https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/osquery.py#L149-L161
15
rackerlabs/lambda-uploader
lambda_uploader/package.py
Package.virtualenv
def virtualenv(self, virtualenv): ''' Sets the virtual environment for the lambda package If this is not set then package_dependencies will create a new one. Takes a path to a virtualenv or a boolean if the virtualenv creation should be skipped. ''' # If a boole...
python
def virtualenv(self, virtualenv): ''' Sets the virtual environment for the lambda package If this is not set then package_dependencies will create a new one. Takes a path to a virtualenv or a boolean if the virtualenv creation should be skipped. ''' # If a boole...
['def', 'virtualenv', '(', 'self', ',', 'virtualenv', ')', ':', '# If a boolean is passed then set the internal _skip_virtualenv flag', 'if', 'isinstance', '(', 'virtualenv', ',', 'bool', ')', ':', 'self', '.', '_skip_virtualenv', '=', 'virtualenv', 'else', ':', 'self', '.', '_virtualenv', '=', 'virtualenv', 'if', 'not...
Sets the virtual environment for the lambda package If this is not set then package_dependencies will create a new one. Takes a path to a virtualenv or a boolean if the virtualenv creation should be skipped.
['Sets', 'the', 'virtual', 'environment', 'for', 'the', 'lambda', 'package']
train
https://github.com/rackerlabs/lambda-uploader/blob/a5036e60d45d1a4fdc07df071f5b6e3b113388d4/lambda_uploader/package.py#L114-L133
16
agoragames/kairos
kairos/cassandra_backend.py
CassandraSet._insert_stmt
def _insert_stmt(self, name, value, timestamp, interval, config): '''Helper to generate the insert statement.''' # Calculate the TTL and abort if inserting into the past expire, ttl = config['expire'], config['ttl'](timestamp) if expire and not ttl: return None i_time = config['i_calc'].to_bu...
python
def _insert_stmt(self, name, value, timestamp, interval, config): '''Helper to generate the insert statement.''' # Calculate the TTL and abort if inserting into the past expire, ttl = config['expire'], config['ttl'](timestamp) if expire and not ttl: return None i_time = config['i_calc'].to_bu...
['def', '_insert_stmt', '(', 'self', ',', 'name', ',', 'value', ',', 'timestamp', ',', 'interval', ',', 'config', ')', ':', '# Calculate the TTL and abort if inserting into the past', 'expire', ',', 'ttl', '=', 'config', '[', "'expire'", ']', ',', 'config', '[', "'ttl'", ']', '(', 'timestamp', ')', 'if', 'expire', 'and...
Helper to generate the insert statement.
['Helper', 'to', 'generate', 'the', 'insert', 'statement', '.']
train
https://github.com/agoragames/kairos/blob/0b062d543b0f4a46df460fa0eb6ec281232ab179/kairos/cassandra_backend.py#L646-L665
17
pavelsof/ipalint
ipalint/read.py
Reader._determine_dialect
def _determine_dialect(self, lines): """ Expects a non-empty [] of strings; these would normally be the first few lines of a csv file. Returns the most likely Dialect named tuple or None if the data seems to form a single column. Ensures that using the returned dialect, all the lines given will have the sa...
python
def _determine_dialect(self, lines): """ Expects a non-empty [] of strings; these would normally be the first few lines of a csv file. Returns the most likely Dialect named tuple or None if the data seems to form a single column. Ensures that using the returned dialect, all the lines given will have the sa...
['def', '_determine_dialect', '(', 'self', ',', 'lines', ')', ':', 'permuts', '=', '[', '(', 'quotechar', ',', 'escapechar', ')', 'for', 'quotechar', 'in', 'CSV_QUOTECHARS', 'for', 'escapechar', 'in', 'CSV_ESCAPECHARS', ']', 'for', 'delim', 'in', 'CSV_DELIMITERS', ':', 'counts', '=', '[', 'line', '.', 'count', '(', 'de...
Expects a non-empty [] of strings; these would normally be the first few lines of a csv file. Returns the most likely Dialect named tuple or None if the data seems to form a single column. Ensures that using the returned dialect, all the lines given will have the same number of columns. Helper for the get_d...
['Expects', 'a', 'non', '-', 'empty', '[]', 'of', 'strings', ';', 'these', 'would', 'normally', 'be', 'the', 'first', 'few', 'lines', 'of', 'a', 'csv', 'file', '.', 'Returns', 'the', 'most', 'likely', 'Dialect', 'named', 'tuple', 'or', 'None', 'if', 'the', 'data', 'seems', 'to', 'form', 'a', 'single', 'column', '.']
train
https://github.com/pavelsof/ipalint/blob/763e5979ede6980cbfc746b06fd007b379762eeb/ipalint/read.py#L177-L218
18
angr/angr
angr/utils/graph.py
compute_dominance_frontier
def compute_dominance_frontier(graph, domtree): """ Compute a dominance frontier based on the given post-dominator tree. This implementation is based on figure 2 of paper An Efficient Method of Computing Static Single Assignment Form by Ron Cytron, etc. :param graph: The graph where we want to c...
python
def compute_dominance_frontier(graph, domtree): """ Compute a dominance frontier based on the given post-dominator tree. This implementation is based on figure 2 of paper An Efficient Method of Computing Static Single Assignment Form by Ron Cytron, etc. :param graph: The graph where we want to c...
['def', 'compute_dominance_frontier', '(', 'graph', ',', 'domtree', ')', ':', 'df', '=', '{', '}', '# Perform a post-order search on the dominator tree', 'for', 'x', 'in', 'networkx', '.', 'dfs_postorder_nodes', '(', 'domtree', ')', ':', 'if', 'x', 'not', 'in', 'graph', ':', '# Skip nodes that are not in the graph', 'c...
Compute a dominance frontier based on the given post-dominator tree. This implementation is based on figure 2 of paper An Efficient Method of Computing Static Single Assignment Form by Ron Cytron, etc. :param graph: The graph where we want to compute the dominance frontier. :param domtree: The domin...
['Compute', 'a', 'dominance', 'frontier', 'based', 'on', 'the', 'given', 'post', '-', 'dominator', 'tree', '.']
train
https://github.com/angr/angr/blob/4e2f97d56af5419ee73bdb30482c8dd8ff5f3e40/angr/utils/graph.py#L63-L104
19
CI-WATER/mapkit
mapkit/RasterConverter.py
RasterConverter.isNumber
def isNumber(self, value): """ Validate whether a value is a number or not """ try: str(value) float(value) return True except ValueError: return False
python
def isNumber(self, value): """ Validate whether a value is a number or not """ try: str(value) float(value) return True except ValueError: return False
['def', 'isNumber', '(', 'self', ',', 'value', ')', ':', 'try', ':', 'str', '(', 'value', ')', 'float', '(', 'value', ')', 'return', 'True', 'except', 'ValueError', ':', 'return', 'False']
Validate whether a value is a number or not
['Validate', 'whether', 'a', 'value', 'is', 'a', 'number', 'or', 'not']
train
https://github.com/CI-WATER/mapkit/blob/ce5fbded6af7adabdf1eec85631c6811ef8ecc34/mapkit/RasterConverter.py#L1097-L1107
20
CamDavidsonPilon/lifelines
lifelines/utils/__init__.py
survival_table_from_events
def survival_table_from_events( death_times, event_observed, birth_times=None, columns=["removed", "observed", "censored", "entrance", "at_risk"], weights=None, collapse=False, intervals=None, ): # pylint: disable=dangerous-default-value,too-many-locals """ Parameters ----------...
python
def survival_table_from_events( death_times, event_observed, birth_times=None, columns=["removed", "observed", "censored", "entrance", "at_risk"], weights=None, collapse=False, intervals=None, ): # pylint: disable=dangerous-default-value,too-many-locals """ Parameters ----------...
['def', 'survival_table_from_events', '(', 'death_times', ',', 'event_observed', ',', 'birth_times', '=', 'None', ',', 'columns', '=', '[', '"removed"', ',', '"observed"', ',', '"censored"', ',', '"entrance"', ',', '"at_risk"', ']', ',', 'weights', '=', 'None', ',', 'collapse', '=', 'False', ',', 'intervals', '=', 'Non...
Parameters ---------- death_times: (n,) array represent the event times event_observed: (n,) array 1 if observed event, 0 is censored event. birth_times: a (n,) array, optional representing when the subject was first observed. A subject's death event is then at [birth times + duration ...
['Parameters', '----------', 'death_times', ':', '(', 'n', ')', 'array', 'represent', 'the', 'event', 'times', 'event_observed', ':', '(', 'n', ')', 'array', '1', 'if', 'observed', 'event', '0', 'is', 'censored', 'event', '.', 'birth_times', ':', 'a', '(', 'n', ')', 'array', 'optional', 'representing', 'when', 'the', '...
train
https://github.com/CamDavidsonPilon/lifelines/blob/bdf6be6f1d10eea4c46365ee0ee6a47d8c30edf8/lifelines/utils/__init__.py#L262-L361
21
django-blog-zinnia/cmsplugin-zinnia
cmsplugin_zinnia/cms_plugins.py
CMSRandomEntriesPlugin.render
def render(self, context, instance, placeholder): """ Update the context with plugin's data """ context = super(CMSRandomEntriesPlugin, self).render( context, instance, placeholder) context['template_to_render'] = (str(instance.template_to_render) or ...
python
def render(self, context, instance, placeholder): """ Update the context with plugin's data """ context = super(CMSRandomEntriesPlugin, self).render( context, instance, placeholder) context['template_to_render'] = (str(instance.template_to_render) or ...
['def', 'render', '(', 'self', ',', 'context', ',', 'instance', ',', 'placeholder', ')', ':', 'context', '=', 'super', '(', 'CMSRandomEntriesPlugin', ',', 'self', ')', '.', 'render', '(', 'context', ',', 'instance', ',', 'placeholder', ')', 'context', '[', "'template_to_render'", ']', '=', '(', 'str', '(', 'instance', ...
Update the context with plugin's data
['Update', 'the', 'context', 'with', 'plugin', 's', 'data']
train
https://github.com/django-blog-zinnia/cmsplugin-zinnia/blob/7613c0d9ae29affe9ab97527e4b6d5bef124afdc/cmsplugin_zinnia/cms_plugins.py#L131-L139
22
rjw57/throw
throw/minus/minus.py
Gallery.SaveGallery
def SaveGallery(self, name=None, items=None): """Use this to update the gallery name or change sort order. Specify which attribute (name or items or both) you want to change.""" url = 'http://min.us/api/SaveGallery' if not name: if not self.name: name = self...
python
def SaveGallery(self, name=None, items=None): """Use this to update the gallery name or change sort order. Specify which attribute (name or items or both) you want to change.""" url = 'http://min.us/api/SaveGallery' if not name: if not self.name: name = self...
['def', 'SaveGallery', '(', 'self', ',', 'name', '=', 'None', ',', 'items', '=', 'None', ')', ':', 'url', '=', "'http://min.us/api/SaveGallery'", 'if', 'not', 'name', ':', 'if', 'not', 'self', '.', 'name', ':', 'name', '=', 'self', '.', 'GetItems', '(', ')', '[', '0', ']', 'if', 'self', '.', 'name', ':', 'name', '=', '...
Use this to update the gallery name or change sort order. Specify which attribute (name or items or both) you want to change.
['Use', 'this', 'to', 'update', 'the', 'gallery', 'name', 'or', 'change', 'sort', 'order', '.', 'Specify', 'which', 'attribute', '(', 'name', 'or', 'items', 'or', 'both', ')', 'you', 'want', 'to', 'change', '.']
train
https://github.com/rjw57/throw/blob/74a7116362ba5b45635ab247472b25cfbdece4ee/throw/minus/minus.py#L62-L88
23
letuananh/chirptext
chirptext/cli.py
setup_logging
def setup_logging(filename, log_dir=None, force_setup=False): ''' Try to load logging configuration from a file. Set level to INFO if failed. ''' if not force_setup and ChirpCLI.SETUP_COMPLETED: logging.debug("Master logging has been setup. This call will be ignored.") return if log_dir ...
python
def setup_logging(filename, log_dir=None, force_setup=False): ''' Try to load logging configuration from a file. Set level to INFO if failed. ''' if not force_setup and ChirpCLI.SETUP_COMPLETED: logging.debug("Master logging has been setup. This call will be ignored.") return if log_dir ...
['def', 'setup_logging', '(', 'filename', ',', 'log_dir', '=', 'None', ',', 'force_setup', '=', 'False', ')', ':', 'if', 'not', 'force_setup', 'and', 'ChirpCLI', '.', 'SETUP_COMPLETED', ':', 'logging', '.', 'debug', '(', '"Master logging has been setup. This call will be ignored."', ')', 'return', 'if', 'log_dir', 'and...
Try to load logging configuration from a file. Set level to INFO if failed.
['Try', 'to', 'load', 'logging', 'configuration', 'from', 'a', 'file', '.', 'Set', 'level', 'to', 'INFO', 'if', 'failed', '.']
train
https://github.com/letuananh/chirptext/blob/ce60b47257b272a587c8703ea1f86cd1a45553a7/chirptext/cli.py#L35-L55
24
doconix/django-mako-plus
django_mako_plus/template/adapter.py
MakoTemplateAdapter.name
def name(self): '''Returns the name of this template (if created from a file) or "string" if not''' if self.mako_template.filename: return os.path.basename(self.mako_template.filename) return 'string'
python
def name(self): '''Returns the name of this template (if created from a file) or "string" if not''' if self.mako_template.filename: return os.path.basename(self.mako_template.filename) return 'string'
['def', 'name', '(', 'self', ')', ':', 'if', 'self', '.', 'mako_template', '.', 'filename', ':', 'return', 'os', '.', 'path', '.', 'basename', '(', 'self', '.', 'mako_template', '.', 'filename', ')', 'return', "'string'"]
Returns the name of this template (if created from a file) or "string" if not
['Returns', 'the', 'name', 'of', 'this', 'template', '(', 'if', 'created', 'from', 'a', 'file', ')', 'or', 'string', 'if', 'not']
train
https://github.com/doconix/django-mako-plus/blob/a90f9b4af19e5fa9f83452989cdcaed21569a181/django_mako_plus/template/adapter.py#L39-L43
25
HazyResearch/fonduer
src/fonduer/learning/disc_models/sparse_lstm.py
SparseLSTM.forward
def forward(self, X): """Forward function. :param X: The input (batch) of the model contains word sequences for lstm, features and feature weights. :type X: For word sequences: a list of torch.Tensor pair (word sequence and word mask) of shape (batch_size, sequence_lengt...
python
def forward(self, X): """Forward function. :param X: The input (batch) of the model contains word sequences for lstm, features and feature weights. :type X: For word sequences: a list of torch.Tensor pair (word sequence and word mask) of shape (batch_size, sequence_lengt...
['def', 'forward', '(', 'self', ',', 'X', ')', ':', 's', '=', 'X', '[', ':', '-', '2', ']', 'f', '=', 'X', '[', '-', '2', ']', 'w', '=', 'X', '[', '-', '1', ']', 'batch_size', '=', 'len', '(', 'f', ')', '# Generate lstm weight indices', 'x_idx', '=', 'self', '.', '_cuda', '(', 'torch', '.', 'as_tensor', '(', 'np', '.',...
Forward function. :param X: The input (batch) of the model contains word sequences for lstm, features and feature weights. :type X: For word sequences: a list of torch.Tensor pair (word sequence and word mask) of shape (batch_size, sequence_length). For features: tor...
['Forward', 'function', '.']
train
https://github.com/HazyResearch/fonduer/blob/4520f86a716f03dcca458a9f4bddac75b4e7068f/src/fonduer/learning/disc_models/sparse_lstm.py#L25-L64
26
google/grr
grr/server/grr_response_server/databases/mysql_flows.py
MySQLDBFlowMixin.CountFlowResultsByType
def CountFlowResultsByType(self, client_id, flow_id, cursor=None): """Returns counts of flow results grouped by result type.""" query = ("SELECT type, COUNT(*) FROM flow_results " "FORCE INDEX (flow_results_by_client_id_flow_id_timestamp) " "WHERE client_id = %s AND flow_id = %s " ...
python
def CountFlowResultsByType(self, client_id, flow_id, cursor=None): """Returns counts of flow results grouped by result type.""" query = ("SELECT type, COUNT(*) FROM flow_results " "FORCE INDEX (flow_results_by_client_id_flow_id_timestamp) " "WHERE client_id = %s AND flow_id = %s " ...
['def', 'CountFlowResultsByType', '(', 'self', ',', 'client_id', ',', 'flow_id', ',', 'cursor', '=', 'None', ')', ':', 'query', '=', '(', '"SELECT type, COUNT(*) FROM flow_results "', '"FORCE INDEX (flow_results_by_client_id_flow_id_timestamp) "', '"WHERE client_id = %s AND flow_id = %s "', '"GROUP BY type"', ')', 'arg...
Returns counts of flow results grouped by result type.
['Returns', 'counts', 'of', 'flow', 'results', 'grouped', 'by', 'result', 'type', '.']
train
https://github.com/google/grr/blob/5cef4e8e2f0d5df43ea4877e9c798e0bf60bfe74/grr/server/grr_response_server/databases/mysql_flows.py#L1350-L1360
27
google/grr
grr/client/grr_response_client/client_actions/artifact_collector.py
ArtifactCollector._ProcessGrepSource
def _ProcessGrepSource(self, source): """Find files fulfilling regex conditions.""" attributes = source.base_source.attributes paths = artifact_utils.InterpolateListKbAttributes( attributes["paths"], self.knowledge_base, self.ignore_interpolation_errors) regex = utils.RegexListDisjunctio...
python
def _ProcessGrepSource(self, source): """Find files fulfilling regex conditions.""" attributes = source.base_source.attributes paths = artifact_utils.InterpolateListKbAttributes( attributes["paths"], self.knowledge_base, self.ignore_interpolation_errors) regex = utils.RegexListDisjunctio...
['def', '_ProcessGrepSource', '(', 'self', ',', 'source', ')', ':', 'attributes', '=', 'source', '.', 'base_source', '.', 'attributes', 'paths', '=', 'artifact_utils', '.', 'InterpolateListKbAttributes', '(', 'attributes', '[', '"paths"', ']', ',', 'self', '.', 'knowledge_base', ',', 'self', '.', 'ignore_interpolation_...
Find files fulfilling regex conditions.
['Find', 'files', 'fulfilling', 'regex', 'conditions', '.']
train
https://github.com/google/grr/blob/5cef4e8e2f0d5df43ea4877e9c798e0bf60bfe74/grr/client/grr_response_client/client_actions/artifact_collector.py#L208-L225
28
F5Networks/f5-common-python
f5/multi_device/cluster/__init__.py
ClusterManager.manage_extant
def manage_extant(self, **kwargs): '''Manage an existing cluster :param kwargs: dict -- keyword args in dict ''' self._check_device_number(kwargs['devices']) self.trust_domain = TrustDomain( devices=kwargs['devices'], partition=kwargs['device_group_parti...
python
def manage_extant(self, **kwargs): '''Manage an existing cluster :param kwargs: dict -- keyword args in dict ''' self._check_device_number(kwargs['devices']) self.trust_domain = TrustDomain( devices=kwargs['devices'], partition=kwargs['device_group_parti...
['def', 'manage_extant', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'self', '.', '_check_device_number', '(', 'kwargs', '[', "'devices'", ']', ')', 'self', '.', 'trust_domain', '=', 'TrustDomain', '(', 'devices', '=', 'kwargs', '[', "'devices'", ']', ',', 'partition', '=', 'kwargs', '[', "'device_group_partition'"...
Manage an existing cluster :param kwargs: dict -- keyword args in dict
['Manage', 'an', 'existing', 'cluster']
train
https://github.com/F5Networks/f5-common-python/blob/7e67d5acd757a60e3d5f8c88c534bd72208f5494/f5/multi_device/cluster/__init__.py#L136-L148
29
google/python-gflags
gflags/flagvalues.py
FlagValues._GetFlagsDefinedByModule
def _GetFlagsDefinedByModule(self, module): """Returns the list of flags defined by a module. Args: module: A module object or a module name (a string). Returns: A new list of Flag objects. Caller may update this list as he wishes: none of those changes will affect the internals of this...
python
def _GetFlagsDefinedByModule(self, module): """Returns the list of flags defined by a module. Args: module: A module object or a module name (a string). Returns: A new list of Flag objects. Caller may update this list as he wishes: none of those changes will affect the internals of this...
['def', '_GetFlagsDefinedByModule', '(', 'self', ',', 'module', ')', ':', 'if', 'not', 'isinstance', '(', 'module', ',', 'str', ')', ':', 'module', '=', 'module', '.', '__name__', 'return', 'list', '(', 'self', '.', 'FlagsByModuleDict', '(', ')', '.', 'get', '(', 'module', ',', '[', ']', ')', ')']
Returns the list of flags defined by a module. Args: module: A module object or a module name (a string). Returns: A new list of Flag objects. Caller may update this list as he wishes: none of those changes will affect the internals of this FlagValue object.
['Returns', 'the', 'list', 'of', 'flags', 'defined', 'by', 'a', 'module', '.']
train
https://github.com/google/python-gflags/blob/4f06c3d0d6cbe9b1fb90ee9fb1c082b3bf9285f6/gflags/flagvalues.py#L265-L279
30
openp2pdesign/makerlabs
makerlabs/hackaday_io.py
get_labs
def get_labs(format): """Gets Hackerspaces data from hackaday.io.""" hackerspaces_json = data_from_hackaday_io(hackaday_io_labs_map_url) hackerspaces = {} # Load all the Hackerspaces for i in hackerspaces_json: current_lab = Hackerspace() current_lab.id = i["id"] current_la...
python
def get_labs(format): """Gets Hackerspaces data from hackaday.io.""" hackerspaces_json = data_from_hackaday_io(hackaday_io_labs_map_url) hackerspaces = {} # Load all the Hackerspaces for i in hackerspaces_json: current_lab = Hackerspace() current_lab.id = i["id"] current_la...
['def', 'get_labs', '(', 'format', ')', ':', 'hackerspaces_json', '=', 'data_from_hackaday_io', '(', 'hackaday_io_labs_map_url', ')', 'hackerspaces', '=', '{', '}', '# Load all the Hackerspaces', 'for', 'i', 'in', 'hackerspaces_json', ':', 'current_lab', '=', 'Hackerspace', '(', ')', 'current_lab', '.', 'id', '=', 'i',...
Gets Hackerspaces data from hackaday.io.
['Gets', 'Hackerspaces', 'data', 'from', 'hackaday', '.', 'io', '.']
train
https://github.com/openp2pdesign/makerlabs/blob/b5838440174f10d370abb671358db9a99d7739fd/makerlabs/hackaday_io.py#L57-L137
31
Azure/blobxfer
cli/cli.py
upload
def upload(ctx): """Upload files to Azure Storage""" settings.add_cli_options(ctx.cli_options, settings.TransferAction.Upload) ctx.initialize(settings.TransferAction.Upload) specs = settings.create_upload_specifications( ctx.cli_options, ctx.config) del ctx.cli_options for spec in specs:...
python
def upload(ctx): """Upload files to Azure Storage""" settings.add_cli_options(ctx.cli_options, settings.TransferAction.Upload) ctx.initialize(settings.TransferAction.Upload) specs = settings.create_upload_specifications( ctx.cli_options, ctx.config) del ctx.cli_options for spec in specs:...
['def', 'upload', '(', 'ctx', ')', ':', 'settings', '.', 'add_cli_options', '(', 'ctx', '.', 'cli_options', ',', 'settings', '.', 'TransferAction', '.', 'Upload', ')', 'ctx', '.', 'initialize', '(', 'settings', '.', 'TransferAction', '.', 'Upload', ')', 'specs', '=', 'settings', '.', 'create_upload_specifications', '('...
Upload files to Azure Storage
['Upload', 'files', 'to', 'Azure', 'Storage']
train
https://github.com/Azure/blobxfer/blob/3eccbe7530cc6a20ab2d30f9e034b6f021817f34/cli/cli.py#L1106-L1116
32
mitsei/dlkit
dlkit/json_/resource/sessions.py
ResourceBinAssignmentSession.get_assignable_bin_ids
def get_assignable_bin_ids(self, bin_id): """Gets a list of bins including and under the given bin node in which any resource can be assigned. arg: bin_id (osid.id.Id): the ``Id`` of the ``Bin`` return: (osid.id.IdList) - list of assignable bin ``Ids`` raise: NullArgument - ``bin_id...
python
def get_assignable_bin_ids(self, bin_id): """Gets a list of bins including and under the given bin node in which any resource can be assigned. arg: bin_id (osid.id.Id): the ``Id`` of the ``Bin`` return: (osid.id.IdList) - list of assignable bin ``Ids`` raise: NullArgument - ``bin_id...
['def', 'get_assignable_bin_ids', '(', 'self', ',', 'bin_id', ')', ':', '# Implemented from template for', '# osid.resource.ResourceBinAssignmentSession.get_assignable_bin_ids', '# This will likely be overridden by an authorization adapter', 'mgr', '=', 'self', '.', '_get_provider_manager', '(', "'RESOURCE'", ',', 'loc...
Gets a list of bins including and under the given bin node in which any resource can be assigned. arg: bin_id (osid.id.Id): the ``Id`` of the ``Bin`` return: (osid.id.IdList) - list of assignable bin ``Ids`` raise: NullArgument - ``bin_id`` is ``null`` raise: OperationFailed - unab...
['Gets', 'a', 'list', 'of', 'bins', 'including', 'and', 'under', 'the', 'given', 'bin', 'node', 'in', 'which', 'any', 'resource', 'can', 'be', 'assigned', '.']
train
https://github.com/mitsei/dlkit/blob/445f968a175d61c8d92c0f617a3c17dc1dc7c584/dlkit/json_/resource/sessions.py#L1562-L1581
33
sbusard/wagoner
wagoner/table.py
Table.check
def check(self): """ Check that this table is complete, that is, every character of this table can be followed by a new character. :return: True if the table is complete, False otherwise. """ for character, followers in self.items(): for follower in followers...
python
def check(self): """ Check that this table is complete, that is, every character of this table can be followed by a new character. :return: True if the table is complete, False otherwise. """ for character, followers in self.items(): for follower in followers...
['def', 'check', '(', 'self', ')', ':', 'for', 'character', ',', 'followers', 'in', 'self', '.', 'items', '(', ')', ':', 'for', 'follower', 'in', 'followers', ':', 'if', 'follower', 'not', 'in', 'self', ':', 'return', 'False', 'return', 'True']
Check that this table is complete, that is, every character of this table can be followed by a new character. :return: True if the table is complete, False otherwise.
['Check', 'that', 'this', 'table', 'is', 'complete', 'that', 'is', 'every', 'character', 'of', 'this', 'table', 'can', 'be', 'followed', 'by', 'a', 'new', 'character', '.']
train
https://github.com/sbusard/wagoner/blob/7f83d66bbd0e009e4d4232ffdf319bd5a2a5683b/wagoner/table.py#L81-L92
34
DataDog/integrations-core
openstack/datadog_checks/openstack/openstack.py
KeystoneCatalog.get_neutron_endpoint
def get_neutron_endpoint(cls, json_resp): """ Parse the service catalog returned by the Identity API for an endpoint matching the Neutron service Sends a CRITICAL service check when none are found registered in the Catalog """ catalog = json_resp.get('token', {}).get('catalog', [...
python
def get_neutron_endpoint(cls, json_resp): """ Parse the service catalog returned by the Identity API for an endpoint matching the Neutron service Sends a CRITICAL service check when none are found registered in the Catalog """ catalog = json_resp.get('token', {}).get('catalog', [...
['def', 'get_neutron_endpoint', '(', 'cls', ',', 'json_resp', ')', ':', 'catalog', '=', 'json_resp', '.', 'get', '(', "'token'", ',', '{', '}', ')', '.', 'get', '(', "'catalog'", ',', '[', ']', ')', 'match', '=', "'neutron'", 'neutron_endpoint', '=', 'None', 'for', 'entry', 'in', 'catalog', ':', 'if', 'entry', '[', "'n...
Parse the service catalog returned by the Identity API for an endpoint matching the Neutron service Sends a CRITICAL service check when none are found registered in the Catalog
['Parse', 'the', 'service', 'catalog', 'returned', 'by', 'the', 'Identity', 'API', 'for', 'an', 'endpoint', 'matching', 'the', 'Neutron', 'service', 'Sends', 'a', 'CRITICAL', 'service', 'check', 'when', 'none', 'are', 'found', 'registered', 'in', 'the', 'Catalog']
train
https://github.com/DataDog/integrations-core/blob/ebd41c873cf9f97a8c51bf9459bc6a7536af8acd/openstack/datadog_checks/openstack/openstack.py#L455-L479
35
materialsproject/pymatgen
pymatgen/io/abinit/flows.py
Flow.validate_json_schema
def validate_json_schema(self): """Validate the JSON schema. Return list of errors.""" errors = [] for work in self: for task in work: if not task.get_results().validate_json_schema(): errors.append(task) if not work.get_results().vali...
python
def validate_json_schema(self): """Validate the JSON schema. Return list of errors.""" errors = [] for work in self: for task in work: if not task.get_results().validate_json_schema(): errors.append(task) if not work.get_results().vali...
['def', 'validate_json_schema', '(', 'self', ')', ':', 'errors', '=', '[', ']', 'for', 'work', 'in', 'self', ':', 'for', 'task', 'in', 'work', ':', 'if', 'not', 'task', '.', 'get_results', '(', ')', '.', 'validate_json_schema', '(', ')', ':', 'errors', '.', 'append', '(', 'task', ')', 'if', 'not', 'work', '.', 'get_res...
Validate the JSON schema. Return list of errors.
['Validate', 'the', 'JSON', 'schema', '.', 'Return', 'list', 'of', 'errors', '.']
train
https://github.com/materialsproject/pymatgen/blob/4ca558cf72f8d5f8a1f21dfdfc0181a971c186da/pymatgen/io/abinit/flows.py#L426-L439
36
ga4gh/ga4gh-server
ga4gh/server/datamodel/datasets.py
Dataset.getRnaQuantificationSetByName
def getRnaQuantificationSetByName(self, name): """ Returns the RnaQuantification set with the specified name, or raises an exception otherwise. """ if name not in self._rnaQuantificationSetNameMap: raise exceptions.RnaQuantificationSetNameNotFoundException(name) ...
python
def getRnaQuantificationSetByName(self, name): """ Returns the RnaQuantification set with the specified name, or raises an exception otherwise. """ if name not in self._rnaQuantificationSetNameMap: raise exceptions.RnaQuantificationSetNameNotFoundException(name) ...
['def', 'getRnaQuantificationSetByName', '(', 'self', ',', 'name', ')', ':', 'if', 'name', 'not', 'in', 'self', '.', '_rnaQuantificationSetNameMap', ':', 'raise', 'exceptions', '.', 'RnaQuantificationSetNameNotFoundException', '(', 'name', ')', 'return', 'self', '.', '_rnaQuantificationSetNameMap', '[', 'name', ']']
Returns the RnaQuantification set with the specified name, or raises an exception otherwise.
['Returns', 'the', 'RnaQuantification', 'set', 'with', 'the', 'specified', 'name', 'or', 'raises', 'an', 'exception', 'otherwise', '.']
train
https://github.com/ga4gh/ga4gh-server/blob/1aa18922ef136db8604f6f098cb1732cba6f2a76/ga4gh/server/datamodel/datasets.py#L402-L409
37
dpnova/python-xprintidle
xprintidle.py
_create_modulename
def _create_modulename(cdef_sources, source, sys_version): """ This is the same as CFFI's create modulename except we don't include the CFFI version. """ key = '\x00'.join([sys_version[:3], source, cdef_sources]) key = key.encode('utf-8') k1 = hex(binascii.crc32(key[0::2]) & 0xffffffff) ...
python
def _create_modulename(cdef_sources, source, sys_version): """ This is the same as CFFI's create modulename except we don't include the CFFI version. """ key = '\x00'.join([sys_version[:3], source, cdef_sources]) key = key.encode('utf-8') k1 = hex(binascii.crc32(key[0::2]) & 0xffffffff) ...
['def', '_create_modulename', '(', 'cdef_sources', ',', 'source', ',', 'sys_version', ')', ':', 'key', '=', "'\\x00'", '.', 'join', '(', '[', 'sys_version', '[', ':', '3', ']', ',', 'source', ',', 'cdef_sources', ']', ')', 'key', '=', 'key', '.', 'encode', '(', "'utf-8'", ')', 'k1', '=', 'hex', '(', 'binascii', '.', 'c...
This is the same as CFFI's create modulename except we don't include the CFFI version.
['This', 'is', 'the', 'same', 'as', 'CFFI', 's', 'create', 'modulename', 'except', 'we', 'don', 't', 'include', 'the', 'CFFI', 'version', '.']
train
https://github.com/dpnova/python-xprintidle/blob/cc8f3c13a5dd578073d20f3d42208fcb8e1983b8/xprintidle.py#L10-L21
38
inasafe/inasafe
safe/gui/tools/batch/batch_dialog.py
BatchDialog.populate_table
def populate_table(self, scenario_directory): """ Populate table with files from scenario_directory directory. :param scenario_directory: Path where .txt & .py reside. :type scenario_directory: QString """ parsed_files = [] unparsed_files = [] self.table.clearCon...
python
def populate_table(self, scenario_directory): """ Populate table with files from scenario_directory directory. :param scenario_directory: Path where .txt & .py reside. :type scenario_directory: QString """ parsed_files = [] unparsed_files = [] self.table.clearCon...
['def', 'populate_table', '(', 'self', ',', 'scenario_directory', ')', ':', 'parsed_files', '=', '[', ']', 'unparsed_files', '=', '[', ']', 'self', '.', 'table', '.', 'clearContents', '(', ')', '# Block signal to allow update checking only when the table is ready', 'self', '.', 'table', '.', 'blockSignals', '(', 'True'...
Populate table with files from scenario_directory directory. :param scenario_directory: Path where .txt & .py reside. :type scenario_directory: QString
['Populate', 'table', 'with', 'files', 'from', 'scenario_directory', 'directory', '.']
train
https://github.com/inasafe/inasafe/blob/831d60abba919f6d481dc94a8d988cc205130724/safe/gui/tools/batch/batch_dialog.py#L213-L251
39
sibirrer/lenstronomy
lenstronomy/Sampling/likelihood.py
LikelihoodModule.logL
def logL(self, args): """ routine to compute X2 given variable parameters for a MCMC/PSO chain """ #extract parameters kwargs_lens, kwargs_source, kwargs_lens_light, kwargs_ps, kwargs_cosmo = self.param.args2kwargs(args) #generate image and computes likelihood sel...
python
def logL(self, args): """ routine to compute X2 given variable parameters for a MCMC/PSO chain """ #extract parameters kwargs_lens, kwargs_source, kwargs_lens_light, kwargs_ps, kwargs_cosmo = self.param.args2kwargs(args) #generate image and computes likelihood sel...
['def', 'logL', '(', 'self', ',', 'args', ')', ':', '#extract parameters', 'kwargs_lens', ',', 'kwargs_source', ',', 'kwargs_lens_light', ',', 'kwargs_ps', ',', 'kwargs_cosmo', '=', 'self', '.', 'param', '.', 'args2kwargs', '(', 'args', ')', '#generate image and computes likelihood', 'self', '.', '_reset_point_source_c...
routine to compute X2 given variable parameters for a MCMC/PSO chain
['routine', 'to', 'compute', 'X2', 'given', 'variable', 'parameters', 'for', 'a', 'MCMC', '/', 'PSO', 'chain']
train
https://github.com/sibirrer/lenstronomy/blob/4edb100a4f3f4fdc4fac9b0032d2b0283d0aa1d6/lenstronomy/Sampling/likelihood.py#L96-L123
40
wandb/client
wandb/__init__.py
log
def log(row=None, commit=True, *args, **kargs): """Log a dict to the global run's history. If commit is false, enables multiple calls before commiting. Eg. wandb.log({'train-loss': 0.5, 'accuracy': 0.9}) """ if run is None: raise ValueError( "You must call `wandb.init` in the ...
python
def log(row=None, commit=True, *args, **kargs): """Log a dict to the global run's history. If commit is false, enables multiple calls before commiting. Eg. wandb.log({'train-loss': 0.5, 'accuracy': 0.9}) """ if run is None: raise ValueError( "You must call `wandb.init` in the ...
['def', 'log', '(', 'row', '=', 'None', ',', 'commit', '=', 'True', ',', '*', 'args', ',', '*', '*', 'kargs', ')', ':', 'if', 'run', 'is', 'None', ':', 'raise', 'ValueError', '(', '"You must call `wandb.init` in the same process before calling log"', ')', 'if', 'row', 'is', 'None', ':', 'row', '=', '{', '}', 'if', 'com...
Log a dict to the global run's history. If commit is false, enables multiple calls before commiting. Eg. wandb.log({'train-loss': 0.5, 'accuracy': 0.9})
['Log', 'a', 'dict', 'to', 'the', 'global', 'run', 's', 'history', '.', 'If', 'commit', 'is', 'false', 'enables', 'multiple', 'calls', 'before', 'commiting', '.']
train
https://github.com/wandb/client/blob/7d08954ed5674fee223cd85ed0d8518fe47266b2/wandb/__init__.py#L465-L481
41
aras7/deployr-python-client
deployr_connection.py
DeployRConnection.login
def login(self, username, password, disableautosave=True, print_response=True): """ :param username: :param password: :param disableautosave: boolean :param print_response: print log if required :return: status code, response data """ if type(username) != ...
python
def login(self, username, password, disableautosave=True, print_response=True): """ :param username: :param password: :param disableautosave: boolean :param print_response: print log if required :return: status code, response data """ if type(username) != ...
['def', 'login', '(', 'self', ',', 'username', ',', 'password', ',', 'disableautosave', '=', 'True', ',', 'print_response', '=', 'True', ')', ':', 'if', 'type', '(', 'username', ')', '!=', 'str', ':', 'return', 'False', ',', '"Username must be string"', 'if', 'type', '(', 'password', ')', '!=', 'str', ':', 'return', 'F...
:param username: :param password: :param disableautosave: boolean :param print_response: print log if required :return: status code, response data
[':', 'param', 'username', ':', ':', 'param', 'password', ':', ':', 'param', 'disableautosave', ':', 'boolean', ':', 'param', 'print_response', ':', 'print', 'log', 'if', 'required', ':', 'return', ':', 'status', 'code', 'response', 'data']
train
https://github.com/aras7/deployr-python-client/blob/3ca517ff38e9a7dd1e21fcc88d54537546b9e7e5/deployr_connection.py#L28-L55
42
saltstack/salt
salt/states/zone.py
detached
def detached(name): ''' Ensure zone is detached name : string name of the zone ''' ret = {'name': name, 'changes': {}, 'result': None, 'comment': ''} zones = __salt__['zoneadm.list'](installed=True, configured=True) if name in zones: if zon...
python
def detached(name): ''' Ensure zone is detached name : string name of the zone ''' ret = {'name': name, 'changes': {}, 'result': None, 'comment': ''} zones = __salt__['zoneadm.list'](installed=True, configured=True) if name in zones: if zon...
['def', 'detached', '(', 'name', ')', ':', 'ret', '=', '{', "'name'", ':', 'name', ',', "'changes'", ':', '{', '}', ',', "'result'", ':', 'None', ',', "'comment'", ':', "''", '}', 'zones', '=', '__salt__', '[', "'zoneadm.list'", ']', '(', 'installed', '=', 'True', ',', 'configured', '=', 'True', ')', 'if', 'name', 'in'...
Ensure zone is detached name : string name of the zone
['Ensure', 'zone', 'is', 'detached']
train
https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/states/zone.py#L1100-L1138
43
onelogin/python3-saml
src/onelogin/saml2/response.py
OneLogin_Saml2_Response.__query
def __query(self, query, tagid=None): """ Extracts nodes that match the query from the Response :param query: Xpath Expresion :type query: String :param tagid: Tag ID :type query: String :returns: The queried nodes :rtype: list """ if se...
python
def __query(self, query, tagid=None): """ Extracts nodes that match the query from the Response :param query: Xpath Expresion :type query: String :param tagid: Tag ID :type query: String :returns: The queried nodes :rtype: list """ if se...
['def', '__query', '(', 'self', ',', 'query', ',', 'tagid', '=', 'None', ')', ':', 'if', 'self', '.', 'encrypted', ':', 'document', '=', 'self', '.', 'decrypted_document', 'else', ':', 'document', '=', 'self', '.', 'document', 'return', 'OneLogin_Saml2_XML', '.', 'query', '(', 'document', ',', 'query', ',', 'None', ','...
Extracts nodes that match the query from the Response :param query: Xpath Expresion :type query: String :param tagid: Tag ID :type query: String :returns: The queried nodes :rtype: list
['Extracts', 'nodes', 'that', 'match', 'the', 'query', 'from', 'the', 'Response']
train
https://github.com/onelogin/python3-saml/blob/064b7275fba1e5f39a9116ba1cdcc5d01fc34daa/src/onelogin/saml2/response.py#L760-L777
44
ray-project/ray
python/ray/rllib/env/atari_wrappers.py
wrap_deepmind
def wrap_deepmind(env, dim=84, framestack=True): """Configure environment for DeepMind-style Atari. Note that we assume reward clipping is done outside the wrapper. Args: dim (int): Dimension to resize observations to (dim x dim). framestack (bool): Whether to framestack observations. ...
python
def wrap_deepmind(env, dim=84, framestack=True): """Configure environment for DeepMind-style Atari. Note that we assume reward clipping is done outside the wrapper. Args: dim (int): Dimension to resize observations to (dim x dim). framestack (bool): Whether to framestack observations. ...
['def', 'wrap_deepmind', '(', 'env', ',', 'dim', '=', '84', ',', 'framestack', '=', 'True', ')', ':', 'env', '=', 'MonitorEnv', '(', 'env', ')', 'env', '=', 'NoopResetEnv', '(', 'env', ',', 'noop_max', '=', '30', ')', 'if', '"NoFrameskip"', 'in', 'env', '.', 'spec', '.', 'id', ':', 'env', '=', 'MaxAndSkipEnv', '(', 'en...
Configure environment for DeepMind-style Atari. Note that we assume reward clipping is done outside the wrapper. Args: dim (int): Dimension to resize observations to (dim x dim). framestack (bool): Whether to framestack observations.
['Configure', 'environment', 'for', 'DeepMind', '-', 'style', 'Atari', '.']
train
https://github.com/ray-project/ray/blob/4eade036a0505e244c976f36aaa2d64386b5129b/python/ray/rllib/env/atari_wrappers.py#L270-L291
45
monarch-initiative/dipper
dipper/sources/Reactome.py
Reactome._parse_reactome_association_file
def _parse_reactome_association_file( self, file, limit=None, subject_prefix=None, object_prefix=None): """ Parse ensembl gene to reactome pathway file :param file: file path (not handle) :param limit: limit (int, optional) limit the number of rows processed :return: ...
python
def _parse_reactome_association_file( self, file, limit=None, subject_prefix=None, object_prefix=None): """ Parse ensembl gene to reactome pathway file :param file: file path (not handle) :param limit: limit (int, optional) limit the number of rows processed :return: ...
['def', '_parse_reactome_association_file', '(', 'self', ',', 'file', ',', 'limit', '=', 'None', ',', 'subject_prefix', '=', 'None', ',', 'object_prefix', '=', 'None', ')', ':', 'eco_map', '=', 'Reactome', '.', 'get_eco_map', '(', 'Reactome', '.', 'map_files', '[', "'eco_map'", ']', ')', 'count', '=', '0', 'with', 'ope...
Parse ensembl gene to reactome pathway file :param file: file path (not handle) :param limit: limit (int, optional) limit the number of rows processed :return: None
['Parse', 'ensembl', 'gene', 'to', 'reactome', 'pathway', 'file', ':', 'param', 'file', ':', 'file', 'path', '(', 'not', 'handle', ')', ':', 'param', 'limit', ':', 'limit', '(', 'int', 'optional', ')', 'limit', 'the', 'number', 'of', 'rows', 'processed', ':', 'return', ':', 'None']
train
https://github.com/monarch-initiative/dipper/blob/24cc80db355bbe15776edc5c7b41e0886959ba41/dipper/sources/Reactome.py#L75-L98
46
jtwhite79/pyemu
pyemu/utils/helpers.py
PstFromFlopyModel.build_prior
def build_prior(self, fmt="ascii",filename=None,droptol=None, chunk=None, sparse=False, sigma_range=6): """ build a prior parameter covariance matrix. Parameters ---------- fmt : str the format to save the cov matrix. Options are "ascii","binary"...
python
def build_prior(self, fmt="ascii",filename=None,droptol=None, chunk=None, sparse=False, sigma_range=6): """ build a prior parameter covariance matrix. Parameters ---------- fmt : str the format to save the cov matrix. Options are "ascii","binary"...
['def', 'build_prior', '(', 'self', ',', 'fmt', '=', '"ascii"', ',', 'filename', '=', 'None', ',', 'droptol', '=', 'None', ',', 'chunk', '=', 'None', ',', 'sparse', '=', 'False', ',', 'sigma_range', '=', '6', ')', ':', 'fmt', '=', 'fmt', '.', 'lower', '(', ')', 'acc_fmts', '=', '[', '"ascii"', ',', '"binary"', ',', '"u...
build a prior parameter covariance matrix. Parameters ---------- fmt : str the format to save the cov matrix. Options are "ascii","binary","uncfile", "coo". default is "ascii" filename : str the filename to save the prior cov matr...
['build', 'a', 'prior', 'parameter', 'covariance', 'matrix', '.']
train
https://github.com/jtwhite79/pyemu/blob/c504d8e7a4097cec07655a6318d275739bd8148a/pyemu/utils/helpers.py#L2632-L2744
47
wummel/linkchecker
linkcheck/checker/__init__.py
get_url_from
def get_url_from (base_url, recursion_level, aggregate, parent_url=None, base_ref=None, line=0, column=0, page=0, name=u"", parent_content_type=None, extern=None): """ Get url data from given base data. @param base_url: base url from a link tag @type base_url: string...
python
def get_url_from (base_url, recursion_level, aggregate, parent_url=None, base_ref=None, line=0, column=0, page=0, name=u"", parent_content_type=None, extern=None): """ Get url data from given base data. @param base_url: base url from a link tag @type base_url: string...
['def', 'get_url_from', '(', 'base_url', ',', 'recursion_level', ',', 'aggregate', ',', 'parent_url', '=', 'None', ',', 'base_ref', '=', 'None', ',', 'line', '=', '0', ',', 'column', '=', '0', ',', 'page', '=', '0', ',', 'name', '=', 'u""', ',', 'parent_content_type', '=', 'None', ',', 'extern', '=', 'None', ')', ':', ...
Get url data from given base data. @param base_url: base url from a link tag @type base_url: string or None @param recursion_level: current recursion level @type recursion_level: number @param aggregate: aggregate object @type aggregate: aggregate.Consumer @param parent_url: parent url ...
['Get', 'url', 'data', 'from', 'given', 'base', 'data', '.']
train
https://github.com/wummel/linkchecker/blob/c2ce810c3fb00b895a841a7be6b2e78c64e7b042/linkcheck/checker/__init__.py#L67-L125
48
peercoin/peercoin_rpc
peercoin_rpc/peercoin_rpc.py
Client.sendtoaddress
def sendtoaddress(self, recv_addr, amount, comment=""): """send ammount to address, with optional comment. Returns txid. sendtoaddress(ADDRESS, AMMOUNT, COMMENT)""" return self.req("sendtoaddress", [recv_addr, amount, comment])
python
def sendtoaddress(self, recv_addr, amount, comment=""): """send ammount to address, with optional comment. Returns txid. sendtoaddress(ADDRESS, AMMOUNT, COMMENT)""" return self.req("sendtoaddress", [recv_addr, amount, comment])
['def', 'sendtoaddress', '(', 'self', ',', 'recv_addr', ',', 'amount', ',', 'comment', '=', '""', ')', ':', 'return', 'self', '.', 'req', '(', '"sendtoaddress"', ',', '[', 'recv_addr', ',', 'amount', ',', 'comment', ']', ')']
send ammount to address, with optional comment. Returns txid. sendtoaddress(ADDRESS, AMMOUNT, COMMENT)
['send', 'ammount', 'to', 'address', 'with', 'optional', 'comment', '.', 'Returns', 'txid', '.', 'sendtoaddress', '(', 'ADDRESS', 'AMMOUNT', 'COMMENT', ')']
train
https://github.com/peercoin/peercoin_rpc/blob/6edd854c7fd607ad9f6f4d5eb8b8b7c7fd8c16cc/peercoin_rpc/peercoin_rpc.py#L182-L185
49
tensorflow/hub
examples/image_retraining/retrain.py
variable_summaries
def variable_summaries(var): """Attach a lot of summaries to a Tensor (for TensorBoard visualization).""" with tf.name_scope('summaries'): mean = tf.reduce_mean(var) tf.summary.scalar('mean', mean) with tf.name_scope('stddev'): stddev = tf.sqrt(tf.reduce_mean(tf.square(var - mean))) tf.summary...
python
def variable_summaries(var): """Attach a lot of summaries to a Tensor (for TensorBoard visualization).""" with tf.name_scope('summaries'): mean = tf.reduce_mean(var) tf.summary.scalar('mean', mean) with tf.name_scope('stddev'): stddev = tf.sqrt(tf.reduce_mean(tf.square(var - mean))) tf.summary...
['def', 'variable_summaries', '(', 'var', ')', ':', 'with', 'tf', '.', 'name_scope', '(', "'summaries'", ')', ':', 'mean', '=', 'tf', '.', 'reduce_mean', '(', 'var', ')', 'tf', '.', 'summary', '.', 'scalar', '(', "'mean'", ',', 'mean', ')', 'with', 'tf', '.', 'name_scope', '(', "'stddev'", ')', ':', 'stddev', '=', 'tf'...
Attach a lot of summaries to a Tensor (for TensorBoard visualization).
['Attach', 'a', 'lot', 'of', 'summaries', 'to', 'a', 'Tensor', '(', 'for', 'TensorBoard', 'visualization', ')', '.']
train
https://github.com/tensorflow/hub/blob/09f45963f6787322967b6fec61459f3ac56fbb27/examples/image_retraining/retrain.py#L709-L719
50
henocdz/workon
workon/script.py
WorkOn._is_unique
def _is_unique(self, name, path): """verify if there is a project with given name or path on the database """ project = None try: project = Project.select().where( (Project.name == name) | (Project.path == path) )[0] ...
python
def _is_unique(self, name, path): """verify if there is a project with given name or path on the database """ project = None try: project = Project.select().where( (Project.name == name) | (Project.path == path) )[0] ...
['def', '_is_unique', '(', 'self', ',', 'name', ',', 'path', ')', ':', 'project', '=', 'None', 'try', ':', 'project', '=', 'Project', '.', 'select', '(', ')', '.', 'where', '(', '(', 'Project', '.', 'name', '==', 'name', ')', '|', '(', 'Project', '.', 'path', '==', 'path', ')', ')', '[', '0', ']', 'except', ':', 'pass'...
verify if there is a project with given name or path on the database
['verify', 'if', 'there', 'is', 'a', 'project', 'with', 'given', 'name', 'or', 'path', 'on', 'the', 'database']
train
https://github.com/henocdz/workon/blob/46f1f6dc4ea95d8efd10adf93a06737237a6874d/workon/script.py#L34-L47
51
tensorflow/tensor2tensor
tensor2tensor/models/image_transformer.py
imagetransformer_cifar_tpu_range
def imagetransformer_cifar_tpu_range(rhp): """Range of hyperparameters for vizier.""" # After starting from base, set intervals for some parameters. rhp.set_float("learning_rate", 0.01, 1.0, scale=rhp.LOG_SCALE) rhp.set_discrete("num_decoder_layers", [8, 10, 12, 14, 16]) rhp.set_discrete("hidden_size", [256, ...
python
def imagetransformer_cifar_tpu_range(rhp): """Range of hyperparameters for vizier.""" # After starting from base, set intervals for some parameters. rhp.set_float("learning_rate", 0.01, 1.0, scale=rhp.LOG_SCALE) rhp.set_discrete("num_decoder_layers", [8, 10, 12, 14, 16]) rhp.set_discrete("hidden_size", [256, ...
['def', 'imagetransformer_cifar_tpu_range', '(', 'rhp', ')', ':', '# After starting from base, set intervals for some parameters.', 'rhp', '.', 'set_float', '(', '"learning_rate"', ',', '0.01', ',', '1.0', ',', 'scale', '=', 'rhp', '.', 'LOG_SCALE', ')', 'rhp', '.', 'set_discrete', '(', '"num_decoder_layers"', ',', '['...
Range of hyperparameters for vizier.
['Range', 'of', 'hyperparameters', 'for', 'vizier', '.']
train
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1054-L1062
52
sethmlarson/trytravis
trytravis.py
_wait_for_travis_build
def _wait_for_travis_build(url, commit, committed_at): """ Waits for a Travis build to appear with the given commit SHA """ print('Waiting for a Travis build to appear ' 'for `%s` after `%s`...' % (commit, committed_at)) import requests slug = _slug_from_url(url) start_time = time.time() ...
python
def _wait_for_travis_build(url, commit, committed_at): """ Waits for a Travis build to appear with the given commit SHA """ print('Waiting for a Travis build to appear ' 'for `%s` after `%s`...' % (commit, committed_at)) import requests slug = _slug_from_url(url) start_time = time.time() ...
['def', '_wait_for_travis_build', '(', 'url', ',', 'commit', ',', 'committed_at', ')', ':', 'print', '(', "'Waiting for a Travis build to appear '", "'for `%s` after `%s`...'", '%', '(', 'commit', ',', 'committed_at', ')', ')', 'import', 'requests', 'slug', '=', '_slug_from_url', '(', 'url', ')', 'start_time', '=', 'ti...
Waits for a Travis build to appear with the given commit SHA
['Waits', 'for', 'a', 'Travis', 'build', 'to', 'appear', 'with', 'the', 'given', 'commit', 'SHA']
train
https://github.com/sethmlarson/trytravis/blob/d92ed708fe71d8db93a6df8077d23ee39ec0364e/trytravis.py#L188-L234
53
nameko/nameko
nameko/cli/shell.py
make_nameko_helper
def make_nameko_helper(config): """Create a fake module that provides some convenient access to nameko standalone functionality for interactive shell usage. """ module = ModuleType('nameko') module.__doc__ = """Nameko shell helper for making rpc calls and dispatching events. Usage: >>> n.rpc.se...
python
def make_nameko_helper(config): """Create a fake module that provides some convenient access to nameko standalone functionality for interactive shell usage. """ module = ModuleType('nameko') module.__doc__ = """Nameko shell helper for making rpc calls and dispatching events. Usage: >>> n.rpc.se...
['def', 'make_nameko_helper', '(', 'config', ')', ':', 'module', '=', 'ModuleType', '(', "'nameko'", ')', 'module', '.', '__doc__', '=', '"""Nameko shell helper for making rpc calls and dispatching\nevents.\n\nUsage:\n >>> n.rpc.service.method()\n "reply"\n\n >>> n.dispatch_event(\'service\', \'event_type\', \...
Create a fake module that provides some convenient access to nameko standalone functionality for interactive shell usage.
['Create', 'a', 'fake', 'module', 'that', 'provides', 'some', 'convenient', 'access', 'to', 'nameko', 'standalone', 'functionality', 'for', 'interactive', 'shell', 'usage', '.']
train
https://github.com/nameko/nameko/blob/88d7e5211de4fcc1c34cd7f84d7c77f0619c5f5d/nameko/cli/shell.py#L58-L77
54
earwig/mwparserfromhell
mwparserfromhell/nodes/template.py
Template.add
def add(self, name, value, showkey=None, before=None, preserve_spacing=True): """Add a parameter to the template with a given *name* and *value*. *name* and *value* can be anything parsable by :func:`.utils.parse_anything`; pipes and equal signs are automatically escaped fro...
python
def add(self, name, value, showkey=None, before=None, preserve_spacing=True): """Add a parameter to the template with a given *name* and *value*. *name* and *value* can be anything parsable by :func:`.utils.parse_anything`; pipes and equal signs are automatically escaped fro...
['def', 'add', '(', 'self', ',', 'name', ',', 'value', ',', 'showkey', '=', 'None', ',', 'before', '=', 'None', ',', 'preserve_spacing', '=', 'True', ')', ':', 'name', ',', 'value', '=', 'parse_anything', '(', 'name', ')', ',', 'parse_anything', '(', 'value', ')', 'self', '.', '_surface_escape', '(', 'value', ',', '"|"...
Add a parameter to the template with a given *name* and *value*. *name* and *value* can be anything parsable by :func:`.utils.parse_anything`; pipes and equal signs are automatically escaped from *value* when appropriate. If *name* is already a parameter in the template, we'll replace ...
['Add', 'a', 'parameter', 'to', 'the', 'template', 'with', 'a', 'given', '*', 'name', '*', 'and', '*', 'value', '*', '.']
train
https://github.com/earwig/mwparserfromhell/blob/98dc30902d35c714a70aca8e6616f49d71cb24cc/mwparserfromhell/nodes/template.py#L232-L310
55
liamw9534/bt-manager
bt_manager/audio.py
SBCAudioCodec.register_transport_ready_event
def register_transport_ready_event(self, user_cb, user_arg): """ Register for transport ready events. The `transport ready` event is raised via a user callback. If the endpoint is configured as a source, then the user may then call :py:meth:`write_transport` in order to send da...
python
def register_transport_ready_event(self, user_cb, user_arg): """ Register for transport ready events. The `transport ready` event is raised via a user callback. If the endpoint is configured as a source, then the user may then call :py:meth:`write_transport` in order to send da...
['def', 'register_transport_ready_event', '(', 'self', ',', 'user_cb', ',', 'user_arg', ')', ':', 'self', '.', 'user_cb', '=', 'user_cb', 'self', '.', 'user_arg', '=', 'user_arg']
Register for transport ready events. The `transport ready` event is raised via a user callback. If the endpoint is configured as a source, then the user may then call :py:meth:`write_transport` in order to send data to the associated sink. Otherwise, if the endpoint is configur...
['Register', 'for', 'transport', 'ready', 'events', '.', 'The', 'transport', 'ready', 'event', 'is', 'raised', 'via', 'a', 'user', 'callback', '.', 'If', 'the', 'endpoint', 'is', 'configured', 'as', 'a', 'source', 'then', 'the', 'user', 'may', 'then', 'call', ':', 'py', ':', 'meth', ':', 'write_transport', 'in', 'order...
train
https://github.com/liamw9534/bt-manager/blob/51be2919394ce8134c698359649bfad09eedf4ec/bt_manager/audio.py#L207-L227
56
LesPatamechanix/patalib
src/patalib/anomaly.py
Anomaly.generate_anomaly
def generate_anomaly(self, input_word, list_of_dict_words, num): """ Generate an anomaly. This is done via a Psuedo-random number generator. """ results = [] for i in range(0,num): index = randint(0,len(list_of_dict_words)-1) name = list_of_dict_words[ind...
python
def generate_anomaly(self, input_word, list_of_dict_words, num): """ Generate an anomaly. This is done via a Psuedo-random number generator. """ results = [] for i in range(0,num): index = randint(0,len(list_of_dict_words)-1) name = list_of_dict_words[ind...
['def', 'generate_anomaly', '(', 'self', ',', 'input_word', ',', 'list_of_dict_words', ',', 'num', ')', ':', 'results', '=', '[', ']', 'for', 'i', 'in', 'range', '(', '0', ',', 'num', ')', ':', 'index', '=', 'randint', '(', '0', ',', 'len', '(', 'list_of_dict_words', ')', '-', '1', ')', 'name', '=', 'list_of_dict_words...
Generate an anomaly. This is done via a Psuedo-random number generator.
['Generate', 'an', 'anomaly', '.', 'This', 'is', 'done', 'via', 'a', 'Psuedo', '-', 'random', 'number', 'generator', '.']
train
https://github.com/LesPatamechanix/patalib/blob/d88cca409b1750fdeb88cece048b308f2a710955/src/patalib/anomaly.py#L9-L23
57
cqlengine/cqlengine
cqlengine/query.py
AbstractQuerySet.count
def count(self): """ Returns the number of rows matched by this query """ if self._batch: raise CQLEngineException("Only inserts, updates, and deletes are available in batch mode") if self._result_cache is None: query = self._select_query() query.count = True...
python
def count(self): """ Returns the number of rows matched by this query """ if self._batch: raise CQLEngineException("Only inserts, updates, and deletes are available in batch mode") if self._result_cache is None: query = self._select_query() query.count = True...
['def', 'count', '(', 'self', ')', ':', 'if', 'self', '.', '_batch', ':', 'raise', 'CQLEngineException', '(', '"Only inserts, updates, and deletes are available in batch mode"', ')', 'if', 'self', '.', '_result_cache', 'is', 'None', ':', 'query', '=', 'self', '.', '_select_query', '(', ')', 'query', '.', 'count', '=', ...
Returns the number of rows matched by this query
['Returns', 'the', 'number', 'of', 'rows', 'matched', 'by', 'this', 'query']
train
https://github.com/cqlengine/cqlengine/blob/7079eaf7071cbf5a045e1d1ab57f6d1b5ba3f9dc/cqlengine/query.py#L568-L579
58
chaimleib/intervaltree
intervaltree/intervaltree.py
IntervalTree.remove
def remove(self, interval): """ Removes an interval from the tree, if present. If not, raises ValueError. Completes in O(log n) time. """ #self.verify() if interval not in self: #print(self.all_intervals) raise ValueError self.top_...
python
def remove(self, interval): """ Removes an interval from the tree, if present. If not, raises ValueError. Completes in O(log n) time. """ #self.verify() if interval not in self: #print(self.all_intervals) raise ValueError self.top_...
['def', 'remove', '(', 'self', ',', 'interval', ')', ':', '#self.verify()', 'if', 'interval', 'not', 'in', 'self', ':', '#print(self.all_intervals)', 'raise', 'ValueError', 'self', '.', 'top_node', '=', 'self', '.', 'top_node', '.', 'remove', '(', 'interval', ')', 'self', '.', 'all_intervals', '.', 'remove', '(', 'inte...
Removes an interval from the tree, if present. If not, raises ValueError. Completes in O(log n) time.
['Removes', 'an', 'interval', 'from', 'the', 'tree', 'if', 'present', '.', 'If', 'not', 'raises', 'ValueError', '.']
train
https://github.com/chaimleib/intervaltree/blob/ffb2b1667f8b832e89324a75a175be8440504c9d/intervaltree/intervaltree.py#L356-L369
59
aws/sagemaker-python-sdk
src/sagemaker/predictor.py
RealTimePredictor.delete_model
def delete_model(self): """Deletes the Amazon SageMaker models backing this predictor. """ request_failed = False failed_models = [] for model_name in self._model_names: try: self.sagemaker_session.delete_model(model_name) except Exception...
python
def delete_model(self): """Deletes the Amazon SageMaker models backing this predictor. """ request_failed = False failed_models = [] for model_name in self._model_names: try: self.sagemaker_session.delete_model(model_name) except Exception...
['def', 'delete_model', '(', 'self', ')', ':', 'request_failed', '=', 'False', 'failed_models', '=', '[', ']', 'for', 'model_name', 'in', 'self', '.', '_model_names', ':', 'try', ':', 'self', '.', 'sagemaker_session', '.', 'delete_model', '(', 'model_name', ')', 'except', 'Exception', ':', '# pylint: disable=broad-exce...
Deletes the Amazon SageMaker models backing this predictor.
['Deletes', 'the', 'Amazon', 'SageMaker', 'models', 'backing', 'this', 'predictor', '.']
train
https://github.com/aws/sagemaker-python-sdk/blob/a9e724c7d3f5572b68c3903548c792a59d99799a/src/sagemaker/predictor.py#L131-L146
60
google/mobly
mobly/asserts.py
assert_raises
def assert_raises(expected_exception, extras=None, *args, **kwargs): """Assert that an exception is raised when a function is called. If no exception is raised, test fail. If an exception is raised but not of the expected type, the exception is let through. This should only be used as a context manage...
python
def assert_raises(expected_exception, extras=None, *args, **kwargs): """Assert that an exception is raised when a function is called. If no exception is raised, test fail. If an exception is raised but not of the expected type, the exception is let through. This should only be used as a context manage...
['def', 'assert_raises', '(', 'expected_exception', ',', 'extras', '=', 'None', ',', '*', 'args', ',', '*', '*', 'kwargs', ')', ':', 'context', '=', '_AssertRaisesContext', '(', 'expected_exception', ',', 'extras', '=', 'extras', ')', 'return', 'context']
Assert that an exception is raised when a function is called. If no exception is raised, test fail. If an exception is raised but not of the expected type, the exception is let through. This should only be used as a context manager: with assert_raises(Exception): func() Args: ...
['Assert', 'that', 'an', 'exception', 'is', 'raised', 'when', 'a', 'function', 'is', 'called', '.']
train
https://github.com/google/mobly/blob/38ba2cf7d29a20e6a2fca1718eecb337df38db26/mobly/asserts.py#L60-L77
61
saltstack/salt
salt/modules/mandrill.py
_http_request
def _http_request(url, headers=None, data=None): ''' Make the HTTP request and return the body as python object. ''' if not headers: headers = _get_headers() session = requests.session() log.debug('Querying %s', url) req = session.post(url, ...
python
def _http_request(url, headers=None, data=None): ''' Make the HTTP request and return the body as python object. ''' if not headers: headers = _get_headers() session = requests.session() log.debug('Querying %s', url) req = session.post(url, ...
['def', '_http_request', '(', 'url', ',', 'headers', '=', 'None', ',', 'data', '=', 'None', ')', ':', 'if', 'not', 'headers', ':', 'headers', '=', '_get_headers', '(', ')', 'session', '=', 'requests', '.', 'session', '(', ')', 'log', '.', 'debug', '(', "'Querying %s'", ',', 'url', ')', 'req', '=', 'session', '.', 'post...
Make the HTTP request and return the body as python object.
['Make', 'the', 'HTTP', 'request', 'and', 'return', 'the', 'body', 'as', 'python', 'object', '.']
train
https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/mandrill.py#L101-L132
62
saltstack/salt
salt/returners/sqlite3_return.py
save_load
def save_load(jid, load, minions=None): ''' Save the load to the specified jid ''' log.debug('sqlite3 returner <save_load> called jid: %s load: %s', jid, load) conn = _get_conn(ret=None) cur = conn.cursor() sql = '''INSERT INTO jids (jid, load) VALUES (:jid, :load)''' cur.execute(sql, ...
python
def save_load(jid, load, minions=None): ''' Save the load to the specified jid ''' log.debug('sqlite3 returner <save_load> called jid: %s load: %s', jid, load) conn = _get_conn(ret=None) cur = conn.cursor() sql = '''INSERT INTO jids (jid, load) VALUES (:jid, :load)''' cur.execute(sql, ...
['def', 'save_load', '(', 'jid', ',', 'load', ',', 'minions', '=', 'None', ')', ':', 'log', '.', 'debug', '(', "'sqlite3 returner <save_load> called jid: %s load: %s'", ',', 'jid', ',', 'load', ')', 'conn', '=', '_get_conn', '(', 'ret', '=', 'None', ')', 'cur', '=', 'conn', '.', 'cursor', '(', ')', 'sql', '=', "'''INSE...
Save the load to the specified jid
['Save', 'the', 'load', 'to', 'the', 'specified', 'jid']
train
https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/returners/sqlite3_return.py#L182-L193
63
iotile/coretools
iotilebuild/iotile/build/config/scons-local-3.0.1/SCons/Action.py
CommandGeneratorAction.get_presig
def get_presig(self, target, source, env, executor=None): """Return the signature contents of this action's command line. This strips $(-$) and everything in between the string, since those parts don't affect signatures. """ return self._generate(target, source, env, 1, executor...
python
def get_presig(self, target, source, env, executor=None): """Return the signature contents of this action's command line. This strips $(-$) and everything in between the string, since those parts don't affect signatures. """ return self._generate(target, source, env, 1, executor...
['def', 'get_presig', '(', 'self', ',', 'target', ',', 'source', ',', 'env', ',', 'executor', '=', 'None', ')', ':', 'return', 'self', '.', '_generate', '(', 'target', ',', 'source', ',', 'env', ',', '1', ',', 'executor', ')', '.', 'get_presig', '(', 'target', ',', 'source', ',', 'env', ')']
Return the signature contents of this action's command line. This strips $(-$) and everything in between the string, since those parts don't affect signatures.
['Return', 'the', 'signature', 'contents', 'of', 'this', 'action', 's', 'command', 'line', '.']
train
https://github.com/iotile/coretools/blob/2d794f5f1346b841b0dcd16c9d284e9bf2f3c6ec/iotilebuild/iotile/build/config/scons-local-3.0.1/SCons/Action.py#L1043-L1049
64
cmbruns/pyopenvr
src/openvr/__init__.py
IVROverlay.getOverlayTransformAbsolute
def getOverlayTransformAbsolute(self, ulOverlayHandle): """Gets the transform if it is absolute. Returns an error if the transform is some other type.""" fn = self.function_table.getOverlayTransformAbsolute peTrackingOrigin = ETrackingUniverseOrigin() pmatTrackingOriginToOverlayTransfor...
python
def getOverlayTransformAbsolute(self, ulOverlayHandle): """Gets the transform if it is absolute. Returns an error if the transform is some other type.""" fn = self.function_table.getOverlayTransformAbsolute peTrackingOrigin = ETrackingUniverseOrigin() pmatTrackingOriginToOverlayTransfor...
['def', 'getOverlayTransformAbsolute', '(', 'self', ',', 'ulOverlayHandle', ')', ':', 'fn', '=', 'self', '.', 'function_table', '.', 'getOverlayTransformAbsolute', 'peTrackingOrigin', '=', 'ETrackingUniverseOrigin', '(', ')', 'pmatTrackingOriginToOverlayTransform', '=', 'HmdMatrix34_t', '(', ')', 'result', '=', 'fn', '...
Gets the transform if it is absolute. Returns an error if the transform is some other type.
['Gets', 'the', 'transform', 'if', 'it', 'is', 'absolute', '.', 'Returns', 'an', 'error', 'if', 'the', 'transform', 'is', 'some', 'other', 'type', '.']
train
https://github.com/cmbruns/pyopenvr/blob/68395d26bb3df6ab1f0f059c38d441f962938be6/src/openvr/__init__.py#L4852-L4859
65
gem/oq-engine
openquake/hazardlib/geo/geodetic.py
spherical_to_cartesian
def spherical_to_cartesian(lons, lats, depths=None): """ Return the position vectors (in Cartesian coordinates) of list of spherical coordinates. For equations see: http://mathworld.wolfram.com/SphericalCoordinates.html. Parameters are components of spherical coordinates in a form of scalars, ...
python
def spherical_to_cartesian(lons, lats, depths=None): """ Return the position vectors (in Cartesian coordinates) of list of spherical coordinates. For equations see: http://mathworld.wolfram.com/SphericalCoordinates.html. Parameters are components of spherical coordinates in a form of scalars, ...
['def', 'spherical_to_cartesian', '(', 'lons', ',', 'lats', ',', 'depths', '=', 'None', ')', ':', 'phi', '=', 'numpy', '.', 'radians', '(', 'lons', ')', 'theta', '=', 'numpy', '.', 'radians', '(', 'lats', ')', 'if', 'depths', 'is', 'None', ':', 'rr', '=', 'EARTH_RADIUS', 'else', ':', 'rr', '=', 'EARTH_RADIUS', '-', 'nu...
Return the position vectors (in Cartesian coordinates) of list of spherical coordinates. For equations see: http://mathworld.wolfram.com/SphericalCoordinates.html. Parameters are components of spherical coordinates in a form of scalars, lists or numpy arrays. ``depths`` can be ``None`` in which case i...
['Return', 'the', 'position', 'vectors', '(', 'in', 'Cartesian', 'coordinates', ')', 'of', 'list', 'of', 'spherical', 'coordinates', '.']
train
https://github.com/gem/oq-engine/blob/8294553a0b8aba33fd96437a35065d03547d0040/openquake/hazardlib/geo/geodetic.py#L183-L221
66
iterative/dvc
dvc/command/metrics.py
show_metrics
def show_metrics(metrics, all_branches=False, all_tags=False): """ Args: metrics (list): Where each element is either a `list` if an xpath was specified, otherwise a `str` """ for branch, val in metrics.items(): if all_branches or all_tags: logger.info("{branch}:"...
python
def show_metrics(metrics, all_branches=False, all_tags=False): """ Args: metrics (list): Where each element is either a `list` if an xpath was specified, otherwise a `str` """ for branch, val in metrics.items(): if all_branches or all_tags: logger.info("{branch}:"...
['def', 'show_metrics', '(', 'metrics', ',', 'all_branches', '=', 'False', ',', 'all_tags', '=', 'False', ')', ':', 'for', 'branch', ',', 'val', 'in', 'metrics', '.', 'items', '(', ')', ':', 'if', 'all_branches', 'or', 'all_tags', ':', 'logger', '.', 'info', '(', '"{branch}:"', '.', 'format', '(', 'branch', '=', 'branc...
Args: metrics (list): Where each element is either a `list` if an xpath was specified, otherwise a `str`
['Args', ':', 'metrics', '(', 'list', ')', ':', 'Where', 'each', 'element', 'is', 'either', 'a', 'list', 'if', 'an', 'xpath', 'was', 'specified', 'otherwise', 'a', 'str']
train
https://github.com/iterative/dvc/blob/8bb21261e34c9632453e09090de7ebe50e38d341/dvc/command/metrics.py#L13-L33
67
redbridge/molnctrl
molnctrl/cachemaker.py
monkeycache
def monkeycache(apis): """ Feed this a dictionary of api bananas, it spits out processed cache """ if isinstance(type(apis), type(None)) or apis is None: return {} verbs = set() cache = {} cache['count'] = apis['count'] cache['asyncapis'] = [] apilist = apis['api'] if a...
python
def monkeycache(apis): """ Feed this a dictionary of api bananas, it spits out processed cache """ if isinstance(type(apis), type(None)) or apis is None: return {} verbs = set() cache = {} cache['count'] = apis['count'] cache['asyncapis'] = [] apilist = apis['api'] if a...
['def', 'monkeycache', '(', 'apis', ')', ':', 'if', 'isinstance', '(', 'type', '(', 'apis', ')', ',', 'type', '(', 'None', ')', ')', 'or', 'apis', 'is', 'None', ':', 'return', '{', '}', 'verbs', '=', 'set', '(', ')', 'cache', '=', '{', '}', 'cache', '[', "'count'", ']', '=', 'apis', '[', "'count'", ']', 'cache', '[', "...
Feed this a dictionary of api bananas, it spits out processed cache
['Feed', 'this', 'a', 'dictionary', 'of', 'api', 'bananas', 'it', 'spits', 'out', 'processed', 'cache']
train
https://github.com/redbridge/molnctrl/blob/9990ae7e522ce364bb61a735f774dc28de5f8e60/molnctrl/cachemaker.py#L82-L132
68
portfoliome/foil
foil/ftp.py
download_ftp_url
def download_ftp_url(source_url, target_uri, buffer_size=8192): """Uses urllib. thread safe?""" ensure_file_directory(target_uri) with urllib.request.urlopen(source_url) as source_file: with open(target_uri, 'wb') as target_file: shutil.copyfileobj(source_file, target_file, buffer_size...
python
def download_ftp_url(source_url, target_uri, buffer_size=8192): """Uses urllib. thread safe?""" ensure_file_directory(target_uri) with urllib.request.urlopen(source_url) as source_file: with open(target_uri, 'wb') as target_file: shutil.copyfileobj(source_file, target_file, buffer_size...
['def', 'download_ftp_url', '(', 'source_url', ',', 'target_uri', ',', 'buffer_size', '=', '8192', ')', ':', 'ensure_file_directory', '(', 'target_uri', ')', 'with', 'urllib', '.', 'request', '.', 'urlopen', '(', 'source_url', ')', 'as', 'source_file', ':', 'with', 'open', '(', 'target_uri', ',', "'wb'", ')', 'as', 'ta...
Uses urllib. thread safe?
['Uses', 'urllib', '.', 'thread', 'safe?']
train
https://github.com/portfoliome/foil/blob/b66d8cf4ab048a387d8c7a033b47e922ed6917d6/foil/ftp.py#L90-L97
69
couchbase/couchbase-python-client
couchbase/bucket.py
Bucket.map_get
def map_get(self, key, mapkey): """ Retrieve a value from a map. :param str key: The document ID :param str mapkey: Key within the map to retrieve :return: :class:`~.ValueResult` :raise: :exc:`IndexError` if the mapkey does not exist :raise: :cb_exc:`NotFoundErro...
python
def map_get(self, key, mapkey): """ Retrieve a value from a map. :param str key: The document ID :param str mapkey: Key within the map to retrieve :return: :class:`~.ValueResult` :raise: :exc:`IndexError` if the mapkey does not exist :raise: :cb_exc:`NotFoundErro...
['def', 'map_get', '(', 'self', ',', 'key', ',', 'mapkey', ')', ':', 'op', '=', 'SD', '.', 'get', '(', 'mapkey', ')', 'sdres', '=', 'self', '.', 'lookup_in', '(', 'key', ',', 'op', ')', 'return', 'self', '.', '_wrap_dsop', '(', 'sdres', ',', 'True', ')']
Retrieve a value from a map. :param str key: The document ID :param str mapkey: Key within the map to retrieve :return: :class:`~.ValueResult` :raise: :exc:`IndexError` if the mapkey does not exist :raise: :cb_exc:`NotFoundError` if the document does not exist. .. seeal...
['Retrieve', 'a', 'value', 'from', 'a', 'map', '.']
train
https://github.com/couchbase/couchbase-python-client/blob/a7bada167785bf79a29c39f820d932a433a6a535/couchbase/bucket.py#L2147-L2161
70
persandstrom/python-verisure
verisure/session.py
Session.set_lock_state
def set_lock_state(self, code, device_label, state): """ Lock or unlock Args: code (str): Lock code device_label (str): device label of lock state (str): 'lock' or 'unlock' """ response = None try: response = requests.put( ...
python
def set_lock_state(self, code, device_label, state): """ Lock or unlock Args: code (str): Lock code device_label (str): device label of lock state (str): 'lock' or 'unlock' """ response = None try: response = requests.put( ...
['def', 'set_lock_state', '(', 'self', ',', 'code', ',', 'device_label', ',', 'state', ')', ':', 'response', '=', 'None', 'try', ':', 'response', '=', 'requests', '.', 'put', '(', 'urls', '.', 'set_lockstate', '(', 'self', '.', '_giid', ',', 'device_label', ',', 'state', ')', ',', 'headers', '=', '{', "'Accept'", ':', ...
Lock or unlock Args: code (str): Lock code device_label (str): device label of lock state (str): 'lock' or 'unlock'
['Lock', 'or', 'unlock']
train
https://github.com/persandstrom/python-verisure/blob/babd25e7f8fb2b24f12e4109dfa8a04041e8dcb8/verisure/session.py#L309-L329
71
Sanji-IO/sanji
sanji/message.py
Message.to_event
def to_event(self): """ get rid of id, sign, tunnel and update message type Notice: this method will return a deepcopy """ msg = copy.deepcopy(self) for _ in ["id", "sign", "tunnel", "query", "param"]: if not hasattr(msg, _): continue ...
python
def to_event(self): """ get rid of id, sign, tunnel and update message type Notice: this method will return a deepcopy """ msg = copy.deepcopy(self) for _ in ["id", "sign", "tunnel", "query", "param"]: if not hasattr(msg, _): continue ...
['def', 'to_event', '(', 'self', ')', ':', 'msg', '=', 'copy', '.', 'deepcopy', '(', 'self', ')', 'for', '_', 'in', '[', '"id"', ',', '"sign"', ',', '"tunnel"', ',', '"query"', ',', '"param"', ']', ':', 'if', 'not', 'hasattr', '(', 'msg', ',', '_', ')', ':', 'continue', 'delattr', '(', 'msg', ',', '_', ')', 'msg', '.',...
get rid of id, sign, tunnel and update message type Notice: this method will return a deepcopy
['get', 'rid', 'of', 'id', 'sign', 'tunnel', 'and', 'update', 'message', 'type', 'Notice', ':', 'this', 'method', 'will', 'return', 'a', 'deepcopy']
train
https://github.com/Sanji-IO/sanji/blob/5c54cc2772bdfeae3337f785de1957237b828b34/sanji/message.py#L209-L222
72
evhub/coconut
coconut/compiler/compiler.py
Compiler.destructuring_stmt_handle
def destructuring_stmt_handle(self, original, loc, tokens): """Process match assign blocks.""" internal_assert(len(tokens) == 2, "invalid destructuring assignment tokens", tokens) matches, item = tokens out = match_handle(loc, [matches, "in", item, None]) out += self.pattern_erro...
python
def destructuring_stmt_handle(self, original, loc, tokens): """Process match assign blocks.""" internal_assert(len(tokens) == 2, "invalid destructuring assignment tokens", tokens) matches, item = tokens out = match_handle(loc, [matches, "in", item, None]) out += self.pattern_erro...
['def', 'destructuring_stmt_handle', '(', 'self', ',', 'original', ',', 'loc', ',', 'tokens', ')', ':', 'internal_assert', '(', 'len', '(', 'tokens', ')', '==', '2', ',', '"invalid destructuring assignment tokens"', ',', 'tokens', ')', 'matches', ',', 'item', '=', 'tokens', 'out', '=', 'match_handle', '(', 'loc', ',', ...
Process match assign blocks.
['Process', 'match', 'assign', 'blocks', '.']
train
https://github.com/evhub/coconut/blob/ff97177344e7604e89a0a98a977a87ed2a56fc6d/coconut/compiler/compiler.py#L1424-L1430
73
Erotemic/utool
utool/util_alg.py
knapsack_iterative_int
def knapsack_iterative_int(items, maxweight): r""" Iterative knapsack method Math: maximize \sum_{i \in T} v_i subject to \sum_{i \in T} w_i \leq W Notes: dpmat is the dynamic programming memoization matrix. dpmat[i, w] is the total value of the items with weight at mos...
python
def knapsack_iterative_int(items, maxweight): r""" Iterative knapsack method Math: maximize \sum_{i \in T} v_i subject to \sum_{i \in T} w_i \leq W Notes: dpmat is the dynamic programming memoization matrix. dpmat[i, w] is the total value of the items with weight at mos...
['def', 'knapsack_iterative_int', '(', 'items', ',', 'maxweight', ')', ':', 'values', '=', '[', 't', '[', '0', ']', 'for', 't', 'in', 'items', ']', 'weights', '=', '[', 't', '[', '1', ']', 'for', 't', 'in', 'items', ']', 'maxsize', '=', 'maxweight', '+', '1', '# Sparse representation seems better', 'dpmat', '=', 'defau...
r""" Iterative knapsack method Math: maximize \sum_{i \in T} v_i subject to \sum_{i \in T} w_i \leq W Notes: dpmat is the dynamic programming memoization matrix. dpmat[i, w] is the total value of the items with weight at most W T is idx_subset, the set of indicies i...
['r', 'Iterative', 'knapsack', 'method']
train
https://github.com/Erotemic/utool/blob/3b27e1f4e6e6fb23cd8744af7b7195b57d99e03a/utool/util_alg.py#L1400-L1477
74
jorgenschaefer/elpy
elpy/server.py
ElpyRPCServer.rpc_get_names
def rpc_get_names(self, filename, source, offset): """Get all possible names """ source = get_source(source) if hasattr(self.backend, "rpc_get_names"): return self.backend.rpc_get_names(filename, source, offset) else: raise Fault("get_names not implemente...
python
def rpc_get_names(self, filename, source, offset): """Get all possible names """ source = get_source(source) if hasattr(self.backend, "rpc_get_names"): return self.backend.rpc_get_names(filename, source, offset) else: raise Fault("get_names not implemente...
['def', 'rpc_get_names', '(', 'self', ',', 'filename', ',', 'source', ',', 'offset', ')', ':', 'source', '=', 'get_source', '(', 'source', ')', 'if', 'hasattr', '(', 'self', '.', 'backend', ',', '"rpc_get_names"', ')', ':', 'return', 'self', '.', 'backend', '.', 'rpc_get_names', '(', 'filename', ',', 'source', ',', 'of...
Get all possible names
['Get', 'all', 'possible', 'names']
train
https://github.com/jorgenschaefer/elpy/blob/ffd982f829b11e53f2be187c7b770423341f29bc/elpy/server.py#L199-L208
75
praekeltfoundation/seaworthy
seaworthy/helpers.py
fetch_image
def fetch_image(client, name): """ Fetch an image if it isn't already present. This works like ``docker pull`` and will pull the tag ``latest`` if no tag is specified in the image name. """ try: image = client.images.get(name) except docker.errors.ImageNotFound: name, tag = ...
python
def fetch_image(client, name): """ Fetch an image if it isn't already present. This works like ``docker pull`` and will pull the tag ``latest`` if no tag is specified in the image name. """ try: image = client.images.get(name) except docker.errors.ImageNotFound: name, tag = ...
['def', 'fetch_image', '(', 'client', ',', 'name', ')', ':', 'try', ':', 'image', '=', 'client', '.', 'images', '.', 'get', '(', 'name', ')', 'except', 'docker', '.', 'errors', '.', 'ImageNotFound', ':', 'name', ',', 'tag', '=', '_parse_image_tag', '(', 'name', ')', 'tag', '=', "'latest'", 'if', 'tag', 'is', 'None', 'e...
Fetch an image if it isn't already present. This works like ``docker pull`` and will pull the tag ``latest`` if no tag is specified in the image name.
['Fetch', 'an', 'image', 'if', 'it', 'isn', 't', 'already', 'present', '.']
train
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L27-L44
76
F5Networks/f5-common-python
f5/bigip/tm/ltm/pool.py
Members.update
def update(self, **kwargs): """Call this to change the configuration of the service on the device. This method uses HTTP PUT to alter the service state on the device. The attributes of the instance will be packaged as a dictionary. That dictionary will be updated with kwargs. It is t...
python
def update(self, **kwargs): """Call this to change the configuration of the service on the device. This method uses HTTP PUT to alter the service state on the device. The attributes of the instance will be packaged as a dictionary. That dictionary will be updated with kwargs. It is t...
['def', 'update', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'checked', '=', 'self', '.', '_check_member_parameters', '(', '*', '*', 'kwargs', ')', 'return', 'super', '(', 'Members', ',', 'self', ')', '.', '_update', '(', '*', '*', 'checked', ')']
Call this to change the configuration of the service on the device. This method uses HTTP PUT to alter the service state on the device. The attributes of the instance will be packaged as a dictionary. That dictionary will be updated with kwargs. It is then submitted as JSON to the de...
['Call', 'this', 'to', 'change', 'the', 'configuration', 'of', 'the', 'service', 'on', 'the', 'device', '.']
train
https://github.com/F5Networks/f5-common-python/blob/7e67d5acd757a60e3d5f8c88c534bd72208f5494/f5/bigip/tm/ltm/pool.py#L144-L160
77
ethereum/py-evm
eth/estimators/gas.py
binary_gas_search
def binary_gas_search(state: BaseState, transaction: BaseTransaction, tolerance: int=1) -> int: """ Run the transaction with various gas limits, progressively approaching the minimum needed to succeed without an OutOfGas exception. The starting range of possible estimates is: [transaction.intrinsic...
python
def binary_gas_search(state: BaseState, transaction: BaseTransaction, tolerance: int=1) -> int: """ Run the transaction with various gas limits, progressively approaching the minimum needed to succeed without an OutOfGas exception. The starting range of possible estimates is: [transaction.intrinsic...
['def', 'binary_gas_search', '(', 'state', ':', 'BaseState', ',', 'transaction', ':', 'BaseTransaction', ',', 'tolerance', ':', 'int', '=', '1', ')', '->', 'int', ':', 'if', 'not', 'hasattr', '(', 'transaction', ',', "'sender'", ')', ':', 'raise', 'TypeError', '(', '"Transaction is missing attribute sender."', ',', '"I...
Run the transaction with various gas limits, progressively approaching the minimum needed to succeed without an OutOfGas exception. The starting range of possible estimates is: [transaction.intrinsic_gas, state.gas_limit]. After the first OutOfGas exception, the range is: (largest_limit_out_of_gas, sta...
['Run', 'the', 'transaction', 'with', 'various', 'gas', 'limits', 'progressively', 'approaching', 'the', 'minimum', 'needed', 'to', 'succeed', 'without', 'an', 'OutOfGas', 'exception', '.']
train
https://github.com/ethereum/py-evm/blob/58346848f076116381d3274bbcea96b9e2cfcbdf/eth/estimators/gas.py#L28-L78
78
SatelliteQE/nailgun
nailgun/entities.py
ContentViewPuppetModule.read
def read(self, entity=None, attrs=None, ignore=None, params=None): """Provide a default value for ``entity``. By default, ``nailgun.entity_mixins.EntityReadMixin.read provides a default value for ``entity`` like so:: entity = type(self)() However, :class:`ContentViewPuppet...
python
def read(self, entity=None, attrs=None, ignore=None, params=None): """Provide a default value for ``entity``. By default, ``nailgun.entity_mixins.EntityReadMixin.read provides a default value for ``entity`` like so:: entity = type(self)() However, :class:`ContentViewPuppet...
['def', 'read', '(', 'self', ',', 'entity', '=', 'None', ',', 'attrs', '=', 'None', ',', 'ignore', '=', 'None', ',', 'params', '=', 'None', ')', ':', "# read() should not change the state of the object it's called on, but", '# super() alters the attributes of any entity passed in. Creating a new', '# object and passing...
Provide a default value for ``entity``. By default, ``nailgun.entity_mixins.EntityReadMixin.read provides a default value for ``entity`` like so:: entity = type(self)() However, :class:`ContentViewPuppetModule` requires that an ``content_view`` be provided, so this techniq...
['Provide', 'a', 'default', 'value', 'for', 'entity', '.']
train
https://github.com/SatelliteQE/nailgun/blob/c36d8c20862e87bf6975bd48ac1ca40a9e634eaa/nailgun/entities.py#L2453-L2480
79
jordanh/neurio-python
neurio/__init__.py
Client.get_samples_live_last
def get_samples_live_last(self, sensor_id): """Get the last sample recorded by the sensor. Args: sensor_id (string): hexadecimal id of the sensor to query, e.g. ``0x0013A20040B65FAD`` Returns: list: dictionary objects containing sample data """ url = "https://api.neur.io/v1/sam...
python
def get_samples_live_last(self, sensor_id): """Get the last sample recorded by the sensor. Args: sensor_id (string): hexadecimal id of the sensor to query, e.g. ``0x0013A20040B65FAD`` Returns: list: dictionary objects containing sample data """ url = "https://api.neur.io/v1/sam...
['def', 'get_samples_live_last', '(', 'self', ',', 'sensor_id', ')', ':', 'url', '=', '"https://api.neur.io/v1/samples/live/last"', 'headers', '=', 'self', '.', '__gen_headers', '(', ')', 'headers', '[', '"Content-Type"', ']', '=', '"application/json"', 'params', '=', '{', '"sensorId"', ':', 'sensor_id', '}', 'url', '=...
Get the last sample recorded by the sensor. Args: sensor_id (string): hexadecimal id of the sensor to query, e.g. ``0x0013A20040B65FAD`` Returns: list: dictionary objects containing sample data
['Get', 'the', 'last', 'sample', 'recorded', 'by', 'the', 'sensor', '.']
train
https://github.com/jordanh/neurio-python/blob/3a1bcadadb3bb3ad48f2df41c039d8b828ffd9c8/neurio/__init__.py#L439-L458
80
markovmodel/PyEMMA
pyemma/plots/plots2d.py
plot_contour
def plot_contour( xall, yall, zall, ax=None, cmap=None, ncontours=100, vmin=None, vmax=None, levels=None, cbar=True, cax=None, cbar_label=None, cbar_orientation='vertical', norm=None, nbins=100, method='nearest', mask=False, **kwargs): """Plot a two-dimensional contour map by...
python
def plot_contour( xall, yall, zall, ax=None, cmap=None, ncontours=100, vmin=None, vmax=None, levels=None, cbar=True, cax=None, cbar_label=None, cbar_orientation='vertical', norm=None, nbins=100, method='nearest', mask=False, **kwargs): """Plot a two-dimensional contour map by...
['def', 'plot_contour', '(', 'xall', ',', 'yall', ',', 'zall', ',', 'ax', '=', 'None', ',', 'cmap', '=', 'None', ',', 'ncontours', '=', '100', ',', 'vmin', '=', 'None', ',', 'vmax', '=', 'None', ',', 'levels', '=', 'None', ',', 'cbar', '=', 'True', ',', 'cax', '=', 'None', ',', 'cbar_label', '=', 'None', ',', 'cbar_ori...
Plot a two-dimensional contour map by interpolating scattered data on a grid. Parameters ---------- xall : ndarray(T) Sample x-coordinates. yall : ndarray(T) Sample y-coordinates. zall : ndarray(T) Sample z-coordinates. ax : matplotlib.Axes object, optional, default=...
['Plot', 'a', 'two', '-', 'dimensional', 'contour', 'map', 'by', 'interpolating', 'scattered', 'data', 'on', 'a', 'grid', '.']
train
https://github.com/markovmodel/PyEMMA/blob/5c3124398217de05ba5ce9c8fb01519222481ab8/pyemma/plots/plots2d.py#L686-L815
81
tomturner/django-tenants
django_tenants/clone.py
CloneSchema.clone_schema
def clone_schema(self, base_schema_name, new_schema_name): """ Creates a new schema `new_schema_name` as a clone of an existing schema `old_schema_name`. """ connection.set_schema_to_public() cursor = connection.cursor() # check if the clone_schema function alrea...
python
def clone_schema(self, base_schema_name, new_schema_name): """ Creates a new schema `new_schema_name` as a clone of an existing schema `old_schema_name`. """ connection.set_schema_to_public() cursor = connection.cursor() # check if the clone_schema function alrea...
['def', 'clone_schema', '(', 'self', ',', 'base_schema_name', ',', 'new_schema_name', ')', ':', 'connection', '.', 'set_schema_to_public', '(', ')', 'cursor', '=', 'connection', '.', 'cursor', '(', ')', '# check if the clone_schema function already exists in the db', 'try', ':', 'cursor', '.', 'execute', '(', '"SELECT ...
Creates a new schema `new_schema_name` as a clone of an existing schema `old_schema_name`.
['Creates', 'a', 'new', 'schema', 'new_schema_name', 'as', 'a', 'clone', 'of', 'an', 'existing', 'schema', 'old_schema_name', '.']
train
https://github.com/tomturner/django-tenants/blob/f3e06e2b0facee7ed797e5694bcac433df3e5315/django_tenants/clone.py#L213-L233
82
toumorokoshi/jenks
jenks/utils.py
get_configuration_file
def get_configuration_file(): """ return jenks configuration file """ path = os.path.abspath(os.curdir) while path != os.sep: config_path = os.path.join(path, CONFIG_FILE_NAME) if os.path.exists(config_path): return config_path path = os.path.dirname(path) return None
python
def get_configuration_file(): """ return jenks configuration file """ path = os.path.abspath(os.curdir) while path != os.sep: config_path = os.path.join(path, CONFIG_FILE_NAME) if os.path.exists(config_path): return config_path path = os.path.dirname(path) return None
['def', 'get_configuration_file', '(', ')', ':', 'path', '=', 'os', '.', 'path', '.', 'abspath', '(', 'os', '.', 'curdir', ')', 'while', 'path', '!=', 'os', '.', 'sep', ':', 'config_path', '=', 'os', '.', 'path', '.', 'join', '(', 'path', ',', 'CONFIG_FILE_NAME', ')', 'if', 'os', '.', 'path', '.', 'exists', '(', 'confi...
return jenks configuration file
['return', 'jenks', 'configuration', 'file']
train
https://github.com/toumorokoshi/jenks/blob/d3333a7b86ba290b7185aa5b8da75e76a28124f5/jenks/utils.py#L36-L44
83
mitsei/dlkit
dlkit/json_/repository/sessions.py
AssetAdminSession.get_asset_form_for_create
def get_asset_form_for_create(self, asset_record_types): """Gets the asset form for creating new assets. A new form should be requested for each create transaction. arg: asset_record_types (osid.type.Type[]): array of asset record types return: (osid.repository.Asset...
python
def get_asset_form_for_create(self, asset_record_types): """Gets the asset form for creating new assets. A new form should be requested for each create transaction. arg: asset_record_types (osid.type.Type[]): array of asset record types return: (osid.repository.Asset...
['def', 'get_asset_form_for_create', '(', 'self', ',', 'asset_record_types', ')', ':', '# Implemented from template for', '# osid.resource.ResourceAdminSession.get_resource_form_for_create_template', 'for', 'arg', 'in', 'asset_record_types', ':', 'if', 'not', 'isinstance', '(', 'arg', ',', 'ABCType', ')', ':', 'raise',...
Gets the asset form for creating new assets. A new form should be requested for each create transaction. arg: asset_record_types (osid.type.Type[]): array of asset record types return: (osid.repository.AssetForm) - the asset form raise: NullArgument - ``asset_record...
['Gets', 'the', 'asset', 'form', 'for', 'creating', 'new', 'assets', '.']
train
https://github.com/mitsei/dlkit/blob/445f968a175d61c8d92c0f617a3c17dc1dc7c584/dlkit/json_/repository/sessions.py#L1279-L1314
84
martinkosir/neverbounce-python
neverbounce/client.py
NeverBounce.verify
def verify(self, email): """ Verify a single email address. :param str email: Email address to verify. :return: A VerifiedEmail object. """ resp = self._call(endpoint='single', data={'email': email}) return VerifiedEmail(email, resp['result'])
python
def verify(self, email): """ Verify a single email address. :param str email: Email address to verify. :return: A VerifiedEmail object. """ resp = self._call(endpoint='single', data={'email': email}) return VerifiedEmail(email, resp['result'])
['def', 'verify', '(', 'self', ',', 'email', ')', ':', 'resp', '=', 'self', '.', '_call', '(', 'endpoint', '=', "'single'", ',', 'data', '=', '{', "'email'", ':', 'email', '}', ')', 'return', 'VerifiedEmail', '(', 'email', ',', 'resp', '[', "'result'", ']', ')']
Verify a single email address. :param str email: Email address to verify. :return: A VerifiedEmail object.
['Verify', 'a', 'single', 'email', 'address', '.', ':', 'param', 'str', 'email', ':', 'Email', 'address', 'to', 'verify', '.', ':', 'return', ':', 'A', 'VerifiedEmail', 'object', '.']
train
https://github.com/martinkosir/neverbounce-python/blob/8d8b3f381dbff2a753a8770fac0d2bfab80d5bec/neverbounce/client.py#L18-L25
85
JarryShaw/PyPCAPKit
src/protocols/internet/hip.py
HIP._read_para_hip_signature_2
def _read_para_hip_signature_2(self, code, cbit, clen, *, desc, length, version): """Read HIP HIP_SIGNATURE_2 parameter. Structure of HIP HIP_SIGNATURE_2 parameter [RFC 7401]: 0 1 2 3 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9...
python
def _read_para_hip_signature_2(self, code, cbit, clen, *, desc, length, version): """Read HIP HIP_SIGNATURE_2 parameter. Structure of HIP HIP_SIGNATURE_2 parameter [RFC 7401]: 0 1 2 3 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9...
['def', '_read_para_hip_signature_2', '(', 'self', ',', 'code', ',', 'cbit', ',', 'clen', ',', '*', ',', 'desc', ',', 'length', ',', 'version', ')', ':', '_algo', '=', 'self', '.', '_read_unpack', '(', '2', ')', '_sign', '=', 'self', '.', '_read_fileng', '(', 'clen', '-', '2', ')', 'hip_signature_2', '=', 'dict', '(', ...
Read HIP HIP_SIGNATURE_2 parameter. Structure of HIP HIP_SIGNATURE_2 parameter [RFC 7401]: 0 1 2 3 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-...
['Read', 'HIP', 'HIP_SIGNATURE_2', 'parameter', '.']
train
https://github.com/JarryShaw/PyPCAPKit/blob/c7f0da9aebc2cf210bf8f8b912f7d3cbb98ca10e/src/protocols/internet/hip.py#L2023-L2061
86
Preston-Landers/concurrent-log-handler
src/concurrent_log_handler/__init__.py
ConcurrentRotatingFileHandler.emit
def emit(self, record): """ Emit a record. Override from parent class to handle file locking for the duration of rollover and write. This also does the formatting *before* locks are obtained, in case the format itself does logging calls from within. Rollover also occurs while th...
python
def emit(self, record): """ Emit a record. Override from parent class to handle file locking for the duration of rollover and write. This also does the formatting *before* locks are obtained, in case the format itself does logging calls from within. Rollover also occurs while th...
['def', 'emit', '(', 'self', ',', 'record', ')', ':', '# noinspection PyBroadException', 'try', ':', 'msg', '=', 'self', '.', 'format', '(', 'record', ')', 'try', ':', 'self', '.', '_do_lock', '(', ')', 'try', ':', 'if', 'self', '.', 'shouldRollover', '(', 'record', ')', ':', 'self', '.', 'doRollover', '(', ')', 'excep...
Emit a record. Override from parent class to handle file locking for the duration of rollover and write. This also does the formatting *before* locks are obtained, in case the format itself does logging calls from within. Rollover also occurs while the lock is held.
['Emit', 'a', 'record', '.']
train
https://github.com/Preston-Landers/concurrent-log-handler/blob/8e0b8e28c2b12e854853d723b3c28346a3218914/src/concurrent_log_handler/__init__.py#L298-L324
87
Autodesk/aomi
aomi/vault.py
Client.server_version
def server_version(self): """Attempts to determine the version of Vault that a server is running. Some actions will change on older Vault deployments.""" health_url = "%s/v1/sys/health" % self.vault_addr resp = self.session.request('get', health_url, **self._kwargs) if re...
python
def server_version(self): """Attempts to determine the version of Vault that a server is running. Some actions will change on older Vault deployments.""" health_url = "%s/v1/sys/health" % self.vault_addr resp = self.session.request('get', health_url, **self._kwargs) if re...
['def', 'server_version', '(', 'self', ')', ':', 'health_url', '=', '"%s/v1/sys/health"', '%', 'self', '.', 'vault_addr', 'resp', '=', 'self', '.', 'session', '.', 'request', '(', "'get'", ',', 'health_url', ',', '*', '*', 'self', '.', '_kwargs', ')', 'if', 'resp', '.', 'status_code', '==', '200', 'or', 'resp', '.', 's...
Attempts to determine the version of Vault that a server is running. Some actions will change on older Vault deployments.
['Attempts', 'to', 'determine', 'the', 'version', 'of', 'Vault', 'that', 'a', 'server', 'is', 'running', '.', 'Some', 'actions', 'will', 'change', 'on', 'older', 'Vault', 'deployments', '.']
train
https://github.com/Autodesk/aomi/blob/84da2dfb0424837adf9c4ddc1aa352e942bb7a4a/aomi/vault.py#L191-L204
88
ryanmcgrath/twython
twython/streaming/types.py
TwythonStreamerTypes.site
def site(self, **params): """Stream site Accepted params found at: https://dev.twitter.com/docs/api/1.1/get/site """ url = 'https://sitestream.twitter.com/%s/site.json' \ % self.streamer.api_version self.streamer._request(url, params=params)
python
def site(self, **params): """Stream site Accepted params found at: https://dev.twitter.com/docs/api/1.1/get/site """ url = 'https://sitestream.twitter.com/%s/site.json' \ % self.streamer.api_version self.streamer._request(url, params=params)
['def', 'site', '(', 'self', ',', '*', '*', 'params', ')', ':', 'url', '=', "'https://sitestream.twitter.com/%s/site.json'", '%', 'self', '.', 'streamer', '.', 'api_version', 'self', '.', 'streamer', '.', '_request', '(', 'url', ',', 'params', '=', 'params', ')']
Stream site Accepted params found at: https://dev.twitter.com/docs/api/1.1/get/site
['Stream', 'site']
train
https://github.com/ryanmcgrath/twython/blob/7366de80efcbbdfaf615d3f1fea72546196916fc/twython/streaming/types.py#L33-L41
89
rsinger86/drf-flex-fields
rest_flex_fields/serializers.py
FlexFieldsSerializerMixin._get_expand_input
def _get_expand_input(self, passed_settings): """ If expand value is explicitliy passed, just return it. If parsing from request, ensure that the value complies with the "permitted_expands" list passed into the context from the FlexFieldsMixin. """ ...
python
def _get_expand_input(self, passed_settings): """ If expand value is explicitliy passed, just return it. If parsing from request, ensure that the value complies with the "permitted_expands" list passed into the context from the FlexFieldsMixin. """ ...
['def', '_get_expand_input', '(', 'self', ',', 'passed_settings', ')', ':', 'value', '=', 'passed_settings', '.', 'get', '(', '"expand"', ')', 'if', 'len', '(', 'value', ')', '>', '0', ':', 'return', 'value', 'if', 'not', 'self', '.', '_can_access_request', ':', 'return', '[', ']', 'expand', '=', 'self', '.', '_parse_r...
If expand value is explicitliy passed, just return it. If parsing from request, ensure that the value complies with the "permitted_expands" list passed into the context from the FlexFieldsMixin.
['If', 'expand', 'value', 'is', 'explicitliy', 'passed', 'just', 'return', 'it', '.', 'If', 'parsing', 'from', 'request', 'ensure', 'that', 'the', 'value', 'complies', 'with', 'the', 'permitted_expands', 'list', 'passed', 'into', 'the', 'context', 'from', 'the', 'FlexFieldsMixin', '.']
train
https://github.com/rsinger86/drf-flex-fields/blob/56495f15977d76697972acac571792e8fd67003d/rest_flex_fields/serializers.py#L196-L221
90
eyurtsev/FlowCytometryTools
FlowCytometryTools/gui/fc_widget.py
BaseVertex.update_coordinates
def update_coordinates(self, new_coordinates): """ new_coordinates : dict """ for k, v in new_coordinates.items(): if k in self.coordinates: self.coordinates[k] = v for svertex in self.spawn_list: verts = tuple([self.coordinates.get(ch, No...
python
def update_coordinates(self, new_coordinates): """ new_coordinates : dict """ for k, v in new_coordinates.items(): if k in self.coordinates: self.coordinates[k] = v for svertex in self.spawn_list: verts = tuple([self.coordinates.get(ch, No...
['def', 'update_coordinates', '(', 'self', ',', 'new_coordinates', ')', ':', 'for', 'k', ',', 'v', 'in', 'new_coordinates', '.', 'items', '(', ')', ':', 'if', 'k', 'in', 'self', '.', 'coordinates', ':', 'self', '.', 'coordinates', '[', 'k', ']', '=', 'v', 'for', 'svertex', 'in', 'self', '.', 'spawn_list', ':', 'verts',...
new_coordinates : dict
['new_coordinates', ':', 'dict']
train
https://github.com/eyurtsev/FlowCytometryTools/blob/4355632508b875273d68c7e2972c17668bcf7b40/FlowCytometryTools/gui/fc_widget.py#L150-L165
91
abelcarreras/DynaPhoPy
dynaphopy/analysis/fitting/fitting_functions.py
Gaussian_function._function
def _function(self, x, a, b, c, d): """Gaussian PDF function x: coordinate a: peak position b: deviation (sigma) c: area proportional parameter d: base line """ return c/b*np.sqrt(2*np.pi)*np.exp(-(x-a)**2/(2*b**2))+d
python
def _function(self, x, a, b, c, d): """Gaussian PDF function x: coordinate a: peak position b: deviation (sigma) c: area proportional parameter d: base line """ return c/b*np.sqrt(2*np.pi)*np.exp(-(x-a)**2/(2*b**2))+d
['def', '_function', '(', 'self', ',', 'x', ',', 'a', ',', 'b', ',', 'c', ',', 'd', ')', ':', 'return', 'c', '/', 'b', '*', 'np', '.', 'sqrt', '(', '2', '*', 'np', '.', 'pi', ')', '*', 'np', '.', 'exp', '(', '-', '(', 'x', '-', 'a', ')', '**', '2', '/', '(', '2', '*', 'b', '**', '2', ')', ')', '+', 'd']
Gaussian PDF function x: coordinate a: peak position b: deviation (sigma) c: area proportional parameter d: base line
['Gaussian', 'PDF', 'function', 'x', ':', 'coordinate', 'a', ':', 'peak', 'position', 'b', ':', 'deviation', '(', 'sigma', ')', 'c', ':', 'area', 'proportional', 'parameter', 'd', ':', 'base', 'line']
train
https://github.com/abelcarreras/DynaPhoPy/blob/51e99422228e6be84830d659b88a0ca904d9136f/dynaphopy/analysis/fitting/fitting_functions.py#L302-L310
92
woolfson-group/isambard
isambard/optimisation/evo_optimizers.py
CMAES._initial_individual
def _initial_individual(self): """Generates an individual with random parameters within bounds.""" ind = creator.Individual( [random.uniform(-1, 1) for _ in range(len(self.value_means))]) return ind
python
def _initial_individual(self): """Generates an individual with random parameters within bounds.""" ind = creator.Individual( [random.uniform(-1, 1) for _ in range(len(self.value_means))]) return ind
['def', '_initial_individual', '(', 'self', ')', ':', 'ind', '=', 'creator', '.', 'Individual', '(', '[', 'random', '.', 'uniform', '(', '-', '1', ',', '1', ')', 'for', '_', 'in', 'range', '(', 'len', '(', 'self', '.', 'value_means', ')', ')', ']', ')', 'return', 'ind']
Generates an individual with random parameters within bounds.
['Generates', 'an', 'individual', 'with', 'random', 'parameters', 'within', 'bounds', '.']
train
https://github.com/woolfson-group/isambard/blob/ebc33b48a28ad217e18f93b910dfba46e6e71e07/isambard/optimisation/evo_optimizers.py#L472-L477
93
yfpeng/bioc
bioc/biocxml/encoder.py
BioCXMLDocumentWriter.write_collection_info
def write_collection_info(self, collection: BioCCollection): """ Writes the collection information: encoding, version, DTD, source, date, key, infons, etc. """ elem = etree.Element('source') elem.text = collection.source self.__writer.send(elem) elem = et...
python
def write_collection_info(self, collection: BioCCollection): """ Writes the collection information: encoding, version, DTD, source, date, key, infons, etc. """ elem = etree.Element('source') elem.text = collection.source self.__writer.send(elem) elem = et...
['def', 'write_collection_info', '(', 'self', ',', 'collection', ':', 'BioCCollection', ')', ':', 'elem', '=', 'etree', '.', 'Element', '(', "'source'", ')', 'elem', '.', 'text', '=', 'collection', '.', 'source', 'self', '.', '__writer', '.', 'send', '(', 'elem', ')', 'elem', '=', 'etree', '.', 'Element', '(', "'date'"...
Writes the collection information: encoding, version, DTD, source, date, key, infons, etc.
['Writes', 'the', 'collection', 'information', ':', 'encoding', 'version', 'DTD', 'source', 'date', 'key', 'infons', 'etc', '.']
train
https://github.com/yfpeng/bioc/blob/47ddaa010960d9ba673aefe068e7bbaf39f0fff4/bioc/biocxml/encoder.py#L185-L204
94
pymc-devs/pymc
pymc/Matplot.py
autocorrelation
def autocorrelation( data, name, maxlags=100, format='png', reflected=False, suffix='-acf', path='./', fontmap=None, new=True, last=True, rows=1, columns=1, num=1, verbose=1): """ Generate bar plot of the autocorrelation function for a series (usually an MCMC trace). :Arguments: data: P...
python
def autocorrelation( data, name, maxlags=100, format='png', reflected=False, suffix='-acf', path='./', fontmap=None, new=True, last=True, rows=1, columns=1, num=1, verbose=1): """ Generate bar plot of the autocorrelation function for a series (usually an MCMC trace). :Arguments: data: P...
['def', 'autocorrelation', '(', 'data', ',', 'name', ',', 'maxlags', '=', '100', ',', 'format', '=', "'png'", ',', 'reflected', '=', 'False', ',', 'suffix', '=', "'-acf'", ',', 'path', '=', "'./'", ',', 'fontmap', '=', 'None', ',', 'new', '=', 'True', ',', 'last', '=', 'True', ',', 'rows', '=', '1', ',', 'columns', '='...
Generate bar plot of the autocorrelation function for a series (usually an MCMC trace). :Arguments: data: PyMC object, trace or array A trace from an MCMC sample or a PyMC object with one or more traces. name: string The name of the object. maxlags (optional): int ...
['Generate', 'bar', 'plot', 'of', 'the', 'autocorrelation', 'function', 'for', 'a', 'series', '(', 'usually', 'an', 'MCMC', 'trace', ')', '.']
train
https://github.com/pymc-devs/pymc/blob/c6e530210bff4c0d7189b35b2c971bc53f93f7cd/pymc/Matplot.py#L899-L997
95
pywbem/pywbem
pywbem/cim_obj.py
CIMProperty.tocimxml
def tocimxml(self): """ Return the CIM-XML representation of this CIM property, as an object of an appropriate subclass of :term:`Element`. The returned CIM-XML representation is a `PROPERTY`, `PROPERTY.REFERENCE`, or `PROPERTY.ARRAY` element dependent on the property ty...
python
def tocimxml(self): """ Return the CIM-XML representation of this CIM property, as an object of an appropriate subclass of :term:`Element`. The returned CIM-XML representation is a `PROPERTY`, `PROPERTY.REFERENCE`, or `PROPERTY.ARRAY` element dependent on the property ty...
['def', 'tocimxml', '(', 'self', ')', ':', 'qualifiers', '=', '[', 'q', '.', 'tocimxml', '(', ')', 'for', 'q', 'in', 'self', '.', 'qualifiers', '.', 'values', '(', ')', ']', 'if', 'self', '.', 'is_array', ':', '# pylint: disable=no-else-return', 'assert', 'self', '.', 'type', '!=', "'reference'", 'if', 'self', '.', 'va...
Return the CIM-XML representation of this CIM property, as an object of an appropriate subclass of :term:`Element`. The returned CIM-XML representation is a `PROPERTY`, `PROPERTY.REFERENCE`, or `PROPERTY.ARRAY` element dependent on the property type, and consistent with :term:`DSP0201`....
['Return', 'the', 'CIM', '-', 'XML', 'representation', 'of', 'this', 'CIM', 'property', 'as', 'an', 'object', 'of', 'an', 'appropriate', 'subclass', 'of', ':', 'term', ':', 'Element', '.']
train
https://github.com/pywbem/pywbem/blob/e54ecb82c2211e289a268567443d60fdd489f1e4/pywbem/cim_obj.py#L4883-L4967
96
rocky/python-spark
spark_parser/spark.py
GenericParser.get_profile_info
def get_profile_info(self): """Show the accumulated results of how many times each rule was used""" return sorted(self.profile_info.items(), key=lambda kv: kv[1], reverse=False) return
python
def get_profile_info(self): """Show the accumulated results of how many times each rule was used""" return sorted(self.profile_info.items(), key=lambda kv: kv[1], reverse=False) return
['def', 'get_profile_info', '(', 'self', ')', ':', 'return', 'sorted', '(', 'self', '.', 'profile_info', '.', 'items', '(', ')', ',', 'key', '=', 'lambda', 'kv', ':', 'kv', '[', '1', ']', ',', 'reverse', '=', 'False', ')', 'return']
Show the accumulated results of how many times each rule was used
['Show', 'the', 'accumulated', 'results', 'of', 'how', 'many', 'times', 'each', 'rule', 'was', 'used']
train
https://github.com/rocky/python-spark/blob/8899954bcf0e166726841a43e87c23790eb3441f/spark_parser/spark.py#L983-L988
97
scivision/pymap3d
pymap3d/sidereal.py
juliandate
def juliandate(time: datetime) -> float: """ Python datetime to Julian time from D.Vallado Fundamentals of Astrodynamics and Applications p.187 and J. Meeus Astronomical Algorithms 1991 Eqn. 7.1 pg. 61 Parameters ---------- time : datetime.datetime time to convert Results ...
python
def juliandate(time: datetime) -> float: """ Python datetime to Julian time from D.Vallado Fundamentals of Astrodynamics and Applications p.187 and J. Meeus Astronomical Algorithms 1991 Eqn. 7.1 pg. 61 Parameters ---------- time : datetime.datetime time to convert Results ...
['def', 'juliandate', '(', 'time', ':', 'datetime', ')', '->', 'float', ':', 'times', '=', 'np', '.', 'atleast_1d', '(', 'time', ')', 'assert', 'times', '.', 'ndim', '==', '1', 'jd', '=', 'np', '.', 'empty', '(', 'times', '.', 'size', ')', 'for', 'i', ',', 't', 'in', 'enumerate', '(', 'times', ')', ':', 'if', 't', '.',...
Python datetime to Julian time from D.Vallado Fundamentals of Astrodynamics and Applications p.187 and J. Meeus Astronomical Algorithms 1991 Eqn. 7.1 pg. 61 Parameters ---------- time : datetime.datetime time to convert Results ------- jd : float Julian date
['Python', 'datetime', 'to', 'Julian', 'time']
train
https://github.com/scivision/pymap3d/blob/c9cf676594611cdb52ff7e0eca6388c80ed4f63f/pymap3d/sidereal.py#L58-L97
98
Phylliade/ikpy
contrib/transformations.py
pose_to_list
def pose_to_list(pose): """ Convert a Pose or PoseStamped in Python list ((position), (quaternion)) :param pose: geometry_msgs.msg.PoseStamped or geometry_msgs.msg.Pose :return: the equivalent in list ((position), (quaternion)) """ if type(pose) == geometry_msgs.msg.PoseStamped: return [...
python
def pose_to_list(pose): """ Convert a Pose or PoseStamped in Python list ((position), (quaternion)) :param pose: geometry_msgs.msg.PoseStamped or geometry_msgs.msg.Pose :return: the equivalent in list ((position), (quaternion)) """ if type(pose) == geometry_msgs.msg.PoseStamped: return [...
['def', 'pose_to_list', '(', 'pose', ')', ':', 'if', 'type', '(', 'pose', ')', '==', 'geometry_msgs', '.', 'msg', '.', 'PoseStamped', ':', 'return', '[', '[', 'pose', '.', 'pose', '.', 'position', '.', 'x', ',', 'pose', '.', 'pose', '.', 'position', '.', 'y', ',', 'pose', '.', 'pose', '.', 'position', '.', 'z', ']', ',...
Convert a Pose or PoseStamped in Python list ((position), (quaternion)) :param pose: geometry_msgs.msg.PoseStamped or geometry_msgs.msg.Pose :return: the equivalent in list ((position), (quaternion))
['Convert', 'a', 'Pose', 'or', 'PoseStamped', 'in', 'Python', 'list', '((', 'position', ')', '(', 'quaternion', '))', ':', 'param', 'pose', ':', 'geometry_msgs', '.', 'msg', '.', 'PoseStamped', 'or', 'geometry_msgs', '.', 'msg', '.', 'Pose', ':', 'return', ':', 'the', 'equivalent', 'in', 'list', '((', 'position', ')', ...
train
https://github.com/Phylliade/ikpy/blob/60e36d6163136942bf520d952db17123c658d0b6/contrib/transformations.py#L48-L68
99
ARMmbed/mbed-cloud-sdk-python
src/mbed_cloud/device_directory/device_directory.py
DeviceDirectoryAPI.list_device_events
def list_device_events(self, **kwargs): """List all device logs. :param int limit: The number of logs to retrieve. :param str order: The ordering direction, ascending (asc) or descending (desc) :param str after: Get logs after/starting at given `device_event_id` :par...
python
def list_device_events(self, **kwargs): """List all device logs. :param int limit: The number of logs to retrieve. :param str order: The ordering direction, ascending (asc) or descending (desc) :param str after: Get logs after/starting at given `device_event_id` :par...
['def', 'list_device_events', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'kwargs', '=', 'self', '.', '_verify_sort_options', '(', 'kwargs', ')', 'kwargs', '=', 'self', '.', '_verify_filters', '(', 'kwargs', ',', 'DeviceEvent', ',', 'True', ')', 'api', '=', 'self', '.', '_get_api', '(', 'device_directory', '.', 'De...
List all device logs. :param int limit: The number of logs to retrieve. :param str order: The ordering direction, ascending (asc) or descending (desc) :param str after: Get logs after/starting at given `device_event_id` :param dict filters: Dictionary of filters to apply. ...
['List', 'all', 'device', 'logs', '.']
train
https://github.com/ARMmbed/mbed-cloud-sdk-python/blob/c0af86fb2cdd4dc7ed26f236139241067d293509/src/mbed_cloud/device_directory/device_directory.py#L290-L305