signature
stringlengths 29
44.1k
| implementation
stringlengths 0
85.2k
|
---|---|
def progression_linear ( week , start_weight , final_weight , start_week , end_week ) :
"""A linear progression function going through the points
( ' start _ week ' , ' start _ weight ' ) and ( ' end _ week ' , ' final _ weight ' ) , evaluated
in ' week ' .
Parameters
week
The week to evaluate the linear function at .
start _ weight
The weight at ' start _ week ' .
final _ weight
The weight at ' end _ week ' .
start _ week
The number of the first week , typically 1.
end _ week
The number of the final week , e . g . 8.
Returns
weight
The weight at ' week ' .
Examples
> > > progression _ linear ( week = 2 , start _ weight = 100 , final _ weight = 120,
. . . start _ week = 1 , end _ week = 3)
110.0
> > > progression _ linear ( 3 , 100 , 140 , 1 , 5)
120.0""" | # Calculate the slope of the linear function
slope = ( start_weight - final_weight ) / ( start_week - end_week )
# Return the answer y = slope ( x - x _ 0 ) + y _ 0
return slope * ( week - start_week ) + start_weight |
def _replace_image ( image_url , image_tag , ebook_folder , image_name = None ) :
"""Replaces the src of an image to link to the local copy in the images folder of the ebook . Tightly coupled with bs4
package .
Args :
image _ url ( str ) : The url of the image .
image _ tag ( bs4 . element . Tag ) : The bs4 tag containing the image .
ebook _ folder ( str ) : The directory where the ebook files are being saved . This must contain a subdirectory
called " images " .
image _ name ( Option [ str ] ) : The short name to save the image as . Should not contain a directory or an extension .""" | try :
assert isinstance ( image_tag , bs4 . element . Tag )
except AssertionError :
raise TypeError ( "image_tag cannot be of type " + str ( type ( image_tag ) ) )
if image_name is None :
image_name = str ( uuid . uuid4 ( ) )
try :
image_full_path = os . path . join ( ebook_folder , 'images' )
assert os . path . exists ( image_full_path )
image_extension = save_image ( image_url , image_full_path , image_name )
image_tag [ 'src' ] = 'images' + '/' + image_name + '.' + image_extension
except ImageErrorException :
image_tag . decompose ( )
except AssertionError :
raise ValueError ( '%s doesn\'t exist or doesn\'t contain a subdirectory images' % ebook_folder )
except TypeError :
image_tag . decompose ( ) |
def multi_layer_feature ( body , from_layers , num_filters , strides , pads , min_filter = 128 ) :
"""Wrapper function to extract features from base network , attaching extra
layers and SSD specific layers
Parameters
from _ layers : list of str
feature extraction layers , use ' ' for add extra layers
For example :
from _ layers = [ ' relu4_3 ' , ' fc7 ' , ' ' , ' ' , ' ' , ' ' ]
which means extract feature from relu4_3 and fc7 , adding 4 extra layers
on top of fc7
num _ filters : list of int
number of filters for extra layers , you can use - 1 for extracted features ,
however , if normalization and scale is applied , the number of filter for
that layer must be provided .
For example :
num _ filters = [ 512 , - 1 , 512 , 256 , 256 , 256]
strides : list of int
strides for the 3x3 convolution appended , - 1 can be used for extracted
feature layers
pads : list of int
paddings for the 3x3 convolution , - 1 can be used for extracted layers
min _ filter : int
minimum number of filters used in 1x1 convolution
Returns
list of mx . Symbols""" | # arguments check
assert len ( from_layers ) > 0
assert isinstance ( from_layers [ 0 ] , str ) and len ( from_layers [ 0 ] . strip ( ) ) > 0
assert len ( from_layers ) == len ( num_filters ) == len ( strides ) == len ( pads )
internals = body . get_internals ( )
layers = [ ]
for k , params in enumerate ( zip ( from_layers , num_filters , strides , pads ) ) :
from_layer , num_filter , s , p = params
if from_layer . strip ( ) : # extract from base network
layer = internals [ from_layer . strip ( ) + '_output' ]
layers . append ( layer )
else : # attach from last feature layer
assert len ( layers ) > 0
assert num_filter > 0
layer = layers [ - 1 ]
num_1x1 = max ( min_filter , num_filter // 2 )
conv_1x1 = conv_act_layer ( layer , 'multi_feat_%d_conv_1x1' % ( k ) , num_1x1 , kernel = ( 1 , 1 ) , pad = ( 0 , 0 ) , stride = ( 1 , 1 ) , act_type = 'relu' )
conv_3x3 = conv_act_layer ( conv_1x1 , 'multi_feat_%d_conv_3x3' % ( k ) , num_filter , kernel = ( 3 , 3 ) , pad = ( p , p ) , stride = ( s , s ) , act_type = 'relu' )
layers . append ( conv_3x3 )
return layers |
def cell_styles ( self ) :
"""dict of { ( row name , col name ) : style }""" | styles = { }
for colname , col in self . dataframe . items ( ) :
for rowname , value in col . items ( ) :
if isinstance ( value , Value ) and value . style is not None :
style = value . style
if not isinstance ( style , CellStyle ) :
style = self . _named_styles [ style ]
styles [ ( rowname , colname ) ] = style
return styles |
def get_error_probability ( self ) :
"""This means for the base we are talking about how many errors between 0 and 1 do we attribute to it ?
For the ' unobserved ' errors , these can only count when one is adjacent to base
: returns : error probability p ( error _ observed ) + ( 1 - p _ error _ observed ) * error _ unobserved
: rtype : float""" | a = self . _observable . get_error_probability ( )
b = self . _unobservable . get_error_probability ( )
return a + ( 1 - a ) * b |
def add_reorganize_data ( self , name , input_name , output_name , mode = 'SPACE_TO_DEPTH' , block_size = 2 ) :
"""Add a data reorganization layer of type " SPACE _ TO _ DEPTH " or " DEPTH _ TO _ SPACE " .
Parameters
name : str
The name of this layer .
input _ name : str
The input blob name of this layer .
output _ name : str
The output blob name of this layer .
mode : str
- If mode = = ' SPACE _ TO _ DEPTH ' : data is moved from the spatial to the channel dimension .
Input is spatially divided into non - overlapping blocks of size block _ size X block _ size
and data from each block is moved to the channel dimension .
Output CHW dimensions are : [ C * block _ size * block _ size , H / block _ size , C / block _ size ] .
- If mode = = ' DEPTH _ TO _ SPACE ' : data is moved from the channel to the spatial dimension .
Reverse of the operation ' SPACE _ TO _ DEPTH ' .
Output CHW dimensions are : [ C / ( block _ size * block _ size ) , H * block _ size , C * block _ size ] .
block _ size : int
Must be greater than 1 . Must divide H and W , when mode is ' SPACE _ TO _ DEPTH ' . ( block _ size * block _ size )
must divide C when mode is ' DEPTH _ TO _ SPACE ' .
See Also
add _ flatten , add _ reshape""" | spec = self . spec
nn_spec = self . nn_spec
# Add a new layer
spec_layer = nn_spec . layers . add ( )
spec_layer . name = name
spec_layer . input . append ( input_name )
spec_layer . output . append ( output_name )
spec_layer_params = spec_layer . reorganizeData
# Set the parameters
if block_size < 2 :
raise ValueError ( "Invalid block_size value %d. Must be greater than 1." % block_size )
spec_layer_params . blockSize = block_size
if mode == 'SPACE_TO_DEPTH' :
spec_layer_params . mode = _NeuralNetwork_pb2 . ReorganizeDataLayerParams . ReorganizationType . Value ( 'SPACE_TO_DEPTH' )
elif mode == 'DEPTH_TO_SPACE' :
spec_layer_params . mode = _NeuralNetwork_pb2 . ReorganizeDataLayerParams . ReorganizationType . Value ( 'DEPTH_TO_SPACE' )
else :
raise NotImplementedError ( 'Unknown reorganization mode %s ' % mode ) |
def connect_widget ( self , wid , getter = None , setter = None , signal = None , arg = None , update = True , flavour = None ) :
"""Finish set - up by connecting the widget . The model was already
specified in the constructor .
* wid * is a widget instance .
* getter * is a callable . It is passed * wid * and must return its
current value .
* setter * is a callable . It is passed * wid * and the current value of
the model property and must update the widget .
* signal * is a string naming the signal to connect to on * wid * . When
it is emitted we update the model .
* getter * , * setter * and * signal * are optional . Missing values are
guessed from * wid * using
: meth : ` gtkmvc3 . adapters . default . search _ adapter _ info ` . If nothing is
found this raises : exc : ` TypeError ` .
* arg * is an optional value passed to the handler for * signal * . This
doesn ' t do anything unless a subclass overrides the handler .
* update * denotes whether to update the widget from the model
immediately . Otherwise the widget stays unchanged until the first
notification .
* flavour * can be used to select special behaviours about
the adaptation when twice or more possibilities are
possibly handled for the same widget type . See
adapters . default for further information .""" | if wid in self . _wid_info :
raise ValueError ( "Widget " + str ( wid ) + " was already connected" )
wid_type = None
if None in ( getter , setter , signal ) :
w = search_adapter_info ( wid , flavour )
if getter is None :
getter = w [ GETTER ]
if setter is None :
setter = w [ SETTER ]
wid_type = w [ WIDTYPE ]
if signal is None :
signal = w [ SIGNAL ]
# saves information about the widget
self . _wid_info [ wid ] = ( getter , setter , wid_type )
# connects the widget
if signal :
if arg :
wid . connect ( signal , self . _on_wid_changed , arg )
else :
wid . connect ( signal , self . _on_wid_changed )
self . _wid = wid
# updates the widget :
if update :
self . update_widget ( ) |
def get_state_search_path_list ( saltenv = 'base' ) :
'''For the state file system , return a list of paths to search for states''' | # state cache should be updated before running this method
search_list = [ ]
cachedir = __opts__ . get ( 'cachedir' , None )
log . info ( "Searching for files in saltenv: %s" , saltenv )
path = cachedir + os . sep + "files" + os . sep + saltenv
search_list . append ( path )
return search_list |
def update_ids ( self , docids ) :
"""Update id - > pos mapping with new document ids .""" | logger . info ( "updating %i id mappings" % len ( docids ) )
for docid in docids :
if docid is not None :
pos = self . id2pos . get ( docid , None )
if pos is not None :
logger . info ( "replacing existing document %r in %s" % ( docid , self ) )
del self . pos2id [ pos ]
self . id2pos [ docid ] = self . length
try :
del self . id2sims [ docid ]
except :
pass
self . length += 1
self . id2sims . sync ( )
self . update_mappings ( ) |
def background_bin_from_string ( background_bins , data ) :
"""Return template ids for each bin as defined by the format string
Parameters
bins : list of strings
List of strings which define how a background bin is taken from the
list of templates .
data : dict of numpy . ndarrays
Dict with parameter key values and numpy . ndarray values which define
the parameters of the template bank to bin up .
Returns
bins : dict
Dictionary of location indices indexed by a bin name""" | used = numpy . array ( [ ] , dtype = numpy . uint32 )
bins = { }
for mbin in background_bins :
name , bin_type , boundary = tuple ( mbin . split ( ':' ) )
if boundary [ 0 : 2 ] == 'lt' :
member_func = lambda vals , bd = boundary : vals < float ( bd [ 2 : ] )
elif boundary [ 0 : 2 ] == 'gt' :
member_func = lambda vals , bd = boundary : vals > float ( bd [ 2 : ] )
else :
raise RuntimeError ( "Can't parse boundary condition! Must begin " "with 'lt' or 'gt'" )
if bin_type == 'component' and boundary [ 0 : 2 ] == 'lt' : # maximum component mass is less than boundary value
vals = numpy . maximum ( data [ 'mass1' ] , data [ 'mass2' ] )
if bin_type == 'component' and boundary [ 0 : 2 ] == 'gt' : # minimum component mass is greater than bdary
vals = numpy . minimum ( data [ 'mass1' ] , data [ 'mass2' ] )
elif bin_type == 'total' :
vals = data [ 'mass1' ] + data [ 'mass2' ]
elif bin_type == 'chirp' :
vals = pycbc . pnutils . mass1_mass2_to_mchirp_eta ( data [ 'mass1' ] , data [ 'mass2' ] ) [ 0 ]
elif bin_type == 'SEOBNRv2Peak' :
vals = pycbc . pnutils . get_freq ( 'fSEOBNRv2Peak' , data [ 'mass1' ] , data [ 'mass2' ] , data [ 'spin1z' ] , data [ 'spin2z' ] )
elif bin_type == 'SEOBNRv4Peak' :
vals = pycbc . pnutils . get_freq ( 'fSEOBNRv4Peak' , data [ 'mass1' ] , data [ 'mass2' ] , data [ 'spin1z' ] , data [ 'spin2z' ] )
elif bin_type == 'SEOBNRv2duration' :
vals = pycbc . pnutils . get_imr_duration ( data [ 'mass1' ] , data [ 'mass2' ] , data [ 'spin1z' ] , data [ 'spin2z' ] , data [ 'f_lower' ] , approximant = 'SEOBNRv2' )
else :
raise ValueError ( 'Invalid bin type %s' % bin_type )
locs = member_func ( vals )
del vals
# make sure we don ' t reuse anything from an earlier bin
locs = numpy . where ( locs ) [ 0 ]
locs = numpy . delete ( locs , numpy . where ( numpy . in1d ( locs , used ) ) [ 0 ] )
used = numpy . concatenate ( [ used , locs ] )
bins [ name ] = locs
return bins |
def _docstring_key ( self , line ) :
"""Returns the key to use for the docblock immediately preceding
the specified line .""" | decormatch = self . docparser . RE_DECOR . match ( line )
if decormatch is not None :
key = "{}.{}" . format ( self . docelement . name , decormatch . group ( "name" ) )
else :
key = self . element . name
return key |
def print_tweets ( tweets ) :
"""Print a list of tweets one by one separated by " = " s .
Parameters
tweets : list ( dict )
A list of tweets . Each tweet is a dict containing the username of the tweet ' s author ,
the post time , and the tweet body .""" | print ( '=' * 60 )
for index , tweet in enumerate ( tweets ) :
print ( '-' * 60 )
print ( 'Tweet {}:' . format ( index ) )
print ( 'Username:' , tweet [ pytwis_constants . USERNAME_KEY ] )
print ( 'Time:' , datetime . datetime . fromtimestamp ( int ( tweet [ pytwis_constants . TWEET_UNIXTIME_KEY ] ) ) . strftime ( '%Y-%m-%d %H:%M:%S' ) )
print ( 'Body:\n\t' , tweet [ pytwis_constants . TWEET_BODY_KEY ] )
print ( '-' * 60 )
print ( '=' * 60 ) |
def send_metrics_to_cloudwatch ( self , rule , metric , dimensions ) :
"""Send metrics to CloudWatch for the given dimensions""" | timestamp = datetime . datetime . utcfromtimestamp ( metric . timestamp )
self . log . debug ( "CloudWatch: Attempting to publish metric: %s to %s " "with value (%s) for dimensions %s @%s" , rule [ 'name' ] , rule [ 'namespace' ] , str ( metric . value ) , str ( dimensions ) , str ( metric . timestamp ) )
try :
self . connection . put_metric_data ( str ( rule [ 'namespace' ] ) , str ( rule [ 'name' ] ) , str ( metric . value ) , timestamp , str ( rule [ 'unit' ] ) , dimensions )
self . log . debug ( "CloudWatch: Successfully published metric: %s to" " %s with value (%s) for dimensions %s" , rule [ 'name' ] , rule [ 'namespace' ] , str ( metric . value ) , str ( dimensions ) )
except AttributeError as e :
self . log . error ( "CloudWatch: Failed publishing - %s " , str ( e ) )
except Exception as e : # Rough connection re - try logic .
self . log . error ( "CloudWatch: Failed publishing - %s\n%s " , str ( e ) , str ( sys . exc_info ( ) [ 0 ] ) )
self . _bind ( ) |
def parse_signature ( self , statement , element , module = None ) :
"""Parses the specified line as a new version of the signature for ' element ' .
: arg statement : the string that has the new signature .
: arg element : the code element whose signature will be changed .""" | # If the signature changes , the user might not have had a chance to add the
# detailed member information for it yet . Here
# we will just update the modifiers and attributes . Also , since all the mods
# etc . will be overwritten , we don ' t need to handle replace separately .
smatch = self . RE_SIG . match ( statement )
result = ( None , None , None )
eresult = None
if smatch is not None :
name = smatch . group ( "name" ) . strip ( )
modifiers = smatch . group ( "modifiers" ) or [ ]
codetype = smatch . group ( "codetype" )
# If the exec is a function , we also may have a type and kind specified .
if codetype . lower ( ) == "function" :
dtype = smatch . group ( "type" )
kind = smatch . group ( "kind" )
if module is None :
element . update ( name , modifiers , dtype , kind )
else :
eresult = Function ( name , modifiers , dtype , kind , module )
else :
if module is None :
element . update ( name , modifiers )
else :
eresult = Subroutine ( name , modifiers , module )
# The parameter sets are actually driven by the body of the executable
# rather than the call signature . However , the declarations will be
# interpreted as members if we don ' t add the parameters to the ordered
# list of parameter names . Overwrite that list with the new names .
params = re . split ( "[\s,]+" , smatch . group ( "parameters" ) . lower ( ) )
if eresult is None :
element . paramorder = params
else :
eresult . paramorder = params
result = ( eresult , smatch . start ( ) , smatch . end ( ) )
return result |
def compute_results ( self , voting_method , votes = None , winners = 1 , ** kwargs ) :
"""Compute voting results to decide the winner ( s ) from the
: attr : ` votes ` .
The votes should have been made for the current
: attr : ` ~ creamas . vote . VoteOrganizer . candidates ` .
: param voting _ method :
A function which computes the results from the votes . Should
accept at least three parameters : candidates , votes and number of
vote winners . The function should return at least a list of vote
winners . See , e . g . : func : ` ~ creamas . vote . vote _ mean ` or
: func : ` ~ creamas . vote . vote _ best ` . Additional ` ` * * kwargs ` ` are passed
down to the voting method .
: param list votes :
A list of votes by which the voting is performed . Each vote should
have the same set of artifacts in them . If ` ` None ` ` the results
are computed for the current list of
: attr : ` ~ creamas . vote . VoteOrganizer . votes ` .
: param int winners :
The number of vote winners
: returns :
list of : py : class : ` ~ creamas . core . artifact . Artifact ` objects ,
the winning artifacts . Some voting methods may also return a score
associated with each winning artifact .
: rtype : list""" | if votes is None :
votes = self . votes
if len ( votes ) == 0 :
self . _log ( logging . DEBUG , "Could not compute results as there are " "no votes!" )
return [ ]
self . _log ( logging . DEBUG , "Computing results from {} votes." . format ( len ( votes ) ) )
return voting_method ( self . candidates , votes , winners , ** kwargs ) |
def reporter ( self ) :
"""Creates a report of the results""" | logging . info ( 'Creating {} report' . format ( self . analysistype ) )
# Create the path in which the reports are stored
make_path ( self . reportpath )
header = 'Strain,Serotype\n'
data = ''
with open ( os . path . join ( self . reportpath , '{}.csv' . format ( self . analysistype ) ) , 'w' ) as report :
for sample in self . runmetadata . samples :
if sample . general . bestassemblyfile != 'NA' :
data += sample . name + ','
if sample [ self . analysistype ] . results : # Set the O - type as either the appropriate attribute , or O - untypable
if ';' . join ( sample . serosippr . o_set ) == '-' :
otype = 'O-untypeable'
else :
otype = '{oset} ({opid})' . format ( oset = ';' . join ( sample . serosippr . o_set ) , opid = sample . serosippr . best_o_pid )
# Same as above , but for the H - type
if ';' . join ( sample . serosippr . h_set ) == '-' :
htype = 'H-untypeable'
else :
htype = '{hset} ({hpid})' . format ( hset = ';' . join ( sample . serosippr . h_set ) , hpid = sample . serosippr . best_h_pid )
serotype = '{otype}:{htype}' . format ( otype = otype , htype = htype )
# Populate the data string
data += serotype if serotype != 'O-untypeable:H-untypeable' else 'ND'
data += '\n'
else :
data += '\n'
report . write ( header )
report . write ( data ) |
def _create_config_signature ( config ) :
"""return the signature for a config object .
The signature is computed as sha1 digest of the contents of
working _ directory , include _ paths , define _ symbols and
undefine _ symbols .
: param config : Configuration object
: type config : : class : ` parser . xml _ generator _ configuration _ t `
: rtype : str""" | m = hashlib . sha1 ( )
m . update ( config . working_directory . encode ( "utf-8" ) )
for p in config . include_paths :
m . update ( p . encode ( "utf-8" ) )
for p in config . define_symbols :
m . update ( p . encode ( "utf-8" ) )
for p in config . undefine_symbols :
m . update ( p . encode ( "utf-8" ) )
for p in config . cflags :
m . update ( p . encode ( "utf-8" ) )
return m . digest ( ) |
def query_most_pic ( num , kind = '1' ) :
'''Query most pics .''' | return TabPost . select ( ) . where ( ( TabPost . kind == kind ) & ( TabPost . logo != "" ) ) . order_by ( TabPost . view_count . desc ( ) ) . limit ( num ) |
def save_object ( self , obj ) :
"""Save an object with Discipline
Only argument is a Django object . This function saves the object
( regardless of whether it already exists or not ) and registers with
Discipline , creating a new Action object . Do not use obj . save ( ) !""" | obj . save ( )
try :
save_object ( obj , editor = self )
except DisciplineException :
pass |
def get_daily ( self , date = None ) :
"""Get time entries for a date ( defaults to today ) .""" | if date == None :
return self . get ( "/daily.json" )
url = "/daily/{}/{}/{}.json" . format ( date . year , date . month , date . day )
return self . get ( url ) |
def _author_uid_get ( post ) :
"""Get the UID of the post author .
: param Post post : The post object to determine authorship of
: return : Author UID
: rtype : str""" | u = post . meta ( 'author.uid' )
return u if u else str ( current_user . uid ) |
def shorten_text ( self , text ) :
"""Shortens text to fit into the : attr : ` width ` .""" | if len ( text ) > self . width :
return text [ : self . width - 3 ] + '...'
return text |
def insert_asm ( self , addr , asm_code , before_label = False ) :
"""Insert some assembly code at the specific address . There must be an instruction starting at that address .
: param int addr : Address of insertion
: param str asm _ code : The assembly code to insert
: return : None""" | if before_label :
self . _inserted_asm_before_label [ addr ] . append ( asm_code )
else :
self . _inserted_asm_after_label [ addr ] . append ( asm_code ) |
def profile_validation ( self , status ) :
"""Return run total value .""" | self . selected_profile . data . setdefault ( 'validation_pass_count' , 0 )
self . selected_profile . data . setdefault ( 'validation_fail_count' , 0 )
if status :
self . selected_profile . data [ 'validation_pass_count' ] += 1
else :
self . selected_profile . data [ 'validation_fail_count' ] += 1 |
def _from_dict ( cls , _dict ) :
"""Initialize a DocStructure object from a json dictionary .""" | args = { }
if 'section_titles' in _dict :
args [ 'section_titles' ] = [ SectionTitles . _from_dict ( x ) for x in ( _dict . get ( 'section_titles' ) ) ]
if 'leading_sentences' in _dict :
args [ 'leading_sentences' ] = [ LeadingSentence . _from_dict ( x ) for x in ( _dict . get ( 'leading_sentences' ) ) ]
return cls ( ** args ) |
def items ( self , folder_id , subfolder_id , ann_id = None ) :
'''Yields an unodered generator of items in a subfolder .
The generator yields items , which are represented by a tuple
of ` ` content _ id ` ` and ` ` subtopic _ id ` ` . The format of these
identifiers is unspecified .
By default ( with ` ` ann _ id = None ` ` ) , subfolders are shown for all
anonymous users . Optionally , ` ` ann _ id ` ` can be set to a username ,
which restricts the list to only subfolders owned by that user .
: param str folder _ id : Folder id
: param str subfolder _ id : Subfolder id
: param str ann _ id : Username
: rtype : generator of ` ` ( content _ id , subtopic _ id ) ` `''' | self . assert_valid_folder_id ( folder_id )
self . assert_valid_folder_id ( subfolder_id )
ann_id = self . _annotator ( ann_id )
folder_cid = self . wrap_folder_content_id ( ann_id , folder_id )
subfolder_sid = self . wrap_subfolder_subtopic_id ( subfolder_id )
ident = ( folder_cid , subfolder_sid )
if self . store . get ( folder_cid ) is None :
raise KeyError ( folder_id )
for lab in self . label_store . directly_connected ( ident ) :
cid = lab . other ( folder_cid )
subid = lab . subtopic_for ( cid )
yield ( cid , subid ) |
def check ( self , func = None , name = None ) :
"""A decorator to register a new Dockerflow check to be run
when the / _ _ heartbeat _ _ endpoint is called . , e . g . : :
from dockerflow . flask import checks
@ dockerflow . check
def storage _ reachable ( ) :
try :
acme . storage . ping ( )
except SlowConnectionException as exc :
return [ checks . Warning ( exc . msg , id = ' acme . health . 0002 ' ) ]
except StorageException as exc :
return [ checks . Error ( exc . msg , id = ' acme . health . 0001 ' ) ]
or using a custom name : :
@ dockerflow . check ( name = ' acme - storage - check )
def storage _ reachable ( ) :""" | if func is None :
return functools . partial ( self . check , name = name )
if name is None :
name = func . __name__
self . logger . info ( 'Registered Dockerflow check %s' , name )
@ functools . wraps ( func )
def decorated_function ( * args , ** kwargs ) :
self . logger . info ( 'Called Dockerflow check %s' , name )
return func ( * args , ** kwargs )
self . checks [ name ] = decorated_function
return decorated_function |
def _is_qstring ( message ) :
"""Check if its a QString without adding any dep to PyQt5.""" | my_class = str ( message . __class__ )
my_class_name = my_class . replace ( '<class \'' , '' ) . replace ( '\'>' , '' )
if my_class_name == 'PyQt5.QtCore.QString' :
return True
return False |
def ne ( self , other , axis = "columns" , level = None ) :
"""Checks element - wise that this is not equal to other .
Args :
other : A DataFrame or Series or scalar to compare to .
axis : The axis to perform the ne over .
level : The Multilevel index level to apply ne over .
Returns :
A new DataFrame filled with Booleans .""" | return self . _binary_op ( "ne" , other , axis = axis , level = level ) |
def set_autocamera ( self , mode = 'density' ) :
"""- set _ autocamera ( mode = ' density ' ) : By default , Scene defines its
own Camera . However , there is no a general way for doing so . Scene
uses a density criterion for getting the point of view . If this is
not a good option for your problem , you can choose among :
| ' minmax ' | ' density ' | ' median ' | ' mean ' | . If None of the previous methods
work well , you may define the camera params by yourself .""" | self . Camera . set_autocamera ( self . _Particles , mode = mode )
self . _camera_params = self . Camera . get_params ( )
self . _x , self . _y , self . _hsml , self . _kview = self . __compute_scene ( )
self . _m = self . _Particles . _mass [ self . _kview ] |
def cat ( args ) :
"""% prog cat * . pdf - o output . pdf
Concatenate pages from pdf files into a single pdf file .
Page ranges refer to the previously - named file .
A file not followed by a page range means all the pages of the file .
PAGE RANGES are like Python slices .
{ page _ range _ help }
EXAMPLES
pdfcat - o output . pdf head . pdf content . pdf : 6 7 : tail . pdf - 1
Concatenate all of head . pdf , all but page seven of content . pdf ,
and the last page of tail . pdf , producing output . pdf .
pdfcat chapter * . pdf > book . pdf
You can specify the output file by redirection .
pdfcat chapter ? . pdf chapter10 . pdf > book . pdf
In case you don ' t want chapter 10 before chapter 2.""" | p = OptionParser ( cat . __doc__ . format ( page_range_help = PAGE_RANGE_HELP ) )
p . add_option ( "--nosort" , default = False , action = "store_true" , help = "Do not sort file names" )
p . add_option ( "--cleanup" , default = False , action = "store_true" , help = "Remove individual pdfs after merging" )
p . set_outfile ( )
p . set_verbose ( help = "Show page ranges as they are being read" )
opts , args = p . parse_args ( args )
if len ( args ) < 1 :
sys . exit ( not p . print_help ( ) )
outfile = opts . outfile
if outfile in args :
args . remove ( outfile )
if not opts . nosort :
args = natsorted ( args )
filename_page_ranges = parse_filename_page_ranges ( args )
verbose = opts . verbose
fw = must_open ( outfile , "wb" )
merger = PdfFileMerger ( )
in_fs = { }
try :
for ( filename , page_range ) in filename_page_ranges :
if verbose :
print ( filename , page_range , file = sys . stderr )
if filename not in in_fs :
in_fs [ filename ] = open ( filename , "rb" )
merger . append ( in_fs [ filename ] , pages = page_range )
except :
print ( traceback . format_exc ( ) , file = sys . stderr )
print ( "Error while reading " + filename , file = sys . stderr )
sys . exit ( 1 )
merger . write ( fw )
fw . close ( )
if opts . cleanup :
logging . debug ( "Cleaning up {} files" . format ( len ( args ) ) )
for arg in args :
os . remove ( arg ) |
def is_binary ( var , allow_none = False ) :
"""Returns True if var is a binary ( bytes ) objects
Result py - 2 py - 3
b ' bytes literal ' True True
' string literal ' True False
u ' unicode literal ' False False
Also works with the corresponding numpy types .""" | return isinstance ( var , six . binary_type ) or ( var is None and allow_none ) |
def confidence_intervals ( self , X , width = .95 , quantiles = None ) :
"""estimate confidence intervals for the model .
Parameters
X : array - like of shape ( n _ samples , m _ features )
Input data matrix
width : float on [ 0,1 ] , optional
quantiles : array - like of floats in ( 0 , 1 ) , optional
Instead of specifying the prediciton width , one can specify the
quantiles . So ` ` width = . 95 ` ` is equivalent to ` ` quantiles = [ . 025 , . 975 ] ` `
Returns
intervals : np . array of shape ( n _ samples , 2 or len ( quantiles ) )
Notes
Wood 2006 , section 4.9
Confidence intervals based on section 4.8 rely on large sample results to deal with
non - Gaussian distributions , and treat the smoothing parameters as fixed , when in
reality they are estimated from the data .""" | if not self . _is_fitted :
raise AttributeError ( 'GAM has not been fitted. Call fit first.' )
X = check_X ( X , n_feats = self . statistics_ [ 'm_features' ] , edge_knots = self . edge_knots_ , dtypes = self . dtype , features = self . feature , verbose = self . verbose )
return self . _get_quantiles ( X , width , quantiles , prediction = False ) |
def check_common_elements_order ( list1 , list2 ) :
"""Function to verify if the common elements between two given lists maintain the same sequence .
Examples :
check _ common _ elements _ order ( [ ' red ' , ' green ' , ' black ' , ' orange ' ] , [ ' red ' , ' pink ' , ' green ' , ' white ' , ' black ' ] ) - > True
check _ common _ elements _ order ( [ ' red ' , ' pink ' , ' green ' , ' white ' , ' black ' ] , [ ' white ' , ' orange ' , ' pink ' , ' black ' ] ) - > False
check _ common _ elements _ order ( [ ' red ' , ' green ' , ' black ' , ' orange ' ] , [ ' red ' , ' pink ' , ' green ' , ' white ' , ' black ' ] ) - > True
Args :
list1 and list2 : Two lists
Returns :
Boolean value indicating whether common elements between list1 and list2 have the same order in both lists .""" | shared_elements = set ( list1 ) & set ( list2 )
ordered_list1 = [ item for item in list1 if item in shared_elements ]
ordered_list2 = [ item for item in list2 if item in shared_elements ]
return ordered_list1 == ordered_list2 |
def get_quant_NAs ( quantdata , quantheader ) :
"""Takes quantdata in a dict and header with quantkeys
( eg iTRAQ isotopes ) . Returns dict of quant intensities
with missing keys set to NA .""" | out = { }
for qkey in quantheader :
out [ qkey ] = quantdata . get ( qkey , 'NA' )
return out |
def components ( self , extra_params = None ) :
"""All components in this Space""" | return self . api . _get_json ( Component , space = self , rel_path = self . _build_rel_path ( 'ticket_components' ) , extra_params = extra_params , ) |
def infer_modifications ( stmts ) :
"""Return inferred Modification from RegulateActivity + ActiveForm .
This function looks for combinations of Activation / Inhibition Statements
and ActiveForm Statements that imply a Modification Statement .
For example , if we know that A activates B , and phosphorylated B is
active , then we can infer that A leads to the phosphorylation of B .
An additional requirement when making this assumption is that the
activity of B should only be dependent on the modified state and not
other context - otherwise the inferred Modification is not necessarily
warranted .
Parameters
stmts : list [ indra . statements . Statement ]
A list of Statements to infer Modifications from .
Returns
linked _ stmts : list [ indra . mechlinker . LinkedStatement ]
A list of LinkedStatements representing the inferred Statements .""" | linked_stmts = [ ]
for act_stmt in _get_statements_by_type ( stmts , RegulateActivity ) :
for af_stmt in _get_statements_by_type ( stmts , ActiveForm ) :
if not af_stmt . agent . entity_matches ( act_stmt . obj ) :
continue
mods = af_stmt . agent . mods
# Make sure the ActiveForm only involves modified sites
if af_stmt . agent . mutations or af_stmt . agent . bound_conditions or af_stmt . agent . location :
continue
if not af_stmt . agent . mods :
continue
for mod in af_stmt . agent . mods :
evs = act_stmt . evidence + af_stmt . evidence
for ev in evs :
ev . epistemics [ 'direct' ] = False
if mod . is_modified :
mod_type_name = mod . mod_type
else :
mod_type_name = modtype_to_inverse [ mod . mod_type ]
mod_class = modtype_to_modclass [ mod_type_name ]
if not mod_class :
continue
st = mod_class ( act_stmt . subj , act_stmt . obj , mod . residue , mod . position , evidence = evs )
ls = LinkedStatement ( [ act_stmt , af_stmt ] , st )
linked_stmts . append ( ls )
logger . info ( 'inferred: %s' % st )
return linked_stmts |
def AddSpecification ( self , specification ) :
"""Adds a format specification .
Args :
specification ( FormatSpecification ) : format specification .
Raises :
KeyError : if the store already contains a specification with
the same identifier .""" | if specification . identifier in self . _format_specifications :
raise KeyError ( 'Format specification {0:s} is already defined in store.' . format ( specification . identifier ) )
self . _format_specifications [ specification . identifier ] = specification
for signature in specification . signatures :
signature_index = len ( self . _signature_map )
signature_identifier = '{0:s}:{1:d}' . format ( specification . identifier , signature_index )
if signature_identifier in self . _signature_map :
raise KeyError ( 'Signature {0:s} is already defined in map.' . format ( signature_identifier ) )
signature . SetIdentifier ( signature_identifier )
self . _signature_map [ signature_identifier ] = specification |
def parse_string ( self , string ) :
"""Parse ASCII output of JPrintMeta""" | self . log . info ( "Parsing ASCII data" )
if not string :
self . log . warning ( "Empty metadata" )
return
lines = string . splitlines ( )
application_data = [ ]
application = lines [ 0 ] . split ( ) [ 0 ]
self . log . debug ( "Reading meta information for '%s'" % application )
for line in lines :
if application is None :
self . log . debug ( "Reading meta information for '%s'" % application )
application = line . split ( ) [ 0 ]
application_data . append ( line )
if line . startswith ( application + b' Linux' ) :
self . _record_app_data ( application_data )
application_data = [ ]
application = None |
def models ( self ) :
"""Return all the models defined for this module""" | app = get_app ( self . __class__ . __module__ . split ( '.' ) [ - 2 ] )
return get_models ( app ) |
def _append ( self , menu ) :
'''append this menu item to a menu''' | menu . AppendCheckItem ( self . id ( ) , self . name , self . description )
menu . Check ( self . id ( ) , self . checked ) |
def split_by_idxs ( seq , idxs ) :
'''A generator that returns sequence pieces , seperated by indexes specified in idxs .''' | last = 0
for idx in idxs :
if not ( - len ( seq ) <= idx < len ( seq ) ) :
raise KeyError ( f'Idx {idx} is out-of-bounds' )
yield seq [ last : idx ]
last = idx
yield seq [ last : ] |
def cmd_relay ( self , args ) :
'''set relays''' | if len ( args ) == 0 or args [ 0 ] not in [ 'set' , 'repeat' ] :
print ( "Usage: relay <set|repeat>" )
return
if args [ 0 ] == "set" :
if len ( args ) < 3 :
print ( "Usage: relay set <RELAY_NUM> <0|1>" )
return
self . master . mav . command_long_send ( self . target_system , self . target_component , mavutil . mavlink . MAV_CMD_DO_SET_RELAY , 0 , int ( args [ 1 ] ) , int ( args [ 2 ] ) , 0 , 0 , 0 , 0 , 0 )
if args [ 0 ] == "repeat" :
if len ( args ) < 4 :
print ( "Usage: relay repeat <RELAY_NUM> <COUNT> <PERIOD>" )
return
self . master . mav . command_long_send ( self . target_system , self . target_component , mavutil . mavlink . MAV_CMD_DO_REPEAT_RELAY , 0 , int ( args [ 1 ] ) , int ( args [ 2 ] ) , float ( args [ 3 ] ) , 0 , 0 , 0 , 0 ) |
def create_doc_id_from_json ( doc ) -> str :
"""Docs with identical contents get the same ID .
Args :
doc :
Returns : a string with the hash of the given document .""" | return hashlib . sha256 ( json . dumps ( doc , sort_keys = True ) . encode ( 'utf-8' ) ) . hexdigest ( ) |
def asDirect ( self ) :
"""Returns the image data as a direct representation of an
` ` x * y * planes ` ` array . This method is intended to remove the
need for callers to deal with palettes and transparency
themselves . Images with a palette ( colour type 3)
are converted to RGB or RGBA ; images with transparency ( a
` ` tRNS ` ` chunk ) are converted to LA or RGBA as appropriate .
When returned in this format the pixel values represent the
colour value directly without needing to refer to palettes or
transparency information .
Like the : meth : ` read ` method this method returns a 4 - tuple :
( * width * , * height * , * pixels * , * meta * )
This method normally returns pixel values with the bit depth
they have in the source image , but when the source PNG has an
` ` sBIT ` ` chunk it is inspected and can reduce the bit depth of
the result pixels ; pixel values will be reduced according to
the bit depth specified in the ` ` sBIT ` ` chunk ( PNG nerds should
note a single result bit depth is used for all channels ; the
maximum of the ones specified in the ` ` sBIT ` ` chunk . An RGB565
image will be rescaled to 6 - bit RGB666 ) .
The * meta * dictionary that is returned reflects the ` direct `
format and not the original source image . For example , an RGB
source image with a ` ` tRNS ` ` chunk to represent a transparent
colour , will have ` ` planes = 3 ` ` and ` ` alpha = False ` ` for the
source image , but the * meta * dictionary returned by this method
will have ` ` planes = 4 ` ` and ` ` alpha = True ` ` because an alpha
channel is synthesized and added .
* pixels * is the pixel data in boxed row flat pixel format ( just
like the : meth : ` read ` method ) .
All the other aspects of the image data are not changed .""" | self . preamble ( )
# Simple case , no conversion necessary .
if not self . colormap and not self . trns and not self . sbit :
return self . read ( )
x , y , pixels , meta = self . read ( )
if self . colormap :
meta [ 'colormap' ] = False
meta [ 'alpha' ] = bool ( self . trns )
meta [ 'bitdepth' ] = 8
meta [ 'planes' ] = 3 + bool ( self . trns )
plte = self . palette ( )
def iterpal ( pixels ) :
for row in pixels :
row = map ( plte . __getitem__ , row )
yield array ( 'B' , itertools . chain ( * row ) )
pixels = iterpal ( pixels )
elif self . trns : # It would be nice if there was some reasonable way
# of doing this without generating a whole load of
# intermediate tuples . But tuples does seem like the
# easiest way , with no other way clearly much simpler or
# much faster . ( Actually , the L to LA conversion could
# perhaps go faster ( all those 1 - tuples ! ) , but I still
# wonder whether the code proliferation is worth it )
it = self . transparent
maxval = 2 ** meta [ 'bitdepth' ] - 1
planes = meta [ 'planes' ]
meta [ 'alpha' ] = True
meta [ 'planes' ] += 1
typecode = 'BH' [ meta [ 'bitdepth' ] > 8 ]
def itertrns ( pixels ) :
for row in pixels : # For each row we group it into pixels , then form a
# characterisation vector that says whether each
# pixel is opaque or not . Then we convert
# True / False to 0 / maxval ( by multiplication ) ,
# and add it as the extra channel .
row = group ( row , planes )
opa = map ( it . __ne__ , row )
opa = map ( maxval . __mul__ , opa )
opa = zip ( opa )
# convert to 1 - tuples
yield array ( typecode , itertools . chain ( * map ( operator . add , row , opa ) ) )
pixels = itertrns ( pixels )
targetbitdepth = None
if self . sbit :
sbit = struct . unpack ( '%dB' % len ( self . sbit ) , self . sbit )
targetbitdepth = max ( sbit )
if targetbitdepth > meta [ 'bitdepth' ] :
raise Error ( 'sBIT chunk %r exceeds bitdepth %d' % ( sbit , self . bitdepth ) )
if min ( sbit ) <= 0 :
raise Error ( 'sBIT chunk %r has a 0-entry' % sbit )
if targetbitdepth == meta [ 'bitdepth' ] :
targetbitdepth = None
if targetbitdepth :
shift = meta [ 'bitdepth' ] - targetbitdepth
meta [ 'bitdepth' ] = targetbitdepth
def itershift ( pixels ) :
for row in pixels :
yield map ( shift . __rrshift__ , row )
pixels = itershift ( pixels )
return x , y , pixels , meta |
def ordering_step ( self , oneway = False ) :
"""iterator that computes all vertices ordering in their layers
( one layer after the other from top to bottom , to top again unless
oneway is True ) .""" | self . dirv = - 1
crossings = 0
for l in self . layers :
mvmt = l . order ( )
crossings += mvmt
yield ( l , mvmt )
if oneway or ( crossings == 0 ) :
return
self . dirv = + 1
while l :
mvmt = l . order ( )
yield ( l , mvmt )
l = l . nextlayer ( ) |
def _AddMessageMethods ( message_descriptor , cls ) :
"""Adds implementations of all Message methods to cls .""" | _AddListFieldsMethod ( message_descriptor , cls )
_AddHasFieldMethod ( message_descriptor , cls )
_AddClearFieldMethod ( message_descriptor , cls )
if message_descriptor . is_extendable :
_AddClearExtensionMethod ( cls )
_AddHasExtensionMethod ( cls )
_AddEqualsMethod ( message_descriptor , cls )
_AddStrMethod ( message_descriptor , cls )
_AddReprMethod ( message_descriptor , cls )
_AddUnicodeMethod ( message_descriptor , cls )
_AddByteSizeMethod ( message_descriptor , cls )
_AddSerializeToStringMethod ( message_descriptor , cls )
_AddSerializePartialToStringMethod ( message_descriptor , cls )
_AddMergeFromStringMethod ( message_descriptor , cls )
_AddIsInitializedMethod ( message_descriptor , cls )
_AddMergeFromMethod ( cls )
_AddWhichOneofMethod ( message_descriptor , cls )
_AddReduceMethod ( cls )
# Adds methods which do not depend on cls .
cls . Clear = _Clear
cls . DiscardUnknownFields = _DiscardUnknownFields
cls . _SetListener = _SetListener |
def info ( self , section = 'default' ) :
"""Get information and statistics about the server .
If called without argument will return default set of sections .
For available sections , see http : / / redis . io / commands / INFO
: raises ValueError : if section is invalid""" | if not section :
raise ValueError ( "invalid section" )
fut = self . execute ( b'INFO' , section , encoding = 'utf-8' )
return wait_convert ( fut , parse_info ) |
async def is_ready ( self ) :
"""Check if the multi - environment has been fully initialized .
This calls each slave environment managers ' : py : meth : ` is _ ready ` and
checks if the multi - environment itself is ready by calling
: py : meth : ` ~ creamas . mp . MultiEnvironment . check _ ready ` .
. . seealso : :
: py : meth : ` creamas . core . environment . Environment . is _ ready `""" | async def slave_task ( addr , timeout ) :
try :
r_manager = await self . env . connect ( addr , timeout = timeout )
ready = await r_manager . is_ready ( )
if not ready :
return False
except :
return False
return True
if not self . env . is_ready ( ) :
return False
if not self . check_ready ( ) :
return False
rets = await create_tasks ( slave_task , self . addrs , 0.5 )
if not all ( rets ) :
return False
return True |
def _prepare_sort_options ( self , has_pk ) :
"""Prepare sort options for _ values attributes .
If we manager sort by score after getting the result , we do not want to
get values from the first sort call , but only from the last one , after
converting results in zset into keys""" | sort_options = super ( ExtendedCollectionManager , self ) . _prepare_sort_options ( has_pk )
if self . _values : # if we asked for values , we have to use the redis ' sort '
# command , which is able to return other fields .
if not sort_options :
sort_options = { }
sort_options [ 'get' ] = self . _values [ 'fields' ] [ 'keys' ]
if self . _sort_by_sortedset_after :
for key in ( 'get' , 'store' ) :
if key in self . _sort_by_sortedset :
del self . _sort_by_sortedset [ key ]
if sort_options and ( not has_pk or self . _want_score_value ) :
for key in ( 'get' , 'store' ) :
if key in sort_options :
self . _sort_by_sortedset [ key ] = sort_options . pop ( key )
if not sort_options :
sort_options = None
return sort_options |
def calc_2d_ellipse_properties ( cov , nstd = 2 ) :
"""Calculate the properties for 2d ellipse given the covariance matrix .""" | def eigsorted ( cov ) :
vals , vecs = np . linalg . eigh ( cov )
order = vals . argsort ( ) [ : : - 1 ]
return vals [ order ] , vecs [ : , order ]
vals , vecs = eigsorted ( cov )
width , height = 2 * nstd * np . sqrt ( vals [ : 2 ] )
normal = vecs [ : , 2 ] if vecs [ 2 , 2 ] > 0 else - vecs [ : , 2 ]
d = np . cross ( normal , ( 0 , 0 , 1 ) )
M = rotation_matrix ( d )
x_trans = np . dot ( M , ( 1 , 0 , 0 ) )
cos_val = np . dot ( vecs [ : , 0 ] , x_trans ) / np . linalg . norm ( vecs [ : , 0 ] ) / np . linalg . norm ( x_trans )
theta = np . degrees ( np . arccos ( np . clip ( cos_val , - 1 , 1 ) ) )
# if you really want the angle
return { 'width' : width , 'height' : height , 'angle' : theta } , normal |
def to_java_doubles ( m ) :
'''to _ java _ doubles ( m ) yields a java array object for the vector or matrix m .''' | global _java
if _java is None :
_init_registration ( )
m = np . asarray ( m )
dims = len ( m . shape )
if dims > 2 :
raise ValueError ( '1D and 2D arrays supported only' )
bindat = serialize_numpy ( m , 'd' )
return ( _java . jvm . nben . util . Numpy . double2FromBytes ( bindat ) if dims == 2 else _java . jvm . nben . util . Numpy . double1FromBytes ( bindat ) ) |
def build_pdf ( source , texinputs = [ ] , builder = None ) :
"""Builds a LaTeX source to PDF .
Will automatically instantiate an available builder ( or raise a
: class : ` exceptions . RuntimeError ` if none are available ) and build the
supplied source with it .
Parameters are passed on to the builder ' s
: meth : ` ~ latex . build . LatexBuilder . build _ pdf ` function .
: param builder : Specify which builder should be used - ` ` latexmk ` ` ,
` ` pdflatex ` ` or ` ` xelatexmk ` ` .""" | if builder is None :
builders = PREFERRED_BUILDERS
elif builder not in BUILDERS :
raise RuntimeError ( 'Invalid Builder specified' )
else :
builders = ( builder , )
for bld in builders :
bld_cls = BUILDERS [ bld ]
builder = bld_cls ( )
if not builder . is_available ( ) :
continue
return builder . build_pdf ( source , texinputs )
else :
raise RuntimeError ( 'No available builder could be instantiated. ' 'Please make sure LaTeX is installed.' ) |
def stop_containers ( self ) :
"""Stops all containers used by this instance of the backend .""" | while len ( self . _containers ) :
container = self . _containers . pop ( )
try :
container . kill ( signal . SIGKILL )
except docker . errors . APIError : # probably doesn ' t exist anymore
pass |
def ckgpav ( inst , sclkdp , tol , ref ) :
"""Get pointing ( attitude ) and angular velocity
for a specified spacecraft clock time .
http : / / naif . jpl . nasa . gov / pub / naif / toolkit _ docs / C / cspice / ckgpav _ c . html
: param inst : NAIF ID of instrument , spacecraft , or structure .
: type inst : int
: param sclkdp : Encoded spacecraft clock time .
: type sclkdp : float
: param tol : Time tolerance .
: type tol : float
: param ref : Reference frame .
: type ref : str
: return :
C - matrix pointing data ,
Angular velocity vector ,
Output encoded spacecraft clock time .
: rtype : tuple""" | inst = ctypes . c_int ( inst )
sclkdp = ctypes . c_double ( sclkdp )
tol = ctypes . c_double ( tol )
ref = stypes . stringToCharP ( ref )
cmat = stypes . emptyDoubleMatrix ( )
av = stypes . emptyDoubleVector ( 3 )
clkout = ctypes . c_double ( )
found = ctypes . c_int ( )
libspice . ckgpav_c ( inst , sclkdp , tol , ref , cmat , av , ctypes . byref ( clkout ) , ctypes . byref ( found ) )
return stypes . cMatrixToNumpy ( cmat ) , stypes . cVectorToPython ( av ) , clkout . value , bool ( found . value ) |
def gen_locustfile ( testcase_file_path ) :
"""generate locustfile from template .""" | locustfile_path = 'locustfile.py'
template_path = os . path . join ( os . path . dirname ( os . path . realpath ( __file__ ) ) , "templates" , "locustfile_template" )
with io . open ( template_path , encoding = 'utf-8' ) as template :
with io . open ( locustfile_path , 'w' , encoding = 'utf-8' ) as locustfile :
template_content = template . read ( )
template_content = template_content . replace ( "$TESTCASE_FILE" , testcase_file_path )
locustfile . write ( template_content )
return locustfile_path |
def route ( regex , method , name ) :
"""Route the decorated view .
: param regex : A string describing a regular expression to which the request path will be matched .
: param method : A string describing the HTTP method that this view accepts .
: param name : A string describing the name of the URL pattern .
` ` regex ` ` may also be a lambda that accepts the parent resource ' s ` ` prefix ` ` argument and returns
a string describing a regular expression to which the request path will be matched .
` ` name ` ` may also be a lambda that accepts the parent resource ' s ` ` views ` ` argument and returns
a string describing the name of the URL pattern .""" | def decorator ( function ) :
function . route = routes . route ( regex = regex , view = function . __name__ , method = method , name = name )
@ wraps ( function )
def wrapper ( self , * args , ** kwargs ) :
return function ( self , * args , ** kwargs )
return wrapper
return decorator |
def next ( self ) :
"""Pops and returns the first outgoing message from the list .
If message list currently has no messages , the calling thread will
be put to sleep until we have at - least one message in the list that
can be popped and returned .""" | # We pick the first outgoing available and send it .
outgoing_msg = self . outgoing_msg_list . pop_first ( )
# If we do not have any outgoing msg . , we wait .
if outgoing_msg is None :
self . outgoing_msg_event . clear ( )
self . outgoing_msg_event . wait ( )
outgoing_msg = self . outgoing_msg_list . pop_first ( )
return outgoing_msg |
def download ( URLs , dest_dir = '.' , dest_file = None , decompress = False , max_jobs = 5 ) :
'''Download files from specified URL , which should be space , tab or
newline separated URLs . The files will be downloaded to specified
destination . If ` filename . md5 ` files are downloaded , they are used to
validate downloaded ` filename ` . Unless otherwise specified , compressed
files are decompressed . If ` max _ jobs ` is given , a maximum of ` max _ jobs `
concurrent download jobs will be used for each domain . This restriction
applies to domain names and will be applied to multiple download
instances .''' | if env . config [ 'run_mode' ] == 'dryrun' :
print ( f'HINT: download\n{URLs}\n' )
return None
if isinstance ( URLs , str ) :
urls = [ x . strip ( ) for x in URLs . split ( ) if x . strip ( ) ]
else :
urls = list ( URLs )
if not urls :
env . logger . debug ( f'No download URL specified: {URLs}' )
return
if dest_file is not None and len ( urls ) != 1 :
raise RuntimeError ( 'Only one URL is allowed if a destination file is specified.' )
if dest_file is None :
filenames = [ ]
for idx , url in enumerate ( urls ) :
token = urllib . parse . urlparse ( url )
# if no scheme or netloc , the URL is not acceptable
if not all ( [ getattr ( token , qualifying_attr ) for qualifying_attr in ( 'scheme' , 'netloc' ) ] ) :
raise ValueError ( f'Invalid URL {url}' )
filename = os . path . split ( token . path ) [ - 1 ]
if not filename :
raise ValueError ( f'Cannot determine destination file for {url}' )
filenames . append ( os . path . join ( dest_dir , filename ) )
else :
token = urllib . parse . urlparse ( urls [ 0 ] )
if not all ( [ getattr ( token , qualifying_attr ) for qualifying_attr in ( 'scheme' , 'netloc' ) ] ) :
raise ValueError ( f'Invalid URL {url}' )
filenames = [ dest_file ]
succ = [ ( False , None ) for x in urls ]
with ProcessPoolExecutor ( max_workers = max_jobs ) as executor :
for idx , ( url , filename ) in enumerate ( zip ( urls , filenames ) ) : # if there is alot , start download
succ [ idx ] = executor . submit ( downloadURL , url , filename , decompress , idx )
succ = [ x . result ( ) for x in succ ]
# for su , url in zip ( succ , urls ) :
# if not su :
# env . logger . warning ( ' Failed to download { } ' . format ( url ) )
failed = [ y for x , y in zip ( succ , urls ) if not x ]
if failed :
if len ( urls ) == 1 :
raise RuntimeError ( 'Failed to download {urls[0]}' )
else :
raise RuntimeError ( f'Failed to download {failed[0]} ({len(failed)} out of {len(urls)})' )
return 0 |
def main ( github_token , github_api_url , progress ) :
"""A CLI to easily manage GitHub releases , assets and references .""" | global progress_reporter_cls
progress_reporter_cls . reportProgress = sys . stdout . isatty ( ) and progress
if progress_reporter_cls . reportProgress :
progress_reporter_cls = _progress_bar
global _github_token_cli_arg
_github_token_cli_arg = github_token
global _github_api_url
_github_api_url = github_api_url |
def get_max_muO2 ( self , min_voltage = None , max_voltage = None ) :
"""Maximum critical oxygen chemical potential along path .
Args :
min _ voltage : The minimum allowable voltage .
max _ voltage : The maximum allowable voltage .
Returns :
Maximum critical oxygen chemical of all compounds along the
insertion path ( a subset of the path can be chosen by the optional
arguments ) .""" | data = [ ]
for pair in self . _select_in_voltage_range ( min_voltage , max_voltage ) :
if pair . muO2_discharge is not None :
data . extend ( [ d [ 'chempot' ] for d in pair . muO2_discharge ] )
if pair . muO2_charge is not None :
data . extend ( [ d [ 'chempot' ] for d in pair . muO2_discharge ] )
return max ( data ) if len ( data ) > 0 else None |
def get_residuals ( ds , m ) :
"""Using the dataset and model object , calculate the residuals and return
Parameters
ds : dataset object
m : model object
Return
residuals : array of residuals , spec minus model spec""" | model_spectra = get_model_spectra ( ds , m )
resid = ds . test_flux - model_spectra
return resid |
def next_builder ( self ) :
"""Create a new builder based off of this one with its sequence number
incremented .
: return : A new Builder instance
: rtype : : class : ` Builder `""" | sequence = self . sequence + 1
next_builder = Builder ( horizon_uri = self . horizon . horizon_uri , address = self . address , network = self . network , sequence = sequence , fee = self . fee )
next_builder . keypair = self . keypair
return next_builder |
def get_default_config ( self ) :
"""Returns the default collector settings""" | config = super ( NetstatCollector , self ) . get_default_config ( )
config . update ( { 'path' : 'netstat' , } )
return config |
def render_word ( self , text , tag , i ) :
"""Render individual word .
text ( unicode ) : Word text .
tag ( unicode ) : Part - of - speech tag .
i ( int ) : Unique ID , typically word index .
RETURNS ( unicode ) : Rendered SVG markup .""" | y = self . offset_y + self . word_spacing
x = self . offset_x + i * self . distance
if self . direction == "rtl" :
x = self . width - x
html_text = escape_html ( text )
return TPL_DEP_WORDS . format ( text = html_text , tag = tag , x = x , y = y ) |
def _get_singlekws ( skw_matches , spires = False ) :
"""Get single keywords .
: var skw _ matches : dict of { keyword : [ info , . . . ] }
: keyword spires : bool , to get the spires output
: return : list of formatted keywords""" | output = { }
for single_keyword , info in skw_matches :
output [ single_keyword . output ( spires ) ] = len ( info [ 0 ] )
output = [ { 'keyword' : key , 'number' : value } for key , value in output . iteritems ( ) ]
return sorted ( output , key = lambda x : x [ 'number' ] , reverse = True ) |
def offset ( self , offset ) :
"""Move all the intervals in the list by the given ` ` offset ` ` .
: param offset : the shift to be applied
: type offset : : class : ` ~ aeneas . exacttiming . TimeValue `
: raises TypeError : if ` ` offset ` ` is not an instance of ` ` TimeValue ` `""" | self . log ( u"Applying offset to all fragments..." )
self . log ( [ u" Offset %.3f" , offset ] )
for fragment in self . fragments :
fragment . interval . offset ( offset = offset , allow_negative = False , min_begin_value = self . begin , max_end_value = self . end )
self . log ( u"Applying offset to all fragments... done" ) |
def mozjpeg ( ext_args ) :
"""Create argument list for mozjpeg .""" | args = copy . copy ( _MOZJPEG_ARGS )
if Settings . destroy_metadata :
args += [ "-copy" , "none" ]
else :
args += [ "-copy" , "all" ]
args += [ '-outfile' ]
args += [ ext_args . new_filename , ext_args . old_filename ]
extern . run_ext ( args )
return _JPEG_FORMAT |
def _parse_css_color ( color ) :
'''_ parse _ css _ color ( css _ color ) - > gtk . gdk . Color''' | if color . startswith ( "rgb(" ) and color . endswith ( ')' ) :
r , g , b = [ int ( c ) * 257 for c in color [ 4 : - 1 ] . split ( ',' ) ]
return gtk . gdk . Color ( r , g , b )
else :
return gtk . gdk . color_parse ( color ) |
def _makeStoreOwnerPerson ( self ) :
"""Make a L { Person } representing the owner of the store that this
L { Organizer } is installed in .
@ rtype : L { Person }""" | if self . store is None :
return None
userInfo = self . store . findFirst ( signup . UserInfo )
name = u''
if userInfo is not None :
name = userInfo . realName
account = self . store . findUnique ( LoginAccount , LoginAccount . avatars == self . store , None )
ownerPerson = self . createPerson ( name )
if account is not None :
for method in ( self . store . query ( LoginMethod , attributes . AND ( LoginMethod . account == account , LoginMethod . internal == False ) ) ) :
self . createContactItem ( EmailContactType ( self . store ) , ownerPerson , dict ( email = method . localpart + u'@' + method . domain ) )
return ownerPerson |
def update ( self , session , arrays = None , frame = None ) :
'''Creates a frame and writes it to disk .
Args :
arrays : a list of np arrays . Use the " custom " option in the client .
frame : a 2D np array . This way the plugin can be used for video of any
kind , not just the visualization that comes with the plugin .
frame can also be a function , which only is evaluated when the
" frame " option is selected by the client .''' | new_config = self . _get_config ( )
if self . _enough_time_has_passed ( self . previous_config [ 'FPS' ] ) :
self . visualizer . update ( new_config )
self . last_update_time = time . time ( )
final_image = self . _update_frame ( session , arrays , frame , new_config )
self . _update_recording ( final_image , new_config ) |
def _reset ( self , framer ) :
"""Reset the state for the framer . It is safe to call this
method multiple times with the same framer ; the ID of the
framer object will be saved and the state only reset if the
IDs are different . After resetting the state , the framer ' s
` ` init _ state ( ) ` ` method will be called .""" | # Do nothing if we ' re already properly initialized
if id ( framer ) == self . _framer_id :
return
# Reset the state
self . _other = { }
# Initialize the state and save the framer ID
framer . init_state ( self )
self . _framer_id = id ( framer ) |
async def handle_exception ( self , exc : Exception , action : str , request_id ) :
"""Handle any exception that occurs , by sending an appropriate message""" | if isinstance ( exc , APIException ) :
await self . reply ( action = action , errors = self . _format_errors ( exc . detail ) , status = exc . status_code , request_id = request_id )
elif exc == Http404 or isinstance ( exc , Http404 ) :
await self . reply ( action = action , errors = self . _format_errors ( 'Not found' ) , status = 404 , request_id = request_id )
else :
raise exc |
def close_others ( self ) :
"""Closes every editors tabs except the current one .""" | current_widget = self . widget ( self . tab_under_menu ( ) )
if self . _try_close_dirty_tabs ( exept = current_widget ) :
i = 0
while self . count ( ) > 1 :
widget = self . widget ( i )
if widget != current_widget :
self . remove_tab ( i )
else :
i = 1 |
def grid_search ( script : str , params : typing . Iterable [ str ] , dry_run : bool = False ) -> None :
"""Build all grid search parameter configurations and optionally run them .
: param script : String of command prefix , e . g . ` ` cxflow train - v - o log ` ` .
: param params : Iterable collection of strings in standard * * cxflow * * param form , e . g . ` ` ' numerical _ param = [ 1 , 2 ] ' ` `
or ` ` ' text _ param = [ " hello " , " cio " ] ' ` ` .
: param dry _ run : If set to ` ` True ` ` , the built commands will only be printed instead of executed .""" | commands = _build_grid_search_commands ( script = script , params = params )
if dry_run :
logging . warning ( 'Dry run' )
for command in commands :
logging . info ( command )
else :
for command in commands :
try :
completed_process = subprocess . run ( command )
logging . info ( 'Command `%s` completed with exit code %d' , command , completed_process . returncode )
except Exception as _ : # pylint : disable = broad - except
logging . error ( 'Command `%s` failed.' , command ) |
def single_node_env ( num_gpus = 1 ) :
"""Setup environment variables for Hadoop compatibility and GPU allocation""" | import tensorflow as tf
# ensure expanded CLASSPATH w / o glob characters ( required for Spark 2.1 + JNI )
if 'HADOOP_PREFIX' in os . environ and 'TFOS_CLASSPATH_UPDATED' not in os . environ :
classpath = os . environ [ 'CLASSPATH' ]
hadoop_path = os . path . join ( os . environ [ 'HADOOP_PREFIX' ] , 'bin' , 'hadoop' )
hadoop_classpath = subprocess . check_output ( [ hadoop_path , 'classpath' , '--glob' ] ) . decode ( )
os . environ [ 'CLASSPATH' ] = classpath + os . pathsep + hadoop_classpath
os . environ [ 'TFOS_CLASSPATH_UPDATED' ] = '1'
# reserve GPU , if requested
if tf . test . is_built_with_cuda ( ) :
gpus_to_use = gpu_info . get_gpus ( num_gpus )
logging . info ( "Using gpu(s): {0}" . format ( gpus_to_use ) )
os . environ [ 'CUDA_VISIBLE_DEVICES' ] = gpus_to_use
else : # CPU
logging . info ( "Using CPU" )
os . environ [ 'CUDA_VISIBLE_DEVICES' ] = '' |
def close ( self , reply_code = 0 , reply_text = '' , class_id = 0 , method_id = 0 ) :
'''Close this channel . Routes to channel . close .''' | # In the off chance that we call this twice . A good example is if
# there ' s an error in close listeners and so we ' re still inside a
# single call to process _ frames , which will try to close this channel
# if there ' s an exception .
if hasattr ( self , 'channel' ) :
self . channel . close ( reply_code , reply_text , class_id , method_id ) |
def tocimxml ( self ) :
"""Return the CIM - XML representation of this CIM property ,
as an object of an appropriate subclass of : term : ` Element ` .
The returned CIM - XML representation is a ` PROPERTY ` ,
` PROPERTY . REFERENCE ` , or ` PROPERTY . ARRAY ` element dependent on the
property type , and consistent with : term : ` DSP0201 ` . Note that
array properties cannot be of reference type .
The order of qualifiers in the returned CIM - XML representation is
preserved from the : class : ` ~ pywbem . CIMProperty ` object .
Returns :
The CIM - XML representation , as an object of an appropriate subclass
of : term : ` Element ` .""" | qualifiers = [ q . tocimxml ( ) for q in self . qualifiers . values ( ) ]
if self . is_array : # pylint : disable = no - else - return
assert self . type != 'reference'
if self . value is None :
value_xml = None
else :
array_xml = [ ]
for v in self . value :
if v is None :
if SEND_VALUE_NULL :
array_xml . append ( cim_xml . VALUE_NULL ( ) )
else :
array_xml . append ( cim_xml . VALUE ( None ) )
elif self . embedded_object is not None :
assert isinstance ( v , ( CIMInstance , CIMClass ) )
array_xml . append ( cim_xml . VALUE ( v . tocimxml ( ) . toxml ( ) ) )
else :
array_xml . append ( cim_xml . VALUE ( atomic_to_cim_xml ( v ) ) )
value_xml = cim_xml . VALUE_ARRAY ( array_xml )
return cim_xml . PROPERTY_ARRAY ( self . name , self . type , value_xml , self . array_size , self . class_origin , self . propagated , embedded_object = self . embedded_object , qualifiers = qualifiers )
elif self . type == 'reference' : # scalar
if self . value is None :
value_xml = None
else :
value_xml = cim_xml . VALUE_REFERENCE ( self . value . tocimxml ( ) )
return cim_xml . PROPERTY_REFERENCE ( self . name , value_xml , reference_class = self . reference_class , class_origin = self . class_origin , propagated = self . propagated , qualifiers = qualifiers )
else : # scalar non - reference
if self . value is None :
value_xml = None
else :
if self . embedded_object is not None :
assert isinstance ( self . value , ( CIMInstance , CIMClass ) )
value_xml = cim_xml . VALUE ( self . value . tocimxml ( ) . toxml ( ) )
else :
value_xml = cim_xml . VALUE ( atomic_to_cim_xml ( self . value ) )
return cim_xml . PROPERTY ( self . name , self . type , value_xml , class_origin = self . class_origin , propagated = self . propagated , embedded_object = self . embedded_object , qualifiers = qualifiers ) |
def ecg_systole ( ecg , rpeaks , t_waves_ends ) :
"""Returns the localization of systoles and diastoles .
Parameters
ecg : list or ndarray
ECG signal ( preferably filtered ) .
rpeaks : list or ndarray
R peaks localization .
t _ waves _ ends : list or ndarray
T waves localization .
Returns
systole : ndarray
Array indicating where systole ( 1 ) and diastole ( 0 ) .
Example
> > > import neurokit as nk
> > > systole = nk . ecg _ systole ( ecg , rpeaks , t _ waves _ ends )
Notes
* Authors *
- ` Dominique Makowski < https : / / dominiquemakowski . github . io / > ` _
* Details *
- * * Systole / Diastole * * : One prominent channel of body and brain communication is that conveyed by baroreceptors , pressure and stretch - sensitive receptors within the heart and surrounding arteries . Within each cardiac cycle , bursts of baroreceptor afferent activity encoding the strength and timing of each heartbeat are carried via the vagus and glossopharyngeal nerve afferents to the nucleus of the solitary tract . This is the principal route that communicates to the brain the dynamic state of the heart , enabling the representation of cardiovascular arousal within viscerosensory brain regions , and influence ascending neuromodulator systems implicated in emotional and motivational behaviour . Because arterial baroreceptors are activated by the arterial pulse pressure wave , their phasic discharge is maximal during and immediately after the cardiac systole , that is , when the blood is ejected from the heart , and minimal during cardiac diastole , that is , between heartbeats ( Azevedo , 2017 ) .
References
- Azevedo , R . T . , Garfinkel , S . N . , Critchley , H . D . , & Tsakiris , M . ( 2017 ) . Cardiac afferent activity modulates the expression of racial stereotypes . Nature communications , 8.
- Edwards , L . , Ring , C . , McIntyre , D . , & Carroll , D . ( 2001 ) . Modulation of the human nociceptive flexion reflex across the cardiac cycle . Psychophysiology , 38(4 ) , 712-718.
- Gray , M . A . , Rylander , K . , Harrison , N . A . , Wallin , B . G . , & Critchley , H . D . ( 2009 ) . Following one ' s heart : cardiac rhythms gate central initiation of sympathetic reflexes . Journal of Neuroscience , 29(6 ) , 1817-1825.""" | waves = np . array ( [ "" ] * len ( ecg ) )
waves [ rpeaks ] = "R"
waves [ t_waves_ends ] = "T"
systole = [ 0 ]
current = 0
for index , value in enumerate ( waves [ 1 : ] ) :
if waves [ index - 1 ] == "R" :
current = 1
if waves [ index - 1 ] == "T" :
current = 0
systole . append ( current )
return ( systole ) |
def omero_bin ( self , command ) :
"""Runs the omero command - line client with an array of arguments using the
old environment""" | assert isinstance ( command , list )
if not self . old_env :
raise Exception ( 'Old environment not initialised' )
log . info ( "Running [old environment]: %s" , " " . join ( command ) )
self . run ( 'omero' , command , capturestd = True , env = self . old_env ) |
def has_entities ( status ) :
"""Returns true if a Status object has entities .
Args :
status : either a tweepy . Status object or a dict returned from Twitter API""" | try :
if sum ( len ( v ) for v in status . entities . values ( ) ) > 0 :
return True
except AttributeError :
if sum ( len ( v ) for v in status [ 'entities' ] . values ( ) ) > 0 :
return True
return False |
def get_version ( dev_version = False ) :
"""Generates a version string .
Arguments :
dev _ version : Generate a verbose development version from git commits .
Examples :
1.1
1.1 . dev43 # If ' dev _ version ' was passed .""" | if dev_version :
version = git_dev_version ( )
if not version :
raise RuntimeError ( "Could not generate dev version from git." )
return version
return "1!%d.%d" % ( MAJOR , MINOR ) |
def play ( self ) :
"""Send signal to resume playback at the paused offset""" | self . _response [ 'shouldEndSession' ] = True
self . _response [ 'action' ] [ 'audio' ] [ 'interface' ] = 'play'
self . _response [ 'action' ] [ 'audio' ] [ 'sources' ] = [ ]
return self |
def limit ( self , maximum ) :
"""Return a new query , limited to a certain number of results .
Unlike core reporting queries , you cannot specify a starting
point for live queries , just the maximum results returned .
` ` ` python
# first 50
query . limit ( 50)""" | self . meta [ 'limit' ] = maximum
self . raw . update ( { 'max_results' : maximum , } )
return self |
def add_triple ( self , subj : Union [ URIRef , str ] , pred : Union [ URIRef , str ] , obj : Union [ URIRef , Literal , str ] ) -> None :
"""Adds triple to rdflib Graph
Triple can be of any subject , predicate , and object of the entity without a need for order .
Args :
subj : Entity subject
pred : Entity predicate
obj : Entity object
Example :
In [ 1 ] : add _ triple (
. . . : ' http : / / uri . interlex . org / base / ilx _ 0101431 ' ,
. . . : RDF . type ,
. . . : ' http : / / www . w3 . org / 2002/07 / owl # Class ' )""" | if obj in [ None , "" , " " ] :
return
# Empty objects are bad practice
_subj = self . process_subj_or_pred ( subj )
_pred = self . process_subj_or_pred ( pred )
_obj = self . process_obj ( obj )
self . g . add ( ( _subj , _pred , _obj ) ) |
def xmoe2_v1_l4k_compressed_c4 ( ) :
"""With compressed attention .""" | hparams = xmoe2_v1_l4k ( )
hparams . decoder_layers = [ "compressed_att" if l == "att" else l for l in hparams . decoder_layers ]
hparams . compression_factor = 4
return hparams |
def get_broks ( self , broker_name ) :
"""Send a HTTP request to the satellite ( GET / _ broks )
Get broks from the satellite .
Un - serialize data received .
: param broker _ name : the concerned broker link
: type broker _ name : BrokerLink
: return : Broks list on success , [ ] on failure
: rtype : list""" | res = self . con . get ( '_broks' , { 'broker_name' : broker_name } , wait = False )
logger . debug ( "Got broks from %s: %s" , self . name , res )
return unserialize ( res , True ) |
async def close_wallet_search ( wallet_search_handle : int ) -> None :
"""Close wallet search ( make search handle invalid )
: param wallet _ search _ handle : wallet wallet handle ( created by open _ wallet _ search )
: return : None""" | logger = logging . getLogger ( __name__ )
logger . debug ( "close_wallet_search: >>> wallet_search_handle: %r" , wallet_search_handle )
if not hasattr ( close_wallet_search , "cb" ) :
logger . debug ( "close_wallet_search: Creating callback" )
close_wallet_search . cb = create_cb ( CFUNCTYPE ( None , c_int32 , c_int32 ) )
c_wallet_search_handle = c_int32 ( wallet_search_handle )
res = await do_call ( 'indy_close_wallet_search' , c_wallet_search_handle , close_wallet_search . cb )
logger . debug ( "close_wallet_search: <<< res: %r" , res )
return res |
def most_similar ( self , word , number = 5 ) :
"""Run a similarity query , retrieving number
most similar words .""" | if self . word_vectors is None :
raise Exception ( 'Model must be fit before querying' )
if self . dictionary is None :
raise Exception ( 'No word dictionary supplied' )
try :
word_idx = self . dictionary [ word ]
except KeyError :
raise Exception ( 'Word not in dictionary' )
return self . _similarity_query ( self . word_vectors [ word_idx ] , number ) [ 1 : ] |
def audiorate ( filename ) :
"""Determines the samplerate of the given audio recording file
: param filename : filename of the audiofile
: type filename : str
: returns : int - - samplerate of the recording""" | if '.wav' in filename . lower ( ) :
wf = wave . open ( filename )
fs = wf . getframerate ( )
wf . close ( )
elif '.call' in filename . lower ( ) :
fs = 333333
else :
raise IOError ( "Unsupported audio format for file: {}" . format ( filename ) )
return fs |
def _createBitpattern ( functioncode , value ) :
"""Create the bit pattern that is used for writing single bits .
This is basically a storage of numerical constants .
Args :
* functioncode ( int ) : can be 5 or 15
* value ( int ) : can be 0 or 1
Returns :
The bit pattern ( string ) .
Raises :
TypeError , ValueError""" | _checkFunctioncode ( functioncode , [ 5 , 15 ] )
_checkInt ( value , minvalue = 0 , maxvalue = 1 , description = 'inputvalue' )
if functioncode == 5 :
if value == 0 :
return '\x00\x00'
else :
return '\xff\x00'
elif functioncode == 15 :
if value == 0 :
return '\x00'
else :
return '\x01' |
def flush ( self , hard = False ) :
"""Drop existing entries from the cache .
Args :
hard ( bool ) : If True , all current entries are flushed from the
server ( s ) , which affects all users . If False , only the local
process is affected .""" | if not self . servers :
return
if hard :
self . client . flush_all ( )
self . reset_stats ( )
else :
from uuid import uuid4
tag = uuid4 ( ) . hex
if self . debug :
tag = "flushed" + tag
self . current = tag |
def split_arg_to_name_type_value ( self , args_list ) :
"""Split argument text to name , type , value .""" | for arg in args_list :
arg_type = None
arg_value = None
has_type = False
has_value = False
pos_colon = arg . find ( ':' )
pos_equal = arg . find ( '=' )
if pos_equal > - 1 :
has_value = True
if pos_colon > - 1 :
if not has_value :
has_type = True
elif pos_equal > pos_colon : # exception for def foo ( arg1 = " : " )
has_type = True
if has_value and has_type :
arg_name = arg [ 0 : pos_colon ] . strip ( )
arg_type = arg [ pos_colon + 1 : pos_equal ] . strip ( )
arg_value = arg [ pos_equal + 1 : ] . strip ( )
elif not has_value and has_type :
arg_name = arg [ 0 : pos_colon ] . strip ( )
arg_type = arg [ pos_colon + 1 : ] . strip ( )
elif has_value and not has_type :
arg_name = arg [ 0 : pos_equal ] . strip ( )
arg_value = arg [ pos_equal + 1 : ] . strip ( )
else :
arg_name = arg . strip ( )
self . arg_name_list . append ( arg_name )
self . arg_type_list . append ( arg_type )
self . arg_value_list . append ( arg_value ) |
def n_orifices_per_row ( self ) :
"""Calculate number of orifices at each level given an orifice
diameter .""" | # H is distance from the bottom of the next row of orifices to the
# center of the current row of orifices
H = self . b_rows - 0.5 * self . orifice_diameter
flow_per_orifice = pc . flow_orifice_vert ( self . orifice_diameter , H , con . VC_ORIFICE_RATIO )
n = np . zeros ( self . n_rows )
for i in range ( self . n_rows ) : # calculate the ideal number of orifices at the current row without
# constraining to an integer
flow_needed = self . flow_ramp [ i ] - self . flow_actual ( i , n )
n_orifices_real = ( flow_needed / flow_per_orifice ) . to ( u . dimensionless )
# constrain number of orifices to be less than the max per row and
# greater or equal to 0
n [ i ] = min ( ( max ( 0 , round ( n_orifices_real ) ) ) , self . n_orifices_per_row_max )
return n |
def parts ( self ) :
"""Return an array of batch parts to submit""" | parts = [ ]
upserts = dict ( )
deletes = [ ]
# we keep track of the batch size as we go ( pretty close approximation ! ) so we can chunk it small enough
# to limit the HTTP posts to under 700KB - server limits to 750KB , so play it safe
max_upload_size = 700000
# loop upserts first - fit the deletes in afterward
# ' { " replace _ all " : true , " complete " : false , " guid " : " 6659fbfc - 3f08-42ee - 998c - 9109f650f4b7 " , " upserts " : [ ] , " deletes " : [ ] } '
base_part_size = 118
if not self . replace_all :
base_part_size += 1
# yeah , this is totally overkill : )
part_size = base_part_size
for value in self . upserts :
if ( part_size + self . upserts_size [ value ] ) >= max_upload_size : # this record would put us over the limit - close out the batch part and start a new one
parts . append ( BatchPart ( self . replace_all , upserts , deletes ) )
upserts = dict ( )
deletes = [ ]
part_size = base_part_size
# for the new upserts dict , drop the lower - casing of value
upserts [ self . lower_val_to_val [ value ] ] = self . upserts [ value ]
part_size += self . upserts_size [ value ]
# updating the approximate size of the batch
for value in self . deletes : # delete adds length of string plus quotes , comma and space
if ( part_size + len ( value ) + 4 ) >= max_upload_size :
parts . append ( BatchPart ( self . replace_all , upserts , deletes ) )
upserts = dict ( )
deletes = [ ]
part_size = base_part_size
# for the new deletes set , drop the lower - casing of value
deletes . append ( { 'value' : self . lower_val_to_val [ value ] } )
part_size += len ( value ) + 4
if len ( upserts ) + len ( deletes ) > 0 : # finish the batch
parts . append ( BatchPart ( self . replace_all , upserts , deletes ) )
if len ( parts ) == 0 :
if not self . replace_all :
raise ValueError ( "Batch has no data, and 'replace_all' is False" )
parts . append ( BatchPart ( self . replace_all , dict ( ) , [ ] ) )
# last part finishes the batch
parts [ - 1 ] . set_last_part ( )
return parts |
def _keepVol ( self , vol ) :
"""Mark this volume to be kept in path .""" | if vol is None :
return
if vol in self . extraVolumes :
del self . extraVolumes [ vol ]
return
if vol not in self . paths :
raise Exception ( "%s not in %s" % ( vol , self ) )
paths = [ os . path . basename ( path ) for path in self . paths [ vol ] ]
newPath = self . selectReceivePath ( paths )
if self . _skipDryRun ( logger , 'INFO' ) ( "Copy %s to %s" , vol , newPath ) :
return
self . butterVolumes [ vol . uuid ] . copy ( newPath ) |
def get_sls_opts ( opts , ** kwargs ) :
'''Return a copy of the opts for use , optionally load a local config on top''' | opts = copy . deepcopy ( opts )
if 'localconfig' in kwargs :
return salt . config . minion_config ( kwargs [ 'localconfig' ] , defaults = opts )
if 'saltenv' in kwargs :
saltenv = kwargs [ 'saltenv' ]
if saltenv is not None :
if not isinstance ( saltenv , six . string_types ) :
saltenv = six . text_type ( saltenv )
if opts [ 'lock_saltenv' ] and saltenv != opts [ 'saltenv' ] :
raise CommandExecutionError ( 'lock_saltenv is enabled, saltenv cannot be changed' )
opts [ 'saltenv' ] = kwargs [ 'saltenv' ]
pillarenv = None
if kwargs . get ( 'pillarenv' ) :
pillarenv = kwargs . get ( 'pillarenv' )
if opts . get ( 'pillarenv_from_saltenv' ) and not pillarenv :
pillarenv = kwargs . get ( 'saltenv' )
if pillarenv is not None and not isinstance ( pillarenv , six . string_types ) :
opts [ 'pillarenv' ] = six . text_type ( pillarenv )
else :
opts [ 'pillarenv' ] = pillarenv
return opts |
def _create_latent_variables ( self ) :
"""Creates model latent variables
Returns
None ( changes model attributes )""" | for parm in range ( self . z_no ) :
self . latent_variables . add_z ( 'Scale ' + self . X_names [ parm ] , fam . Flat ( transform = 'exp' ) , fam . Normal ( 0 , 3 ) )
self . latent_variables . z_list [ parm ] . start = - 5.0
self . z_no = len ( self . latent_variables . z_list ) |
def _get_by_id ( collection , id ) :
'''Get item from a list by the id field''' | matches = [ item for item in collection if item . id == id ]
if not matches :
raise ValueError ( 'Could not find a matching item' )
elif len ( matches ) > 1 :
raise ValueError ( 'The id matched {0} items, not 1' . format ( len ( matches ) ) )
return matches [ 0 ] |
def getlang_by_native_name ( native_name ) :
"""Try to lookup a Language object by native _ name , e . g . ' English ' , in internal language list .
Returns None if lookup by language name fails in resources / languagelookup . json .""" | direct_match = _iget ( native_name , _LANGUAGE_NATIVE_NAME_LOOKUP )
if direct_match :
return direct_match
else :
simple_native_name = native_name . split ( ',' ) [ 0 ]
# take part before comma
simple_native_name = simple_native_name . split ( '(' ) [ 0 ] . strip ( )
# and before any bracket
return _LANGUAGE_NATIVE_NAME_LOOKUP . get ( simple_native_name , None ) |
End of preview. Expand
in Dataset Viewer.
README.md exists but content is empty.
Use the Edit dataset card button to edit it.
- Downloads last month
- 205