code stringlengths 4 4.48k | docstring stringlengths 1 6.45k | _id stringlengths 24 24 |
|---|---|---|
def test_games_malformed_turn(self): <NEW_LINE> <INDENT> player1 = LegitPlayer() <NEW_LINE> player1.end_of_game = mock.MagicMock() <NEW_LINE> player2 = LegitPlayer() <NEW_LINE> player2.end_of_game = mock.MagicMock() <NEW_LINE> player2.play_turn = mock.MagicMock(return_value="lolol") <NEW_LINE> player1guard = PlayerGuar... | tests that a player who gives a malformed turn
loses and end_of_games is not called on it | 625941b05166f23b2e1a4ea1 |
def get_neighbor(self, theta, new_places, new_states): <NEW_LINE> <INDENT> assert isinstance(theta, np.ndarray) <NEW_LINE> assert np.all([x < self.states for x in theta]), "Invalid theta" <NEW_LINE> new_theta = np.copy(theta) <NEW_LINE> new_theta.put(new_places, new_states) <NEW_LINE> return new_theta | Function to get a neighbor from a theta
Parameters
---------
theta : np.ndarray
A numpy array of integers describing the state
new_places : np.ndarray
An array of locations for which to swap values
new_states : np.ndarray
An array of new values to swap out
Output:
--------
Returns a neighboring state | 625941b0adb09d7d5db6c4de |
def test_abs_path(self): <NEW_LINE> <INDENT> fname = self.id().split(".")[3] <NEW_LINE> backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup') <NEW_LINE> node = self.make_simple_node(base_dir="{0}/{1}/node".format(module_name, fname)) <NEW_LINE> try: <NEW_LINE> <INDENT> self.run_pb(["init", "-B", os.pat... | failure with backup catalog should be given as absolute path | 625941b0097d151d1a222ba5 |
def test_reschedule_true(self): <NEW_LINE> <INDENT> instance_uuid = self.instance['uuid'] <NEW_LINE> method_args = (None, None, None, None, False, {}) <NEW_LINE> self.mox.StubOutWithMock(self.compute, '_deallocate_network') <NEW_LINE> self.mox.StubOutWithMock(self.compute, '_reschedule') <NEW_LINE> try: <NEW_LINE> <IND... | Test behavior when re-scheduling happens | 625941b08a43f66fc4b53db3 |
def set_randomize(self, randomize): <NEW_LINE> <INDENT> assert isinstance(randomize, bool), LOGGER.error("randomize must be boolean") <NEW_LINE> self.__randomize = randomize | Set whether to randomize moves selection.
:Parameters:
#. randomize (boolean): Whether to pull moves randomly from path
or pull moves in order at every step. | 625941b03317a56b869399ae |
def forward(self, prediction_out, target): <NEW_LINE> <INDENT> target = torch.where(target==1, torch.ones_like(target), -1*torch.ones_like(target)).view(-1) <NEW_LINE> target = target.float() <NEW_LINE> A = 1.0 - target*prediction_out.view(-1) <NEW_LINE> max_squared = torch.nn.Softplus()(A) if self.use_softplus else to... | prediction_out (B,1): f(x) | 625941b0d8ef3951e3243287 |
def traiteLigne(ligne): <NEW_LINE> <INDENT> ligne = ligne.decode("Latin1") <NEW_LINE> newLine = u"" <NEW_LINE> c, m = 0, 0 <NEW_LINE> while c < len(ligne): <NEW_LINE> <INDENT> if ligne[c] == " ": <NEW_LINE> <INDENT> newLine = newLine + ligne[m:c] + "-*-" <NEW_LINE> m = c + 1 <NEW_LINE> <DEDENT> c = c + 1 <NEW_LINE> <DE... | convertit une ligne de 'Latin1' en 'Utf8', avec insertion de -*- | 625941b0d7e4931a7ee9dc65 |
def test_team_builder_config_product_size_materials_id_delete(self): <NEW_LINE> <INDENT> pass | Test case for team_builder_config_product_size_materials_id_delete
Delete a model instance by {{id}} from the data source. | 625941b0adb09d7d5db6c4e0 |
def _get_files(self, m_string_path): <NEW_LINE> <INDENT> for _, _, list_files in os.walk(m_string_path): <NEW_LINE> <INDENT> break <NEW_LINE> <DEDENT> return set(list_files) | this method returns a set of files in the directory of csv files
Requirements:
package os
Inputs:
m_string_path
Type: string
Desc: the path of the directory with bulk upload files
Important Info:
None
Return:
object
Type: set
Desc: files in the directory | 625941b0293b9510aa2c2fde |
def get_page_info(url, page_num): <NEW_LINE> <INDENT> base_url = 'http://www.indeed.com' <NEW_LINE> page_job_descriptions = [] <NEW_LINE> start_num = str(page_num * 10) <NEW_LINE> page_url = ''.join([url, '&start=', start_num]) <NEW_LINE> print('Getting page: ' + page_url) <NEW_LINE> html_page = requests.get(page_url).... | This function takes a URL and a page number as arguments,
combines them into a new URL for search, and returns a two-dimensional list,
where each value of the list is a one-dimensional list that contains a set of words
appearing in one job Ad of this page.
url: string, a base URL before the page number
page_num: int,... | 625941b0596a89723608980e |
def random_walk_normal_fn(scale=1., name=None): <NEW_LINE> <INDENT> def _fn(state_parts, seed): <NEW_LINE> <INDENT> with tf.name_scope(name, 'random_walk_normal_fn', values=[state_parts, scale, seed]): <NEW_LINE> <INDENT> scales = scale if mcmc_util.is_list_like(scale) else [scale] <NEW_LINE> if len(scales) == 1: <NEW_... | Returns a callable that adds a random normal perturbation to the input.
This function returns a callable that accepts a Python `list` of `Tensor`s of
any shapes and `dtypes` representing the state parts of the `current_state`
and a random seed. The supplied argument `scale` must be a `Tensor` or Python
`list` of `Ten... | 625941b00a50d4780f666bda |
def num_chars(s): <NEW_LINE> <INDENT> return len(str(s)) | function to return number of characters | 625941b02c8b7c6e89b35510 |
def dispatch(self): <NEW_LINE> <INDENT> if self.controller.options.serial: <NEW_LINE> <INDENT> self.serial_dispatch() <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> self.parallel_dispatch() | Run the doctests for the controller's specified sources,
by calling :meth:`parallel_dispatch` or :meth:`serial_dispatch`
according to the ``--serial`` option.
EXAMPLES::
sage: from sage.doctest.control import DocTestController, DocTestDefaults
sage: from sage.doctest.forker import DocTestDispatcher
sage: ... | 625941b0d7e4931a7ee9dc67 |
def _validate_network_subnetpools(self, network, subnet_ip_version, new_subnetpool, network_scope): <NEW_LINE> <INDENT> ipv6_pd_subnetpool = new_subnetpool == const.IPV6_PD_POOL_ID <NEW_LINE> if network_scope: <NEW_LINE> <INDENT> if (ipv6_pd_subnetpool or new_subnetpool and new_subnetpool.address_scope_id != network_sc... | Validate all subnets on the given network have been allocated from
the same subnet pool as new_subnetpool if no address scope is
used. If address scopes are used, validate that all subnets on the
given network participate in the same address scope. | 625941b0293b9510aa2c2fe0 |
def items(self): <NEW_LINE> <INDENT> return self._list | Returns all items in this list as a python list. | 625941b0d8ef3951e324328b |
def store_attr(names=None, self=None, but='', cast=False, **attrs): <NEW_LINE> <INDENT> fr = sys._getframe(1) <NEW_LINE> args = fr.f_code.co_varnames[:fr.f_code.co_argcount] <NEW_LINE> if self: args = ('self', *args) <NEW_LINE> else: self = fr.f_locals[args[0]] <NEW_LINE> if not hasattr(self, '__stored_args__'): self._... | Store params named in comma-separated `names` from calling context into attrs in `self` | 625941b082261d6c526ab1eb |
def generate_seed(): <NEW_LINE> <INDENT> return random.randint(0, 100000000000) | Generate a random seed to be used in worldgen
:return: random integer | 625941b03346ee7daa2b2ab5 |
def resolve_unique_keys(self, info, **args): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> return json.dumps(self.unique_keys) <NEW_LINE> <DEDENT> except AttributeError as exception: <NEW_LINE> <INDENT> logger.exception(exception) <NEW_LINE> return [] | Return a string dump of the unique keys. This is a string because
we don't have a polymorphic GraphQL representation of why
might be defined as a unique key and it is therefore easier to
just return it as a string and have the client parse it.
..note:
The AttributeError must be handled here now because at one
... | 625941b03c8af77a43ae34ee |
def annotations(self): <NEW_LINE> <INDENT> return map(self.create_annotation_node, self.document.annotations) | Returns a list of annotation nodes | 625941b07b180e01f3dc4558 |
def main(req: func.HttpRequest) -> func.HttpResponse: <NEW_LINE> <INDENT> logging.info("Python HTTP trigger function processed a request.") <NEW_LINE> name = req.params.get("name") <NEW_LINE> if not name: <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> req_body = req.get_json() <NEW_LINE> <DEDENT> except ValueError: <NEW_... | Says hello to user. | 625941b05e10d32532c5ec79 |
def _Plus(lhs=None, rhs=None, out=None, name=None, **kwargs): <NEW_LINE> <INDENT> return (0,) | Adds arguments element-wise.
The storage type of ``elemwise_add`` output depends on storage types of inputs
- elemwise_add(row_sparse, row_sparse) = row_sparse
- otherwise, ``elemwise_add`` generates output with default storage
Parameters
----------
lhs : NDArray
first input
rhs : NDArray
second inpu... | 625941b0046cf37aa974ca9a |
def _add_keys(resource: Dict[str, Any], path: Tuple[str, ...], items: Dict[str, Any]): <NEW_LINE> <INDENT> if not any(i is not None for i in items.values()): <NEW_LINE> <INDENT> return resource <NEW_LINE> <DEDENT> obj = resource <NEW_LINE> for p in path: <NEW_LINE> <INDENT> if p not in obj: <NEW_LINE> <INDENT> obj[p] =... | A helper function to manipulate k8s resource configs. It will update a dict in resource
definition given a path (think jsonpath). Will ignore items set to None. Will not update
resource if all items are None. Note that it updates resource dict in-place.
Examples:
>>> _add_keys({}, ("foo", "bar"), {"A": 1, ... | 625941b02c8b7c6e89b35514 |
def read_processed(fileloc, flux_file, ivar_file): <NEW_LINE> <INDENT> with open(fileloc + flux_file,"rb") as file: <NEW_LINE> <INDENT> f = pickle.load(file) <NEW_LINE> <DEDENT> with open(fileloc + ivar_file,"rb") as file: <NEW_LINE> <INDENT> i = pickle.load(file) <NEW_LINE> <DEDENT> return f, i | Reads processed fluxes, ivars from disk
Returns fluxes and ivars 2D arrays
Args:
fileloc : location of the processed files
flux_file : name of the processed fluxes pkl file
ivar_file: name of the processed ivars pkl file | 625941b0d7e4931a7ee9dc6b |
def _url_request(self, url, request_parameters, request_type='GET', files=None, repeat=0, error_text="Error", raise_on_failure=True): <NEW_LINE> <INDENT> if files is not None: <NEW_LINE> <INDENT> mpf = _MultiPartForm(param_dict=request_parameters, files=files) <NEW_LINE> req = request(url) <NEW_LINE> body = mpf.make_re... | Send a new request and format the json response.
Keyword arguments:
url - the url of the request
request_parameters - a dictionay containing the name of the parameter and its correspoinsding value
request_type - the type of request: 'GET', 'POST'
files - the files to be uploaded
repeat - the nuber of times to repeat th... | 625941b0293b9510aa2c2fe3 |
def _has_intersection(self, point, query): <NEW_LINE> <INDENT> has_intersection = False <NEW_LINE> if point == query[1]: <NEW_LINE> <INDENT> has_intersection = False <NEW_LINE> <DEDENT> elif point >= query[0] and point < query[1]: <NEW_LINE> <INDENT> has_intersection = True <NEW_LINE> <DEDENT> return has_intersection | Verify if a point has intersection with a query | 625941b0d164cc6175782a97 |
def create_channel(self, config): <NEW_LINE> <INDENT> with self.lock: <NEW_LINE> <INDENT> self._create_channel(config) <NEW_LINE> <DEDENT> logger.info('Added channel `%s` to AMQP connector `%s`', config.name, self.config.name) | Creates a channel.
| 625941b03c8af77a43ae34f0 |
def TestFunction(data_in=None): <NEW_LINE> <INDENT> pass | Define functions here. | 625941b0046cf37aa974ca9c |
def get_toolbar(self, language, fieldname, checksum): <NEW_LINE> <INDENT> groups = [] <NEW_LINE> extra_params = COPY_TEMPLATE.format( ugettext(u'Loading…'), reverse('js-get', kwargs={'checksum': checksum}), checksum, ) <NEW_LINE> groups.append( GROUP_TEMPLATE.format( '', BUTTON_TEMPLATE.format( 'copy-text', ugettext('F... | Returns toolbar HTML code. | 625941b0d8ef3951e324328f |
def write(self): <NEW_LINE> <INDENT> dirs = [self.recipes_dir] <NEW_LINE> if self.expectations: <NEW_LINE> <INDENT> dirs.append(self.expect_dir) <NEW_LINE> <DEDENT> for d in dirs: <NEW_LINE> <INDENT> if not os.path.exists(d): <NEW_LINE> <INDENT> os.makedirs(d) <NEW_LINE> <DEDENT> <DEDENT> with open(self.recipe_path, 'w... | Writes the recipe to disk. | 625941b0d7e4931a7ee9dc6d |
def DescribeFleetUtilization(self, request): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> params = request._serialize() <NEW_LINE> body = self.call("DescribeFleetUtilization", params) <NEW_LINE> response = json.loads(body) <NEW_LINE> if "Error" not in response["Response"]: <NEW_LINE> <INDENT> model = models.DescribeFle... | 本接口(DescribeFleetUtilization)用于查询服务器舰队的利用率信息。
:param request: Request instance for DescribeFleetUtilization.
:type request: :class:`tencentcloud.gse.v20191112.models.DescribeFleetUtilizationRequest`
:rtype: :class:`tencentcloud.gse.v20191112.models.DescribeFleetUtilizationResponse` | 625941b0d164cc6175782a98 |
def align_segments(s1, s2, data1, data2): <NEW_LINE> <INDENT> lengths_s1 = length_segment(s1, data1) <NEW_LINE> lengths_s2 = length_segment(s2, data2) <NEW_LINE> ratios_s1 = [l/lengths_s1[-1] for l in lengths_s1] <NEW_LINE> ratios_s2 = [l/lengths_s2[-1] for l in lengths_s2] <NEW_LINE> len_s1 = lengths_s1[-1] <NEW_LINE>... | Compute the alignment of segments s1 and s2,
such that the first and last elements of s1 and s2 are the same, but nothing else.
:return_type: list of (list of QPointF, int)
:returns: List of wall parts such that the first point is the vertex.
The integer is the id of the point (if it corresponds to one). | 625941b08a43f66fc4b53dbd |
def crypt_private(passwd, passwd_hash, hash_prefix='$P$'): <NEW_LINE> <INDENT> itoa64 = './0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz' <NEW_LINE> output = '*0' <NEW_LINE> if passwd_hash[0:2] == output: <NEW_LINE> <INDENT> output = '*1' <NEW_LINE> <DEDENT> if passwd_hash[0:3] != hash_prefix: <NEW_LIN... | Hash password, using same salt and number of
iterations as in passwd_hash.
This is useful when you want to check password match.
In this case you pass your raw password and password
hash to this function and then compare its return
value with password hash again:
is_valid = (crypt_private(passwd, hash) == hash... | 625941b0d164cc6175782a99 |
def testIsGrantha_FalseCases(self): <NEW_LINE> <INDENT> print("Testing TamilLetter.is_grantha() with aytham", end=' ') <NEW_LINE> letter = AYTHAM <NEW_LINE> self.assertFalse(TamilLetter.is_grantha(letter), "\'%s\' is Aytham, but is_grantha() returned True" %letter) <NEW_LINE> print(".... pass") <NEW_LINE> print("Testin... | is_grantha() should return False for aytham and for all vowels, combinations and vallinam, mellinam and idaiyinam consonants | 625941b07b180e01f3dc455c |
def CalcFFT(self): <NEW_LINE> <INDENT> if not (self.flags["timeCalculated"]): <NEW_LINE> <INDENT> self.CalcTime() <NEW_LINE> <DEDENT> self.FFTData = np.fft.fft(self.Data[:, :4], axis=0) / (self.Data.shape[0]) <NEW_LINE> self.fftFreqs = np.fft.fftfreq(self.Data.shape[0], self.MeanDeltaT) <NEW_LINE> self.flags["FFTCalcul... | Returns
-------
None. | 625941b076d4e153a657e883 |
def test_database_populate(cli_runner, db): <NEW_LINE> <INDENT> result = cli_runner(cli.database_cli, ['populate']) <NEW_LINE> assert result.exit_code == 0 <NEW_LINE> assert result.output.endswith('Database populate completed.\n') | Test `claimstore database populate` command. | 625941b02c8b7c6e89b35518 |
def __repr__(self): <NEW_LINE> <INDENT> return FS_REPR.format(type(self).__name__, super(SimpleNamespace, self).__repr__()) | Object representation | 625941b082261d6c526ab1f1 |
def _make_estimator(self, append=True): <NEW_LINE> <INDENT> estimator = clone(self.base_estimator_) <NEW_LINE> estimator.set_params(**dict((p, getattr(self, p)) for p in self.estimator_params)) <NEW_LINE> if append: <NEW_LINE> <INDENT> self.estimators_.append(estimator) <NEW_LINE> <DEDENT> return estimator | Make and configure a copy of the `base_estimator_` attribute.
Warning: This method should be used to properly instantiate new
sub-estimators. | 625941b05166f23b2e1a4ead |
def load_dataset(self, ds, level=1): <NEW_LINE> <INDENT> is_parent = level==1 <NEW_LINE> levsep = is_parent and "/--------" or "|__.." <NEW_LINE> treelog.info( "%s%s%s (%s)", level * ' ', levsep, ds.__class__.__name__, (is_parent and "parent" or level)) <NEW_LINE> for ref_ds in ds.meta.references: <NEW_LINE> <INDENT> ... | load this dataset and all its dependent datasets.
level is essentially the order of processing (going from dataset to
dependent datasets). Child datasets are always loaded before the
parent. The level is important for visualizing the chain of
dependencies : 0 is the bottom, and thus should be the first set of
ob... | 625941b08e71fb1e9831d502 |
def get_minibatch(self, sentences, tokenize=False, add_start_end=True): <NEW_LINE> <INDENT> if tokenize: <NEW_LINE> <INDENT> sentences = [nltk.word_tokenize(sentence) for sentence in sentences] <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> sentences = [sentence.split() for sentence in sentences] <NEW_LINE> <DEDENT> if ... | Prepare minibatch. | 625941b06aa9bd52df036af7 |
def __init__(self, size=0, position=(0, 0)): <NEW_LINE> <INDENT> self.__size = size <NEW_LINE> if type(self.__size) != int: <NEW_LINE> <INDENT> raise TypeError('size must be an integer') <NEW_LINE> <DEDENT> if self.__size < 0: <NEW_LINE> <INDENT> raise ValueError('size must be >= 0') <NEW_LINE> <DEDENT> self.__position... | Example of docstring on the __init__ method.
Args:
size (int): size of square.
position (tuple): position of square | 625941b03317a56b869399ba |
def Clone(self): <NEW_LINE> <INDENT> pass | Clone(self: Vector3DAnimation) -> Vector3DAnimation
Creates a modifiable clone of this System.Windows.Media.Animation.Vector3DAnimation,making deep
copies of this object's values. When copying dependency properties,this method copies resource
references and data bindings (but they might no longer resolve) b... | 625941b0046cf37aa974caa0 |
def join_path(*args): <NEW_LINE> <INDENT> return '/'.join(args) | Join an arbitrary number of strings as a Linux/Unix file path
Used in place of ``os.path.join`` so tests can run on Windows or Linux
without generating false positives.
Example::
>>> join_path('some', 'directory')
some/directory
>>> join_path('/home', )
:Returns: String
:params args: **Required** The st... | 625941b0d8ef3951e3243293 |
def parser_xls_repescagem(self, file): <NEW_LINE> <INDENT> wb = xlrd.open_workbook(file) <NEW_LINE> s = wb.sheet_by_name('Plan1') <NEW_LINE> result = {} <NEW_LINE> for row in range(s.nrows): <NEW_LINE> <INDENT> for col in range(s.ncols): <NEW_LINE> <INDENT> cel = str(s.cell(row, col)).replace('text:', '').replace("'", ... | Parseia o arquivo .xls de forma mais custosa, buscando,
em cada célula da planilha, a presença dos nomes dos campos.
Esse método deve ser usado apenas para os arquivos que não puderam
ser parseados pelo método .parser_xls_file | 625941b082261d6c526ab1f3 |
def rpminfo(ui): <NEW_LINE> <INDENT> result = set() <NEW_LINE> rpmbin = ui.config("rage", "rpmbin", "rpm") <NEW_LINE> for name in ["hg", "hg.real"]: <NEW_LINE> <INDENT> path = which(name) <NEW_LINE> if not path: <NEW_LINE> <INDENT> continue <NEW_LINE> <DEDENT> result.add(shcmd("%s -qf %s" % (rpmbin, path), check=False)... | FBONLY: Information about RPM packages | 625941b05166f23b2e1a4eaf |
def _eval(self, indiv): <NEW_LINE> <INDENT> x = indiv[0] <NEW_LINE> return (exp(-2*log(2)*((x-0.08)/0.854)**2))*sin(5*pi*(x**(3.0/4)-0.05))**6 | f(x) = (exp^(...))*sin^6(5*pi*x)
| 625941b08e71fb1e9831d504 |
def testParseSystem(self): <NEW_LINE> <INDENT> knowledge_base_values = {u'current_control_set': u'ControlSet001'} <NEW_LINE> test_file = self._GetTestFilePath([u'SYSTEM']) <NEW_LINE> event_queue_consumer = self._ParseFile( self._parser, test_file, knowledge_base_values=knowledge_base_values) <NEW_LINE> event_objects = ... | Tests the Parse function on a SYSTEM file. | 625941b0bf627c535bc12f25 |
def default_frame(self): <NEW_LINE> <INDENT> return self._def_frame | Return the default frame of on ``self``.
OUTPUT:
- a local frame as an instance of
:class:`~sage.manifolds.local_frame.LocalFrame`
EXAMPLES::
sage: M = Manifold(3, 'M', structure='top')
sage: E = M.vector_bundle(2, 'E')
sage: e = E.local_frame('e')
sage: E.default_frame()
Local frame (E|_M, (e... | 625941b08a43f66fc4b53dc1 |
def vim_global(name, kind=str): <NEW_LINE> <INDENT> ret = GLOBALS_DEFAULTS.get(name, None) <NEW_LINE> try: <NEW_LINE> <INDENT> vname = "autotag" + name <NEW_LINE> v_buffer = "b:" + vname <NEW_LINE> exists_buffer = (vim.eval("exists('%s')" % v_buffer) == "1") <NEW_LINE> v_global = "g:" + vname <NEW_LINE> exists_global =... | Get global variable from vim, cast it appropriately | 625941b0293b9510aa2c2fe9 |
def p_statement_block(p): <NEW_LINE> <INDENT> p[0] = PT_Block(p[1]) | statement : block | 625941b03317a56b869399bc |
def __init__(self, api_url, key): <NEW_LINE> <INDENT> self.api_url = api_url <NEW_LINE> self.headers = {'content-type': CONTENT_TYPE_JSON, 'X-Api-Key': key} <NEW_LINE> self.printer_last_reading = [{}, None] <NEW_LINE> self.job_last_reading = [{}, None] | Initialize OctoPrint API and set headers needed later. | 625941b0d8ef3951e3243295 |
def _collectSubstratesFromResults( results ): <NEW_LINE> <INDENT> substrates = collections.OrderedDict() <NEW_LINE> for res in results: <NEW_LINE> <INDENT> for lbl in res.foundLabels: <NEW_LINE> <INDENT> if lbl.sub is None: <NEW_LINE> <INDENT> fpath = misc.getLabelableFilePath(res.original) <NEW_LINE> <DEDENT> else: <N... | Collect all image substrates of found labels;
if the foundLabel does not have a path, use the path from
the original | 625941b073bcbd0ca4b2bdcd |
def _get_target_V( self, batch: ReplayBufferSample, task_info: TaskInfo ) -> TensorType: <NEW_LINE> <INDENT> mtobs = MTObs(env_obs=batch.next_env_obs, task_obs=None, task_info=task_info) <NEW_LINE> _, policy_action, log_pi, _ = self.actor(mtobs=mtobs) <NEW_LINE> target_Q1, target_Q2 = self.critic_target(mtobs=mtobs, ac... | Compute the target values.
Args:
batch (ReplayBufferSample): batch from the replay buffer.
task_info (TaskInfo): task_info object.
Returns:
TensorType: target values. | 625941b04f6381625f114797 |
def chunks(l, n): <NEW_LINE> <INDENT> return [l[i:i + n] for i in range(0, len(l), n)] | split list l into chunks of size n | 625941b08e71fb1e9831d506 |
def _fill_context(self, context, model): <NEW_LINE> <INDENT> return context | Fills the template generation context with data from the model.
The default implementation does nothing. You can override this if your template requires additional
context data that can be filled from the model.
Args:
context - The template context.
model - The model of the currently rendered row
Returns:
... | 625941b03346ee7daa2b2abf |
def offspring(self, parent1, parent2): <NEW_LINE> <INDENT> child = np.random.rand(Hypers.nchrom) <NEW_LINE> for n in range(Hypers.nchrom): <NEW_LINE> <INDENT> rando = np.random.rand(1) <NEW_LINE> child[n] = rando*parent1[n] + (1-rando)*parent2[n] <NEW_LINE> <DEDENT> return(child) | Return the child of both parents, taking each gene from only one parent at random. | 625941b0d164cc6175782a9f |
def __init__(self, auth, home=None): <NEW_LINE> <INDENT> self.auth = auth <NEW_LINE> self.homedata = None <NEW_LINE> self.home_ids = [] <NEW_LINE> self.home_names = [] <NEW_LINE> self.room_names = [] <NEW_LINE> self.schedules = [] <NEW_LINE> self.home = home <NEW_LINE> self.home_id = None | Initialize the HomeData object. | 625941b06aa9bd52df036af9 |
@click.option('--username', prompt=True, default=lambda: os.environ.get('USER', '')) <NEW_LINE> @account.command(help='Connect to baguette.io.') <NEW_LINE> def login(username): <NEW_LINE> <INDENT> password = click.prompt('Password', hide_input=True) <NEW_LINE> if api.login(username, password): <NEW_LINE> <INDENT> click... | Connect to baguette.io using username/password.
:param username: The username to log in with.
:type username: str
:returns: The status of the login.
:rtype: bool | 625941b03317a56b869399be |
def leg_count(self): <NEW_LINE> <INDENT> return self.legs | Return the number of legs this Arthropod possesses. | 625941b00a50d4780f666be8 |
def get_all_time_entry(self, user, spent_on=None): <NEW_LINE> <INDENT> redmine = Redmine(url=self.url, key=user.authkey) <NEW_LINE> try: <NEW_LINE> <INDENT> time_entries = list() <NEW_LINE> r_user_id = redmine.auth().id <NEW_LINE> r_time_entries = redmine.time_entry.filter(user_id=r_user_id, spent_on=spent_on) <NEW_LIN... | Get all time entry from redmine for user.
:param tracktime.models.User user:
:param spent_on:
:rtype: list | 625941b00a366e3fb873e56e |
def testFixedDisplay1(self): <NEW_LINE> <INDENT> V.ArithmeticClass(Options(dict(arithmetic='fixed', precision=6, display=7))) <NEW_LINE> self.assertEqual(F.display, 6) <NEW_LINE> self.assertTrue(F.info.find('display') < 0) | fixed display must be <= precision | 625941b0d8ef3951e3243297 |
def draw_side1(x, y, window): <NEW_LINE> <INDENT> draw_die_face(x, y, window) <NEW_LINE> draw_die_hole(x, y, window) | - This function draws the side of a die with one dot in the center
- draw_die_face(x, y, window)
- draw_die_hole(x, y, window) | 625941b0bf627c535bc12f29 |
def sortList(self, head): <NEW_LINE> <INDENT> if not head or not head.next: <NEW_LINE> <INDENT> return head <NEW_LINE> <DEDENT> slow = head <NEW_LINE> fast = head <NEW_LINE> while fast.next and fast.next.next: <NEW_LINE> <INDENT> slow = slow.next <NEW_LINE> fast = fast.next.next <NEW_LINE> <DEDENT> A = head <NEW_LINE> ... | :type head: ListNode
:rtype: ListNode | 625941b03c8af77a43ae34fa |
def finalize(self): <NEW_LINE> <INDENT> if self._is_finalized: <NEW_LINE> <INDENT> raise SHMACIssue.already_finalized() <NEW_LINE> <DEDENT> self._is_finalized = True <NEW_LINE> self._set_final_result() <NEW_LINE> self._result_is_ready = True <NEW_LINE> self._mac = DeletedAttribute(SHMACIssue.already_finalized) <NEW_LIN... | Caps off the instance's validation hash object & populates the
instance's final result with an HMAC of its state. This signals
the end of a stream of data that can be validated with the
current instance. | 625941b06aa9bd52df036afb |
def __init__(self): <NEW_LINE> <INDENT> self.throw() | Konstruktor bezparametrowy | 625941b0f9cc0f698b14035b |
def check_for_msn_groups_headers(self, msg, target=None): <NEW_LINE> <INDENT> to = ''.join(msg.get_decoded_header('To')) <NEW_LINE> if not Regex(r"<(\S+)\@groups\.msn\.com>").search(to): <NEW_LINE> <INDENT> return False <NEW_LINE> <DEDENT> listname = Regex(r"<(\S+)\@groups\.msn\.com>").match(to).groups()[0] <NEW_LINE> ... | Check if the email's destination is a msn group | 625941b0b57a9660fec335d3 |
def re_compile_maybe_verbose(regex: str) -> Pattern[str]: <NEW_LINE> <INDENT> if '\n' in regex: <NEW_LINE> <INDENT> regex = '(?x)' + regex <NEW_LINE> <DEDENT> compiled: Pattern[str] = re.compile(regex) <NEW_LINE> return compiled | Compile a regular expression string in `regex`.
If it contains newlines, use verbose mode. | 625941b04a966d76dd550d66 |
def read(self): <NEW_LINE> <INDENT> while not self.buffer: <NEW_LINE> <INDENT> self._read() <NEW_LINE> <DEDENT> data = self.buffer.pop(0) <NEW_LINE> try: <NEW_LINE> <INDENT> json_data = json_converter.loads(data) <NEW_LINE> _logger.debug("Data read %r", json_data) <NEW_LINE> <DEDENT> except ValueError: <NEW_LINE> <INDE... | This method waits until new data is available at the connection
or in the buffer and returns it to the caller. | 625941b0507cdc57c6306a25 |
def get_property(self, name): <NEW_LINE> <INDENT> if name != "mode": <NEW_LINE> <INDENT> raise UnsupportedProperty(name) <NEW_LINE> <DEDENT> sound_mode = self.entity.attributes.get(media_player.ATTR_SOUND_MODE) <NEW_LINE> if sound_mode and sound_mode.upper() in self.VALID_SOUND_MODES: <NEW_LINE> <INDENT> return sound_m... | Read and return a property. | 625941b0b5575c28eb68dd58 |
def get_move(self, game, time_left): <NEW_LINE> <INDENT> self.time_left = time_left <NEW_LINE> self.best_move = (-1, -1) <NEW_LINE> if game.get_legal_moves(): <NEW_LINE> <INDENT> self.best_move = game.get_legal_moves()[0] <NEW_LINE> <DEDENT> try: <NEW_LINE> <INDENT> for depth in range(0, 10000): <NEW_LINE> <INDENT> sel... | Search for the best move from the available legal moves and return a
result before the time limit expires.
Modify the get_move() method from the MinimaxPlayer class to implement
iterative deepening search instead of fixed-depth search.
**********************************************************************
NOTE: If ti... | 625941b0e5267d203edcd9fe |
def approximate(func, from_val, to_val, degree): <NEW_LINE> <INDENT> print('------ Started Calculating Approximation ------') <NEW_LINE> print() <NEW_LINE> print('f(x) = {0}'.format(func)) <NEW_LINE> print() <NEW_LINE> start_time = time.time() <NEW_LINE> orth_basis = orthonormal_basis(_get_float_value(from_val), _get_f... | Approximate given continuous function over real numbers, using a polynomial of given degree, with inner product
defined as <f, g> = INTEGRATE f(x) * g(x) dx from a to b.
:param func: Function to approximate represented in a string, with variable as 'x'.
:param from_val: a as a string.
:param to_val: b as a string.
:par... | 625941b02c8b7c6e89b35522 |
def spin_wigner(rho, theta, phi): <NEW_LINE> <INDENT> if rho.type == 'bra': <NEW_LINE> <INDENT> rho = rho.dag() <NEW_LINE> <DEDENT> if rho.type == 'ket': <NEW_LINE> <INDENT> rho = ket2dm(rho) <NEW_LINE> <DEDENT> J = rho.shape[0] <NEW_LINE> j = (J - 1) / 2 <NEW_LINE> THETA, PHI = meshgrid(theta, phi) <NEW_LINE> W = np.z... | Wigner function for a spin-j system.
The spin W function is normal when integrated over the surface of the
sphere
.. math:: \sqrt{\frac{4 \pi}{2j + 1}}\int_\phi \int_\theta
W(\theta,\phi) \sin(\theta) d\theta d\phi = 1
Parameters
----------
state : qobj
A state vector or density matrix for a spin-j qu... | 625941b04f6381625f11479d |
def get_location_string(self): <NEW_LINE> <INDENT> location_string = ('Rows ' + str(self.r_1) + '-' + str(self.r_2) + ' of ' + str(self.total_rows) + '\nCols ' + str(self.c_1) + '-' + str(self.c_2) + ' of ' + str(self.total_cols)) <NEW_LINE> return location_string | Creates a header showing where the top left df corner is | 625941b08a43f66fc4b53dc9 |
def test_operator_exceptions(self): <NEW_LINE> <INDENT> radix1 = Radix(0, [], [], [], 3) <NEW_LINE> radix2 = Radix(0, [], [], [], 2) <NEW_LINE> with self.assertRaises(BasesError): <NEW_LINE> <INDENT> radix1 > radix2 <NEW_LINE> <DEDENT> with self.assertRaises(BasesError): <NEW_LINE> <INDENT> radix1 < radix2 <NEW_LINE> <... | Test that comparsion operators yield exceptions. | 625941b03346ee7daa2b2ac5 |
def deserialize(self, str): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> if self.ONELEG is None: <NEW_LINE> <INDENT> self.ONELEG = hexapodservice.msg.legjoints() <NEW_LINE> <DEDENT> if self.ALLLEGS is None: <NEW_LINE> <INDENT> self.ALLLEGS = hexapodservice.msg.legs() <NEW_LINE> <DEDENT> end = 0 <NEW_LINE> _x = self <NE... | unpack serialized message in str into this message instance
:param str: byte array of serialized message, ``str`` | 625941b05e10d32532c5ec89 |
def computeActionFromValues(self, state): <NEW_LINE> <INDENT> value = util.Counter() <NEW_LINE> actions = self.mdp.getPossibleActions(state) <NEW_LINE> for action in actions: <NEW_LINE> <INDENT> value[action] = self.getQValue(state, action) <NEW_LINE> <DEDENT> return value.argMax() <NEW_LINE> util.raiseNotDefined() | The policy is the best action in the given state
according to the values currently stored in self.values.
You may break ties any way you see fit. Note that if
there are no legal actions, which is the case at the
terminal state, you should return None. | 625941b0e5267d203edcda00 |
def get_budget(self): <NEW_LINE> <INDENT> return self.budget | Returns base budget for Department | 625941b0099cdd3c635f09bd |
def intersect(self, nums1, nums2): <NEW_LINE> <INDENT> num_result = [] <NEW_LINE> for i in range(len(nums1)): <NEW_LINE> <INDENT> if nums1[i] in nums2: <NEW_LINE> <INDENT> num_result.append(nums1[i]) <NEW_LINE> nums2.remove(nums1[i]) <NEW_LINE> <DEDENT> <DEDENT> return num_result | :type nums1: List[int]
:type nums2: List[int]
:rtype: List[int] | 625941b0046cf37aa974caaa |
def partition(self, head, x): <NEW_LINE> <INDENT> if head is None: <NEW_LINE> <INDENT> return None <NEW_LINE> <DEDENT> leftDummy = ListNode(0) <NEW_LINE> left = leftDummy <NEW_LINE> rightDummy = ListNode(0) <NEW_LINE> right = rightDummy <NEW_LINE> node = head <NEW_LINE> while node: <NEW_LINE> <INDENT> if node.val < x: ... | :type head: ListNode
:type x: int
:rtype: ListNode | 625941b0d7e4931a7ee9dc7b |
def getMappingTagToDirectory(tag): <NEW_LINE> <INDENT> tag = tag.replace("VO.glast.org-","") <NEW_LINE> items = tag.split("/") <NEW_LINE> if len(items)>3: <NEW_LINE> <INDENT> return S_ERROR("Bad tag structure") <NEW_LINE> <DEDENT> soft_os = items[0] <NEW_LINE> variant = items[1] <NEW_LINE> package = items[2] <NEW_LINE>... | Returns the directory given a tag name
| 625941b0d8ef3951e324329d |
def edit_link_tag(self,pk,text): <NEW_LINE> <INDENT> query_str = self.request.GET.urlencode() <NEW_LINE> print("query_str",query_str) <NEW_LINE> params = QueryDict(mutable=True) <NEW_LINE> params[self.config._query_param_key] = query_str <NEW_LINE> print("params", params) <NEW_LINE> print("params_urlencode",params.urle... | 此函数的作用是返回一个可以点击的标签
:param pk:
:param text:
:return: 返回一个标签, | 625941b04f6381625f11479f |
def recording_setting_update(self, meeting_id, **kwargs): <NEW_LINE> <INDENT> kwargs['_return_http_data_only'] = True <NEW_LINE> if kwargs.get('async_req'): <NEW_LINE> <INDENT> return self.recording_setting_update_with_http_info(meeting_id, **kwargs) <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> (data) = self.recording... | Get Meeting Recording Settings # noqa: E501
Retrieve settings applied to a meeting's [Cloud Recording](https://support.zoom.us/hc/en-us/articles/203741855-Cloud-Recording).<br><br> **Scopes**: `recording:read:admin` `recording:read`<br> **[Rate Limit Label](https://marketplace.zoom.us/docs/api-reference/rate-limits... | 625941b0b57a9660fec335d7 |
def config_setup( env: LibraryEnvironment, site_list, arbitrator_list, instance_name=None, overwrite_existing=False, ): <NEW_LINE> <INDENT> instance_name = instance_name or constants.DEFAULT_INSTANCE_NAME <NEW_LINE> report_processor = env.report_processor <NEW_LINE> report_processor.report_list( config_validators.check... | create booth configuration
env
list site_list -- site adresses of multisite
list arbitrator_list -- arbitrator adresses of multisite
string instance_name -- booth instance name
bool overwrite_existing -- allow overwriting existing files | 625941b03c8af77a43ae3500 |
def XTRA_echo2(socket_handler, params): <NEW_LINE> <INDENT> MANAGER.app_log.warning("Extra command test2") <NEW_LINE> print(params) <NEW_LINE> data = "Echo2, you said {}".format((',').join(params)) <NEW_LINE> MANAGER.execute_filter_hook('send_data_back', {'socket': socket_handler, 'data': data}, first_only=True) | This command takes inline param(s): XTRA_echo2 hello 2 | 625941b026238365f5f0ebca |
def to_json(self, value): <NEW_LINE> <INDENT> return to_json(value) | Converts value to JSON.
This can be overridden if necessary. | 625941b0293b9510aa2c2ff4 |
@app.route('/signin', methods=['GET', 'POST']) <NEW_LINE> def signin(): <NEW_LINE> <INDENT> username = request.values.getlist('username')[0] <NEW_LINE> password = request.values.getlist('password')[0] <NEW_LINE> nickname = request.values.getlist('nickname')[0] <NEW_LINE> gender = request.values.getlist('gender')[0] <NE... | 若用户名不冲突 存库并返回注册成功信息
若冲突 返回错误信息 | 625941b0627d3e7fe0d68ba8 |
def t_run_av_scores(): <NEW_LINE> <INDENT> alg_params = solver.AlgoParams(generations=500, pop_size=30, num_children=20, mutation_rate=0.1, crossover_rate=0.2) <NEW_LINE> algs = [algo_ea] <NEW_LINE> gens = [gen_schwefel] <NEW_LINE> for gen in gens: <NEW_LINE> <INDENT> print() <NEW_LINE> print() <NEW_LINE> print(gen.__n... | Call compare_algs() [prints average fitness results for range of algorithms] for
multiple generators | 625941b04f88993c3716bdcf |
def create_record(hostname, port): <NEW_LINE> <INDENT> p = {} <NEW_LINE> for f in min_rqd_fields: <NEW_LINE> <INDENT> p[f] = getattr(port, f, None) <NEW_LINE> <DEDENT> p['errors'] = [] <NEW_LINE> p['errors_cat'] = create_error_categories() <NEW_LINE> try: <NEW_LINE> <INDENT> p["lldpnbr"] = str((p["lldpnbr"].nbr.model.h... | Make sure each object has the minimum required fields
to make it easier for reporting later | 625941b0099cdd3c635f09bf |
def Simulate(self): <NEW_LINE> <INDENT> self.setDynamicFunction() <NEW_LINE> t = np.arange(0,self.finalTime,self.separation) <NEW_LINE> return odeint(self.DynamicFunction,self.initCondition,t,()) | simulation routine using the Odeint solver from the scipy optimize package , Odeint is a python implementation
of the ODEPACK package in FORTRAN which uses a multi-step solver in the non stiff case | 625941b0d7e4931a7ee9dc7d |
def get_layers(name): <NEW_LINE> <INDENT> assert name == "cornet_irl" <NEW_LINE> return ["V1.output", "V2.output", "V4.output", "IT.output"] | This method returns a list of string layer names to consider per model. The benchmarks maps brain regions to
layers and uses this list as a set of possible layers. The lists doesn't have to contain all layers, the less the
faster the benchmark process works. Additionally the given layers have to produce an activations ... | 625941b0507cdc57c6306a2b |
def nlm_coeff(k, nmax, lmax): <NEW_LINE> <INDENT> i = 0 <NEW_LINE> for n in range(nmax+1): <NEW_LINE> <INDENT> for l in range(lmax+1): <NEW_LINE> <INDENT> for m in range(l+1): <NEW_LINE> <INDENT> if (i==k): <NEW_LINE> <INDENT> return(n,l,m) <NEW_LINE> break <NEW_LINE> <DEDENT> i+=1 <NEW_LINE> <DEDENT> <DEDENT> <DEDENT>... | Returns the coefficients n, l, m for a given number of
the.....
e.g k=1 is n=0, l=1, m=0
Input:
------
k: Number of the coefficient. e.g
nmax: maximum n in the expansion.
lmax: maximum l in the expansion.
Output:
-------
(n, l, m) | 625941b0d164cc6175782aa8 |
def update(self, obj_index, option, value, set_index=None): <NEW_LINE> <INDENT> option = Options.get(option) <NEW_LINE> if option.symmetric: <NEW_LINE> <INDENT> self._update_symmetric_option(obj_index, option, value) <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> self._update_asymmetric_option(obj_index, set_index, opti... | Updates an object option value. | 625941b0b57a9660fec335d9 |
def get_linked_workflow(workflow, steps, settings): <NEW_LINE> <INDENT> workflow_nodes = {} <NEW_LINE> for step_title, step in workflow[CWL_STEPS].items(): <NEW_LINE> <INDENT> step_name = get_step_name_from_title(step_title, workflow) <NEW_LINE> workflow_nodes[step_name] = { 'inputs': {}, 'outputs': {}, 'ancestors': {}... | Determines static workflow using CWL definitions. This will attempt to get
all steps defined in the 'workflow' and construct a dict of nodes to be
displayed.
:param workflow: (dict) A CWL workflow dictionary.
:param steps: (dict) A dict of CWL step dictionaries.
:param settings: (dict) A CWL arguments dictionary.
:... | 625941b07b180e01f3dc456c |
def test_module_instructions_print(self): <NEW_LINE> <INDENT> module_tuples = getmembers(commands, ismodule) <NEW_LINE> with mock.patch("buf.commands.help.print") as mock_print: <NEW_LINE> <INDENT> for module_name, module in module_tuples: <NEW_LINE> <INDENT> test_options_dict = {"<subcommand_name>": module_name} <NEW_... | Tests that a module's 'instructions' docstring is printed when the modules name is called with help. | 625941b076d4e153a657e893 |
def decode_b64_file(src, dst): <NEW_LINE> <INDENT> with open(src, 'r') as source_file: <NEW_LINE> <INDENT> with open(dst, 'w') as destination_file: <NEW_LINE> <INDENT> for line in source_file: <NEW_LINE> <INDENT> destination_file.write(base64.b64decode(line)) | Read src base64 encoded file and output its content to dst file
| 625941b08a349b6b435e7ed9 |
def control_samples(controls: Sequence[dict], sampleslist: Sequence[str]): <NEW_LINE> <INDENT> def __process_control__(trait_data): <NEW_LINE> <INDENT> def __process_sample__(acc, sample): <NEW_LINE> <INDENT> if sample in trait_data["data"].keys(): <NEW_LINE> <INDENT> sample_item = trait_data["data"][sample] <NEW_LINE>... | Fetches data for the control traits.
This migrates `web/webqtl/correlation/correlationFunction.controlStrain` in
GN1, with a few modifications to the arguments passed in.
PARAMETERS:
controls: A map of sample names to trait data. Equivalent to the `cvals`
value in the corresponding source function in GN1.
samples... | 625941b06fece00bbac2d49c |
def reset(self): <NEW_LINE> <INDENT> self.d0 = np.random.randint(10, 30) <NEW_LINE> self.d = self.d0 <NEW_LINE> self.init_v = np.random.random() * SPEED_LIMIT <NEW_LINE> self.v_head = self.init_v <NEW_LINE> self.v = self.v_head <NEW_LINE> self.step_number = 0 <NEW_LINE> return self.v_cal_raw | 初始化环境, 返回未经攻击的前车速度的观测值
:return
v_head: 前车速度
v: 自车速度 | 625941b0293b9510aa2c2ff7 |
def decode(self, v): <NEW_LINE> <INDENT> immediate = None <NEW_LINE> if getattr(self.parent, "imm7_align4", False): <NEW_LINE> <INDENT> v = self.parent.imm7_align4.value & 0x1F <NEW_LINE> immediate = v << 2 <NEW_LINE> <DEDENT> elif getattr(self.parent, "imm7", False): <NEW_LINE> <INDENT> immediate = self.parent.imm7.va... | Modify the decoded value using the previously decoded
immediate. | 625941b03346ee7daa2b2acb |
def logconfig(level, file=None): <NEW_LINE> <INDENT> logger = logging.getLogger() <NEW_LINE> logging.addLevelName(100, "CRITICAL") <NEW_LINE> logging.addLevelName(60, "ERROR") <NEW_LINE> logging.addLevelName(50, "WARNING") <NEW_LINE> logging.addLevelName(40, "DEBUG") <NEW_LINE> logging.addLevelName(30, "DEBUG2") <NEW_L... | Initialize the python logger for Current. | 625941b0627d3e7fe0d68bac |
def calculate_file_url(name_bytes, base_url): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> rel_name = base64.b64decode(name_bytes).decode('utf8') <NEW_LINE> <DEDENT> except binascii.Error: <NEW_LINE> <INDENT> return None <NEW_LINE> <DEDENT> full_url = urllib.parse.urljoin(base_url, rel_name) <NEW_LINE> log.debug('rel_n... | Calculate and return a full URL for file name.
File name is interpreted as base64-encoded bytes and joined to base URL. | 625941b0046cf37aa974cab0 |
def locRec_dict(dataDF, locColName): <NEW_LINE> <INDENT> temp_dict = {} <NEW_LINE> locList = dataDF[locColName].unique() <NEW_LINE> rowIndex = dataDF.index.values <NEW_LINE> for loc in locList: <NEW_LINE> <INDENT> recs = [] <NEW_LINE> for row in rowIndex: <NEW_LINE> <INDENT> rowloc = dataDF.loc[row, "Location"] <NEW_LI... | function to turn recommendation dataframe into dictionary by location;
:param dataDF: dataframe of recommendations
:param locColName: column name that will be reference of dictionary key
:return: | 625941b03539df3088e2e0b1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.