id
int64
20
338k
vocab_size
int64
2
671
ast_levels
int64
4
32
nloc
int64
1
451
n_ast_nodes
int64
12
5.6k
n_identifiers
int64
1
186
n_ast_errors
int64
0
10
n_words
int64
2
2.17k
n_whitespaces
int64
2
13.8k
fun_name
stringlengths
2
73
commit_message
stringlengths
51
15.3k
url
stringlengths
31
59
code
stringlengths
51
31k
ast_errors
stringlengths
0
1.46k
token_counts
int64
6
3.32k
file_name
stringlengths
5
56
language
stringclasses
1 value
path
stringlengths
7
134
commit_id
stringlengths
40
40
repo
stringlengths
3
28
complexity
int64
1
153
159,095
25
9
6
105
14
1
27
47
test_cli_log_level_debug_used
Configurable logging for libraries (#10614) * Make library level logging to be configurable Fixes https://github.com/RasaHQ/rasa/issues/10203 * Create log level documentation under cheatsheet in Rasa docs * Add log docs to `rasa shell --debug` (and others)
https://github.com/RasaHQ/rasa.git
def test_cli_log_level_debug_used(): configure_logging_and_warnings(logging.DEBUG) rasa_logger = logging.getLogger("rasa") rasa_logger.level == logging.DEBUG matplotlib_logger = logging.getLogger("matplotlib") # Default log level for libraries is currently ERROR matplotlib_logger.level == logging.ERROR @mock.patch.dict(os.environ, {"LOG_LEVEL": "WARNING"})
@mock.patch.dict(os.environ, {"LOG_LEVEL": "WARNING"})
41
test_common.py
Python
tests/utils/test_common.py
f00148b089d326c952880a0e5e6bd4b2dcb98ce5
rasa
1
101,184
21
12
10
126
15
0
24
150
_obtain_mask
lib.detected_face.Mask - Add source + target offset and coverage to set_sub_crop method
https://github.com/deepfakes/faceswap.git
def _obtain_mask(cls, detected_face, mask_type): mask = detected_face.mask.get(mask_type) if not mask: return None if mask.stored_centering != "face": face = AlignedFace(detected_face.landmarks_xy) mask.set_sub_crop(face.pose.offset[mask.stored_centering], face.pose.offset["face"], centering="face") return mask.mask.squeeze()
77
viewport.py
Python
tools/manual/faceviewer/viewport.py
32950897376b48e0f08b46385602e4df902cf49e
faceswap
3
259,191
14
12
6
62
6
0
15
45
_estimator_has
MNT Replace if_delegate_has_method with available_if in ensemble and semi_supervised (#20545) Co-authored-by: Guillaume Lemaitre <[email protected]> Co-authored-by: Jérémie du Boisberranger <[email protected]>
https://github.com/scikit-learn/scikit-learn.git
def _estimator_has(attr): return lambda self: ( hasattr(self.estimators_[0], attr) if hasattr(self, "estimators_") else hasattr(self.base_estimator, attr) )
39
_bagging.py
Python
sklearn/ensemble/_bagging.py
a794c58692a1f3e7a85a42d8c7f7ddd5fcf18baa
scikit-learn
2
286,398
90
14
82
406
36
0
114
880
call_find
More Fixes to Crypto + key sort (#3244) * fix #3095 - autocomplete and command working + key sort * fix #3056 * fix [Bug] bugs #3048 * fix [Bug] bug #3017 * sort -> sortby, not ascend, tests * fix my goof ups Co-authored-by: james <[email protected]>
https://github.com/OpenBB-finance/OpenBBTerminal.git
def call_find(self, other_args): parser = argparse.ArgumentParser( prog="find", add_help=False, formatter_class=argparse.ArgumentDefaultsHelpFormatter, description=, ) parser.add_argument( "-c", "--coin", help="Symbol Name or Id of Coin", dest="coin", required="-h" not in other_args, type=str, ) parser.add_argument( "-k", "--key", dest="key", help="Specify by which column you would like to search: symbol, name, id", type=str, choices=FIND_KEYS, default="symbol", ) parser.add_argument( "-l", "--limit", default=10, dest="limit", help="Number of records to display", type=check_positive, ) parser.add_argument( "-s", "--skip", default=0, dest="skip", help="Skip n of records", type=check_positive, ) if other_args and not other_args[0][0] == "-": other_args.insert(0, "-c") ns_parser = self.parse_known_args_and_warn( parser, other_args, EXPORT_ONLY_RAW_DATA_ALLOWED, ) # TODO: merge find + display_all_coins if ns_parser: if ns_parser.coin == "ALL": display_all_coins( symbol=ns_parser.coin, source=ns_parser.source, limit=ns_parser.limit, skip=ns_parser.skip, show_all=True, export=ns_parser.export, ) else: find( query=ns_parser.coin, source=ns_parser.source, key=ns_parser.key, limit=ns_parser.limit, export=ns_parser.export, )
257
crypto_controller.py
Python
openbb_terminal/cryptocurrency/crypto_controller.py
09f753da1c2a2f03c41fe6a3ca2eb79f6ea58995
OpenBBTerminal
5
190,078
40
13
18
213
18
0
55
153
generate_tex_file
Migrate more `os.path` to `pathlib` (#2980) * Migrate more `os.path` to `pathlib` * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fix type errors with recent pathlib code * pathlib fixes * more pathlib fixes * remove unused imports introduced by pathlib migration * convert `open()` calls to pathlib * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Migrate tex_file_writing to pathlib * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * converted more old code to pathlib, and fixed a bug in module_ops * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fix test failures * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fix test failures * Apply suggestions from code review Co-authored-by: Benjamin Hackl <[email protected]> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com> Co-authored-by: Benjamin Hackl <[email protected]>
https://github.com/ManimCommunity/manim.git
def generate_tex_file(expression, environment=None, tex_template=None): if tex_template is None: tex_template = config["tex_template"] if environment is not None: output = tex_template.get_texcode_for_expression_in_env(expression, environment) else: output = tex_template.get_texcode_for_expression(expression) tex_dir = config.get_dir("tex_dir") if not tex_dir.exists(): tex_dir.mkdir() result = tex_dir / (tex_hash(output) + ".tex") if not result.exists(): logger.info( "Writing %(expression)s to %(path)s", {"expression": expression, "path": f"{result}"}, ) result.write_text(output, encoding="utf-8") return result
122
tex_file_writing.py
Python
manim/utils/tex_file_writing.py
9d1f066d637cb15baea10e6907ab85efff8fb36f
manim
5
246,370
55
12
29
295
23
0
88
330
test_in_flight_requests_stop_being_in_flight
Add more tests for in-flight state query duplication. (#12033)
https://github.com/matrix-org/synapse.git
def test_in_flight_requests_stop_being_in_flight(self) -> None: req1 = ensureDeferred( self.state_datastore._get_state_for_group_using_inflight_cache( 42, StateFilter.all() ) ) self.pump(by=0.1) # This should have gone to the database self.assertEqual(len(self.get_state_group_calls), 1) self.assertFalse(req1.called) # Complete the request right away. self._complete_request_fake(*self.get_state_group_calls[0]) self.assertTrue(req1.called) # Send off another request req2 = ensureDeferred( self.state_datastore._get_state_for_group_using_inflight_cache( 42, StateFilter.all() ) ) self.pump(by=0.1) # It should have gone to the database again, because the previous request # isn't in-flight and therefore isn't available for deduplication. self.assertEqual(len(self.get_state_group_calls), 2) self.assertFalse(req2.called) # Complete the request right away. self._complete_request_fake(*self.get_state_group_calls[1]) self.assertTrue(req2.called) groups, sf, d = self.get_state_group_calls[0] self.assertEqual(self.get_success(req1), FAKE_STATE) self.assertEqual(self.get_success(req2), FAKE_STATE)
186
test_state_store.py
Python
tests/storage/databases/test_state_store.py
546b9c9e648f5e2b25bb7c8350570787ff9befae
synapse
1
266,036
24
10
7
129
12
0
34
97
test_cf_data
Closes #10052: The cf attribute now returns deserialized custom field data
https://github.com/netbox-community/netbox.git
def test_cf_data(self): site = Site(name='Test Site', slug='test-site') # Check custom field data on new instance site.custom_field_data['foo'] = 'abc' self.assertEqual(site.cf['foo'], 'abc') # Check custom field data from database site.save() site = Site.objects.get(name='Test Site') self.assertEqual(site.cf['foo'], 'abc')
69
test_customfields.py
Python
netbox/extras/tests/test_customfields.py
ea6d86e6c4bb6037465410db6205a7471bc81a6c
netbox
1
200,289
37
19
11
176
16
0
54
150
convert_to_native_paths
runtests.py: Undo auto-formatting, re-add changes to blacklist for scipy, numpy
https://github.com/sympy/sympy.git
def convert_to_native_paths(lst): newlst = [] for i, rv in enumerate(lst): rv = os.path.join(*rv.split("/")) # on windows the slash after the colon is dropped if sys.platform == "win32": pos = rv.find(':') if pos != -1: if rv[pos + 1] != '\\': rv = rv[:pos + 1] + '\\' + rv[pos + 1:] newlst.append(os.path.normcase(rv)) return newlst
101
runtests.py
Python
sympy/testing/runtests.py
6d2bbf80752549276a968fd4af78231c569d55c5
sympy
5
60,371
171
19
31
424
32
0
258
556
ProcessFile
Balanced joint maximum mean discrepancy for deep transfer learning
https://github.com/jindongwang/transferlearning.git
def ProcessFile(filename, vlevel, extra_check_functions=[]): _SetVerboseLevel(vlevel) try: # Support the UNIX convention of using "-" for stdin. Note that # we are not opening the file with universal newline support # (which codecs doesn't support anyway), so the resulting lines do # contain trailing '\r' characters if we are reading a file that # has CRLF endings. # If after the split a trailing '\r' is present, it is removed # below. If it is not expected to be present (i.e. os.linesep != # '\r\n' as in Windows), a warning is issued below if this file # is processed. if filename == '-': lines = codecs.StreamReaderWriter(sys.stdin, codecs.getreader('utf8'), codecs.getwriter('utf8'), 'replace').read().split('\n') else: lines = codecs.open(filename, 'r', 'utf8', 'replace').read().split('\n') carriage_return_found = False # Remove trailing '\r'. for linenum in range(len(lines)): if lines[linenum].endswith('\r'): lines[linenum] = lines[linenum].rstrip('\r') carriage_return_found = True except IOError: sys.stderr.write( "Skipping input '%s': Can't open for reading\n" % filename) return # Note, if no dot is found, this will give the entire filename as the ext. file_extension = filename[filename.rfind('.') + 1:] # When reading from stdin, the extension is unknown, so no cpplint tests # should rely on the extension. if filename != '-' and file_extension not in _valid_extensions: sys.stderr.write('Ignoring %s; not a valid file name ' '(%s)\n' % (filename, ', '.join(_valid_extensions))) else: ProcessFileData(filename, file_extension, lines, Error, extra_check_functions) if carriage_return_found and os.linesep != '\r\n': # Use 0 for linenum since outputting only one error for potentially # several lines. Error(filename, 0, 'whitespace/newline', 1, 'One or more unexpected \\r (^M) found;' 'better to use only a \\n') sys.stderr.write('Done processing %s\n' % filename)
230
cpp_lint.py
Python
code/deep/BJMMD/caffe/scripts/cpp_lint.py
cc4d0564756ca067516f71718a3d135996525909
transferlearning
9
36,020
17
6
8
23
5
0
17
38
default_batch_size
Add ONNX export for ViT (#15658) * Add ONNX support for ViT * Refactor to use generic preprocessor * Add vision dep to tests * Extend ONNX slow tests to ViT * Add dummy image generator * Use model_type to determine modality * Add deprecation warnings for tokenizer argument * Add warning when overwriting the preprocessor * Add optional args to docstrings * Add minimum PyTorch version to OnnxConfig * Refactor OnnxConfig class variables from CONSTANT_NAME to snake_case * Add reasonable value for default atol Co-authored-by: Sylvain Gugger <[email protected]>
https://github.com/huggingface/transformers.git
def default_batch_size(self) -> int: # Using 2 avoid ONNX making assumption about single sample batch return OnnxConfig.default_fixed_batch
12
config.py
Python
src/transformers/onnx/config.py
50dd314d939a86f3a81e19af01459f449fbaeeca
transformers
1
272,037
10
11
4
63
10
0
11
39
_tracking_metadata
Reformatting the codebase with black. PiperOrigin-RevId: 450093126
https://github.com/keras-team/keras.git
def _tracking_metadata(self): metadata = json.loads(super()._tracking_metadata) metadata["_is_feature_layer"] = True return json.dumps(metadata, default=json_utils.get_json_type)
37
dense_features.py
Python
keras/feature_column/dense_features.py
84afc5193d38057e2e2badf9c889ea87d80d8fbf
keras
1
176,155
76
9
56
334
5
0
93
649
tutte_graph
Docstrings for the small.py module (#5240) * added description for the first 5 small graphs * modified descriptions based on comment and added description for two more functions * added doctrings to all the functions * Minor touchups. Co-authored-by: Ross Barnowski <[email protected]>
https://github.com/networkx/networkx.git
def tutte_graph(create_using=None): description = [ "adjacencylist", "Tutte's Graph", 46, [ [2, 3, 4], [5, 27], [11, 12], [19, 20], [6, 34], [7, 30], [8, 28], [9, 15], [10, 39], [11, 38], [40], [13, 40], [14, 36], [15, 16], [35], [17, 23], [18, 45], [19, 44], [46], [21, 46], [22, 42], [23, 24], [41], [25, 28], [26, 33], [27, 32], [34], [29], [30, 33], [31], [32, 34], [33], [], [], [36, 39], [37], [38, 40], [39], [], [], [42, 45], [43], [44, 46], [45], [], [], ], ] G = make_small_undirected_graph(description, create_using) return G
267
small.py
Python
networkx/generators/small.py
dec723f072eb997a497a159dbe8674cd39999ee9
networkx
1
111,539
49
11
9
250
21
1
53
92
test_append_invalid_alias
Refactor KB for easier customization (#11268) * Add implementation of batching + backwards compatibility fixes. Tests indicate issue with batch disambiguation for custom singular entity lookups. * Fix tests. Add distinction w.r.t. batch size. * Remove redundant and add new comments. * Adjust comments. Fix variable naming in EL prediction. * Fix mypy errors. * Remove KB entity type config option. Change return types of candidate retrieval functions to Iterable from Iterator. Fix various other issues. * Update spacy/pipeline/entity_linker.py Co-authored-by: Paul O'Leary McCann <[email protected]> * Update spacy/pipeline/entity_linker.py Co-authored-by: Paul O'Leary McCann <[email protected]> * Update spacy/kb_base.pyx Co-authored-by: Paul O'Leary McCann <[email protected]> * Update spacy/kb_base.pyx Co-authored-by: Paul O'Leary McCann <[email protected]> * Update spacy/pipeline/entity_linker.py Co-authored-by: Paul O'Leary McCann <[email protected]> * Add error messages to NotImplementedErrors. Remove redundant comment. * Fix imports. * Remove redundant comments. * Rename KnowledgeBase to InMemoryLookupKB and BaseKnowledgeBase to KnowledgeBase. * Fix tests. * Update spacy/errors.py Co-authored-by: Sofie Van Landeghem <[email protected]> * Move KB into subdirectory. * Adjust imports after KB move to dedicated subdirectory. * Fix config imports. * Move Candidate + retrieval functions to separate module. Fix other, small issues. * Fix docstrings and error message w.r.t. class names. Fix typing for candidate retrieval functions. * Update spacy/kb/kb_in_memory.pyx Co-authored-by: Sofie Van Landeghem <[email protected]> * Update spacy/ml/models/entity_linker.py Co-authored-by: Sofie Van Landeghem <[email protected]> * Fix typing. * Change typing of mentions to be Span instead of Union[Span, str]. * Update docs. * Update EntityLinker and _architecture docs. * Update website/docs/api/entitylinker.md Co-authored-by: Paul O'Leary McCann <[email protected]> * Adjust message for E1046. * Re-add section for Candidate in kb.md, add reference to dedicated page. * Update docs and docstrings. * Re-add section + reference for KnowledgeBase.get_alias_candidates() in docs. * Update spacy/kb/candidate.pyx * Update spacy/kb/kb_in_memory.pyx * Update spacy/pipeline/legacy/entity_linker.py * Remove canididate.md. Remove mistakenly added config snippet in entity_linker.py. Co-authored-by: Paul O'Leary McCann <[email protected]> Co-authored-by: Sofie Van Landeghem <[email protected]>
https://github.com/explosion/spaCy.git
def test_append_invalid_alias(nlp): mykb = InMemoryLookupKB(nlp.vocab, entity_vector_length=1) # adding entities mykb.add_entity(entity="Q1", freq=27, entity_vector=[1]) mykb.add_entity(entity="Q2", freq=12, entity_vector=[2]) mykb.add_entity(entity="Q3", freq=5, entity_vector=[3]) # adding aliases mykb.add_alias(alias="douglas", entities=["Q2", "Q3"], probabilities=[0.8, 0.1]) mykb.add_alias(alias="adam", entities=["Q2"], probabilities=[0.9]) # append an alias - should fail because the entities and probabilities vectors are not of equal length with pytest.raises(ValueError): mykb.append_alias(alias="douglas", entity="Q1", prior_prob=0.2) @pytest.mark.filterwarnings("ignore:\\[W036")
@pytest.mark.filterwarnings("ignore:\\[W036")
148
test_entity_linker.py
Python
spacy/tests/pipeline/test_entity_linker.py
1f23c615d7a7326ca5a38a7d768b8b70caaa0e17
spaCy
1
102,696
36
11
9
106
17
0
41
88
subscribe_to_coin_updates
Merge standalone wallet into main (#9793) * wallet changes from pac * cat changes * pool tests * pooling tests passing * offers * lint * mempool_mode * black * linting * workflow files * flake8 * more cleanup * renamed * remove obsolete test, don't cast announcement * memos are not only bytes32 * trade renames * fix rpcs, block_record * wallet rpc, recompile settlement clvm * key derivation * clvm tests * lgtm issues and wallet peers * stash * rename * mypy linting * flake8 * bad initializer * flaky tests * Make CAT wallets only create on verified hints (#9651) * fix clvm tests * return to log lvl warn * check puzzle unhardened * public key, not bytes. api caching change * precommit changes * remove unused import * mypy ci file, tests * ensure balance before creating a tx * Remove CAT logic from full node test (#9741) * Add confirmations and sleeps for wallet (#9742) * use pool executor * rever merge mistakes/cleanup * Fix trade test flakiness (#9751) * remove precommit * older version of black * lint only in super linter * Make announcements in RPC be objects instead of bytes (#9752) * Make announcements in RPC be objects instead of bytes * Lint * misc hint'ish cleanup (#9753) * misc hint'ish cleanup * unremove some ci bits * Use main cached_bls.py * Fix bad merge in main_pac (#9774) * Fix bad merge at 71da0487b9cd5564453ec24b76f1ac773c272b75 * Remove unused ignores * more unused ignores * Fix bad merge at 3b143e705057d6c14e2fb3e00078aceff0552d7e * One more byte32.from_hexstr * Remove obsolete test * remove commented out * remove duplicate payment object * remove long sync * remove unused test, noise * memos type * bytes32 * make it clear it's a single state at a time * copy over asset ids from pacr * file endl linter * Update chia/server/ws_connection.py Co-authored-by: dustinface <[email protected]> Co-authored-by: Matt Hauff <[email protected]> Co-authored-by: Kyle Altendorf <[email protected]> Co-authored-by: dustinface <[email protected]>
https://github.com/Chia-Network/chia-blockchain.git
async def subscribe_to_coin_updates(self, coin_names, peer, height=uint32(0)): msg = wallet_protocol.RegisterForCoinUpdates(coin_names, height) all_coins_state: Optional[RespondToCoinUpdates] = await peer.register_interest_in_coin(msg) # State for untrusted sync is processed only in wp sync | or short sync backwards if all_coins_state is not None and self.is_trusted(peer): await self.wallet_state_manager.new_coin_state(all_coins_state.coin_states, peer)
67
wallet_node.py
Python
chia/wallet/wallet_node.py
89f15f591cc3cc3e8ae40e95ffc802f7f2561ece
chia-blockchain
3
278,610
59
17
15
194
22
1
72
193
softmax
Remove pylint comments. PiperOrigin-RevId: 452353044
https://github.com/keras-team/keras.git
def softmax(x, axis=-1): if x.shape.rank > 1: if isinstance(axis, int): output = tf.nn.softmax(x, axis=axis) else: # nn.softmax does not support tuple axis. e = tf.exp(x - tf.reduce_max(x, axis=axis, keepdims=True)) s = tf.reduce_sum(e, axis=axis, keepdims=True) output = e / s else: raise ValueError( "Cannot apply softmax to a tensor that is 1D. " f"Received input: {x}" ) # Cache the logits to use for crossentropy loss. output._keras_logits = x return output @keras_export("keras.activations.elu") @tf.__internal__.dispatch.add_dispatch_support
@keras_export("keras.activations.elu") @tf.__internal__.dispatch.add_dispatch_support
104
activations.py
Python
keras/activations.py
3613c3defc39c236fb1592c4f7ba1a9cc887343a
keras
3
137,498
5
6
23
31
6
1
5
12
test_stop_job_gracefully
[Jobs] Use SIGTERM followed by SIGKILL to stop a job (#30851) Currently, when user wants to stop a job, we directly send a `SIGKILL` signal. Instead, we want to send a `SIGTERM` signal first, then send a `SIGKILL` signal after a few seconds if the child process still has not terminated.
https://github.com/ray-project/ray.git
async def test_stop_job_gracefully(job_manager): entrypoint =
entrypoint = """python -c \"
65
test_job_manager.py
Python
dashboard/modules/job/tests/test_job_manager.py
22af73253cd48fa134aee33c145b541c99acdc8b
ray
1
260,458
10
11
4
75
12
0
11
23
test_agglomerative_clustering_memory_mapped
MNT Deprecate `affinity` in `AgglomerativeClustering` (#23470) Co-authored-by: Thomas J. Fan <[email protected]> Co-authored-by: Guillaume Lemaitre <[email protected]>
https://github.com/scikit-learn/scikit-learn.git
def test_agglomerative_clustering_memory_mapped(): rng = np.random.RandomState(0) Xmm = create_memmap_backed_data(rng.randn(50, 100)) AgglomerativeClustering(metric="euclidean", linkage="single").fit(Xmm)
43
test_hierarchical.py
Python
sklearn/cluster/tests/test_hierarchical.py
a5d50cf3c7611b4343f446a97266ea77a4175afa
scikit-learn
1
198,198
46
13
16
211
28
0
61
193
sort_args_by_name
Rename files for array expression conversions in order to avoid naming conflicts in TAB-completion of the corresponding functions
https://github.com/sympy/sympy.git
def sort_args_by_name(self): expr = self.expr if not isinstance(expr, ArrayTensorProduct): return self args = expr.args sorted_data = sorted(enumerate(args), key=lambda x: default_sort_key(x[1])) pos_sorted, args_sorted = zip(*sorted_data) reordering_map = {i: pos_sorted.index(i) for i, arg in enumerate(args)} contraction_tuples = self._get_contraction_tuples() contraction_tuples = [[(reordering_map[j], k) for j, k in i] for i in contraction_tuples] c_tp = _array_tensor_product(*args_sorted) new_contr_indices = self._contraction_tuples_to_contraction_indices( c_tp, contraction_tuples ) return _array_contraction(c_tp, *new_contr_indices)
135
array_expressions.py
Python
sympy/tensor/array/expressions/array_expressions.py
a69c49bec6caf2cb460dc4eedf0fec184db92f0e
sympy
5
288,141
6
6
3
22
4
0
6
20
is_connected
Add ESPHome BleakClient (#78911) Co-authored-by: Paulus Schoutsen <[email protected]>
https://github.com/home-assistant/core.git
def is_connected(self) -> bool: return self._is_connected
12
client.py
Python
homeassistant/components/esphome/bluetooth/client.py
7042d6d35be54865b1252c0b28a50cce1a92eabc
core
1
276,209
7
11
5
61
10
0
7
30
_get_assets_dir
Reformatting the codebase with black. PiperOrigin-RevId: 450093126
https://github.com/keras-team/keras.git
def _get_assets_dir(export_dir): return tf.io.gfile.join( tf.compat.as_text(export_dir), tf.compat.as_text(tf.saved_model.ASSETS_DIRECTORY), )
38
saved_model_experimental.py
Python
keras/saving/saved_model_experimental.py
84afc5193d38057e2e2badf9c889ea87d80d8fbf
keras
1
156,048
40
14
8
163
18
0
43
87
new_da_object
absolufy-imports - No relative - PEP8 (#8796) Conversation in https://github.com/dask/distributed/issues/5889
https://github.com/dask/dask.git
def new_da_object(dsk, name, chunks, meta=None, dtype=None): if is_dataframe_like(meta) or is_series_like(meta) or is_index_like(meta): from dask.dataframe.core import new_dd_object assert all(len(c) == 1 for c in chunks[1:]) divisions = [None] * (len(chunks[0]) + 1) return new_dd_object(dsk, name, meta, divisions) else: return Array(dsk, name=name, chunks=chunks, meta=meta, dtype=dtype)
111
core.py
Python
dask/array/core.py
cccb9d8d8e33a891396b1275c2448c352ef40c27
dask
5
191,686
11
10
5
55
7
0
11
27
test_run_multiple_args_error
change run to use args and kwargs (#367) Before, `run` was not able to be called with multiple arguments. This expands the functionality.
https://github.com/hwchase17/langchain.git
def test_run_multiple_args_error() -> None: chain = FakeChain() with pytest.raises(ValueError): chain.run("bar", "foo")
28
test_base.py
Python
tests/unit_tests/chains/test_base.py
8d0869c6d3ed63b2b15d4f75ea664e089dcc569d
langchain
1
107,167
21
11
5
106
16
1
22
40
test_constrained_layout4
ENH: implement and use base layout_engine for more flexible layout.
https://github.com/matplotlib/matplotlib.git
def test_constrained_layout4(): fig, axs = plt.subplots(2, 2, layout="constrained") for ax in axs.flat: pcm = example_pcolor(ax, fontsize=24) fig.colorbar(pcm, ax=axs, pad=0.01, shrink=0.6) @image_comparison(['constrained_layout5.png'], tol=0.002)
@image_comparison(['constrained_layout5.png'], tol=0.002)
60
test_constrainedlayout.py
Python
lib/matplotlib/tests/test_constrainedlayout.py
ec4dfbc3c83866f487ff0bc9c87b0d43a1c02b22
matplotlib
2
20,474
18
11
6
93
12
0
19
69
check
check point progress on only bringing in pip==22.0.4 (#4966) * vendor in pip==22.0.4 * updating vendor packaging version * update pipdeptree to fix pipenv graph with new version of pip. * Vendoring of pip-shims 0.7.0 * Vendoring of requirementslib 1.6.3 * Update pip index safety restrictions patch for pip==22.0.4 * Update patches * exclude pyptoject.toml from black to see if that helps. * Move this part of the hash collection back to the top (like prior implementation) because it affects the outcome of this test now in pip 22.0.4
https://github.com/pypa/pipenv.git
def check(self, pattern): if self.eos: raise EndOfText() if pattern not in self._re_cache: self._re_cache[pattern] = re.compile(pattern, self.flags) return self._re_cache[pattern].match(self.data, self.pos)
60
scanner.py
Python
pipenv/patched/notpip/_vendor/pygments/scanner.py
f3166e673fe8d40277b804d35d77dcdb760fc3b3
pipenv
3
5,819
23
11
4
67
6
0
36
58
extract_text_from_element
PR - Fix `extract_text_from_element()`and `find_element*()` to `find_element()` (#6438) * Updated getUserData() and find_element* Signed-off-by: elulcao <[email protected]> Thanks @breuerfelix for reviewing, 🚀 People in this thread please let me know if something is not OK, IG changed a lot these days. 🤗 @her
https://github.com/InstaPy/InstaPy.git
def extract_text_from_element(elem): # if element is valid and contains text withou spaces if elem and hasattr(elem, "text") and elem.text and not re.search("\s", elem.text): return elem.text # if the element is not valid, return None return None
38
util.py
Python
instapy/util.py
2a157d452611d37cf50ccb7d56ff1a06e9790ecb
InstaPy
5
98,080
49
15
23
261
40
0
55
279
delete_alert_rule
feat(workflow): Add audit logs on add/edit/remove metric alerts (#33296)
https://github.com/getsentry/sentry.git
def delete_alert_rule(alert_rule, user=None): if alert_rule.status == AlertRuleStatus.SNAPSHOT.value: raise AlreadyDeletedError() with transaction.atomic(): if user: create_audit_entry_from_user( user, organization_id=alert_rule.organization_id, target_object=alert_rule.id, data=alert_rule.get_audit_log_data(), event=AuditLogEntryEvent.ALERT_RULE_REMOVE, ) incidents = Incident.objects.filter(alert_rule=alert_rule) bulk_delete_snuba_subscriptions(list(alert_rule.snuba_query.subscriptions.all())) if incidents.exists(): alert_rule.update(status=AlertRuleStatus.SNAPSHOT.value) AlertRuleActivity.objects.create( alert_rule=alert_rule, user=user, type=AlertRuleActivityType.DELETED.value ) else: alert_rule.delete() if alert_rule.id: # Change the incident status asynchronously, which could take awhile with many incidents due to snapshot creations. tasks.auto_resolve_snapshot_incidents.apply_async(kwargs={"alert_rule_id": alert_rule.id})
162
logic.py
Python
src/sentry/incidents/logic.py
5b01e6a61cbbb62bde6e1cacf155e00c5e5bc432
sentry
5
177,063
128
15
23
407
30
0
239
508
strategy_saturation_largest_first
strategy_saturation_largest_first now accepts partial colorings (#5888) The coloring strategy function `strategy_saturation_largest_first` allows the user to pass in a dict with a partial node coloring. This was always allowed, but previously the partial coloring dict was ignored. The partial coloring is also verified within the function.
https://github.com/networkx/networkx.git
def strategy_saturation_largest_first(G, colors): distinct_colors = {v: set() for v in G} # Add the node color assignments given in colors to the # distinct colors set for each neighbor of that node for vertex, color in colors.items(): for neighbor in G[vertex]: distinct_colors[neighbor].add(color) # Check that the color assignments in colors are valid # i.e. no neighboring nodes have the same color if len(colors) >= 2: for vertex, color in colors.items(): if color in distinct_colors[vertex]: raise nx.NetworkXError( "Neighboring vertices must have different colors" ) # If 0 nodes have been colored, simply choose the node of highest degree. if not colors: node = max(G, key=G.degree) yield node # Add the color 0 to the distinct colors set for each # neighbor of that node. for v in G[node]: distinct_colors[v].add(0) for i in range(len(G) - len(colors)): # Compute the maximum saturation and the set of nodes that # achieve that saturation. saturation = {v: len(c) for v, c in distinct_colors.items() if v not in colors} # Yield the node with the highest saturation, and break ties by # degree. node = max(saturation, key=lambda v: (saturation[v], G.degree(v))) yield node # Update the distinct color sets for the neighbors. color = colors[node] for v in G[node]: distinct_colors[v].add(color) #: Dictionary mapping name of a strategy as a string to the strategy function. STRATEGIES = { "largest_first": strategy_largest_first, "random_sequential": strategy_random_sequential, "smallest_last": strategy_smallest_last, "independent_set": strategy_independent_set, "connected_sequential_bfs": strategy_connected_sequential_bfs, "connected_sequential_dfs": strategy_connected_sequential_dfs, "connected_sequential": strategy_connected_sequential, "saturation_largest_first": strategy_saturation_largest_first, "DSATUR": strategy_saturation_largest_first, }
209
greedy_coloring.py
Python
networkx/algorithms/coloring/greedy_coloring.py
77d7ddac9a7c69ff086dd825e55454f300f4242b
networkx
13
60,244
20
10
9
83
3
0
34
74
choose_color_by_layertype
Balanced joint maximum mean discrepancy for deep transfer learning
https://github.com/jindongwang/transferlearning.git
def choose_color_by_layertype(layertype): color = '#6495ED' # Default if layertype == 'Convolution' or layertype == 'Deconvolution': color = '#FF5050' elif layertype == 'Pooling': color = '#FF9900' elif layertype == 'InnerProduct': color = '#CC33FF' return color
39
draw.py
Python
code/deep/BJMMD/caffe/python/caffe/draw.py
cc4d0564756ca067516f71718a3d135996525909
transferlearning
5
259,211
35
16
13
145
18
0
48
210
_compute_transformed_categories
ENH Adds infrequent categories to OneHotEncoder (#16018) * ENH Completely adds infrequent categories * STY Linting * STY Linting * DOC Improves wording * DOC Lint * BUG Fixes * CLN Address comments * CLN Address comments * DOC Uses math to description float min_frequency * DOC Adds comment regarding drop * BUG Fixes method name * DOC Clearer docstring * TST Adds more tests * FIX Fixes mege * CLN More pythonic * CLN Address comments * STY Flake8 * CLN Address comments * DOC Fix * MRG * WIP * ENH Address comments * STY Fix * ENH Use functiion call instead of property * ENH Adds counts feature * CLN Rename variables * DOC More details * CLN Remove unneeded line * CLN Less lines is less complicated * CLN Less diffs * CLN Improves readiabilty * BUG Fix * CLN Address comments * TST Fix * CLN Address comments * CLN Address comments * CLN Move docstring to userguide * DOC Better wrapping * TST Adds test to handle_unknown='error' * ENH Spelling error in docstring * BUG Fixes counter with nan values * BUG Removes unneeded test * BUG Fixes issue * ENH Sync with main * DOC Correct settings * DOC Adds docstring * DOC Immprove user guide * DOC Move to 1.0 * DOC Update docs * TST Remove test * DOC Update docstring * STY Linting * DOC Address comments * ENH Neater code * DOC Update explaination for auto * Update sklearn/preprocessing/_encoders.py Co-authored-by: Roman Yurchak <[email protected]> * TST Uses docstring instead of comments * TST Remove call to fit * TST Spelling error * ENH Adds support for drop + infrequent categories * ENH Adds infrequent_if_exist option * DOC Address comments for user guide * DOC Address comments for whats_new * DOC Update docstring based on comments * CLN Update test with suggestions * ENH Adds computed property infrequent_categories_ * DOC Adds where the infrequent column is located * TST Adds more test for infrequent_categories_ * DOC Adds docstring for _compute_drop_idx * CLN Moves _convert_to_infrequent_idx into its own method * TST Increases test coverage * TST Adds failing test * CLN Careful consideration of dropped and inverse_transform * STY Linting * DOC Adds docstrinb about dropping infrequent * DOC Uses only * DOC Numpydoc * TST Includes test for get_feature_names_out * DOC Move whats new * DOC Address docstring comments * DOC Docstring changes * TST Better comments * TST Adds check for handle_unknown='ignore' for infrequent * CLN Make _infrequent_indices private * CLN Change min_frequency default to None * DOC Adds comments * ENH adds support for max_categories=1 * ENH Describe lexicon ordering for ties * DOC Better docstring * STY Fix * CLN Error when explicity dropping an infrequent category * STY Grammar Co-authored-by: Joel Nothman <[email protected]> Co-authored-by: Roman Yurchak <[email protected]> Co-authored-by: Guillaume Lemaitre <[email protected]>
https://github.com/scikit-learn/scikit-learn.git
def _compute_transformed_categories(self, i, remove_dropped=True): cats = self.categories_[i] if self._infrequent_enabled: infreq_map = self._default_to_infrequent_mappings[i] if infreq_map is not None: frequent_mask = infreq_map < infreq_map.max() infrequent_cat = "infrequent_sklearn" # infrequent category is always at the end cats = np.concatenate( (cats[frequent_mask], np.array([infrequent_cat], dtype=object)) ) if remove_dropped: cats = self._remove_dropped_categories(cats, i) return cats
92
_encoders.py
Python
sklearn/preprocessing/_encoders.py
7f0006c8aad1a09621ad19c3db19c3ff0555a183
scikit-learn
4
32,877
30
10
14
86
9
0
33
60
get_key_to_not_convert
`bitsandbytes` - `Linear8bitLt` integration into `transformers` models (#17901) * first commit * correct replace function * add final changes - works like charm! - cannot implement tests yet - tested * clean up a bit * add bitsandbytes dependencies * working version - added import function - added bitsandbytes utils file * small fix * small fix - fix import issue * fix import issues * Apply suggestions from code review Co-authored-by: Sylvain Gugger <[email protected]> * refactor a bit - move bitsandbytes utils to utils - change comments on functions * reformat docstring - reformat docstring on init_empty_weights_8bit * Update src/transformers/__init__.py Co-authored-by: Sylvain Gugger <[email protected]> * revert bad formatting * change to bitsandbytes * refactor a bit - remove init8bit since it is useless * more refactoring - fixed init empty weights issue - added threshold param * small hack to make it work * Update src/transformers/modeling_utils.py * Update src/transformers/modeling_utils.py * revmoe the small hack * modify utils file * make style + refactor a bit * create correctly device map * add correct dtype for device map creation * Apply suggestions from code review Co-authored-by: Sylvain Gugger <[email protected]> * apply suggestions - remove with torch.grad - do not rely on Python bool magic! * add docstring - add docstring for new kwargs * add docstring - comment `replace_8bit_linear` function - fix weird formatting * - added more documentation - added new utility function for memory footprint tracking - colab demo to add * few modifs - typo doc - force cast into float16 when load_in_8bit is enabled * added colab link * add test architecture + docstring a bit * refactor a bit testing class * make style + refactor a bit * enhance checks - add more checks - start writing saving test * clean up a bit * male style * add more details on doc * add more tests - still needs to fix 2 tests * replace by "or" - could not fix it from GitHub GUI Co-authored-by: Sylvain Gugger <[email protected]> * refactor a bit testing code + add readme * make style * fix import issue * Update src/transformers/modeling_utils.py Co-authored-by: Michael Benayoun <[email protected]> * add few comments * add more doctring + make style * more docstring * raise error when loaded in 8bit * make style * add warning if loaded on CPU * add small sanity check * fix small comment * add bitsandbytes on dockerfile * Improve documentation - improve documentation from comments * add few comments * slow tests pass on the VM but not on the CI VM * Fix merge conflict * make style * another test should pass on a multi gpu setup * fix bad import in testing file * Fix slow tests - remove dummy batches - no more CUDA illegal memory errors * odify dockerfile * Update docs/source/en/main_classes/model.mdx * Update Dockerfile * Update model.mdx * Update Dockerfile * Apply suggestions from code review * few modifications - lm head can stay on disk/cpu - change model name so that test pass * change test value - change test value to the correct output - torch bmm changed to baddmm in bloom modeling when merging * modify installation guidelines * Apply suggestions from code review Co-authored-by: Sylvain Gugger <[email protected]> * Apply suggestions from code review Co-authored-by: Sylvain Gugger <[email protected]> * Apply suggestions from code review Co-authored-by: Sylvain Gugger <[email protected]> * replace `n`by `name` * merge `load_in_8bit` and `low_cpu_mem_usage` * first try - keep the lm head in full precision * better check - check the attribute `base_model_prefix` instead of computing the number of parameters * added more tests * Update src/transformers/utils/bitsandbytes.py Co-authored-by: Sylvain Gugger <[email protected]> * Merge branch 'integration-8bit' of https://github.com/younesbelkada/transformers into integration-8bit * improve documentation - fix typos for installation - change title in the documentation Co-authored-by: Sylvain Gugger <[email protected]> Co-authored-by: Michael Benayoun <[email protected]>
https://github.com/huggingface/transformers.git
def get_key_to_not_convert(model): r # Ignore this for base models (BertModel, GPT2Model, etc.) if not hasattr(model, model.base_model_prefix): return "" # otherwise they have an attached head list_modules = list(model.named_parameters()) last_name = list_modules[-1][0] return last_name.split(".")[0]
50
bitsandbytes.py
Python
src/transformers/utils/bitsandbytes.py
4a51075a96d2049f368b5f3dd6c0e9f08f599b62
transformers
2
156,649
10
8
2
58
6
0
10
24
split
Include known inconsistency in DataFrame `str.split` accessor docstring (#9177)
https://github.com/dask/dask.git
def split(self, pat=None, n=-1, expand=False): return self._split("split", pat=pat, n=n, expand=expand)
38
accessor.py
Python
dask/dataframe/accessor.py
dadfd9b681997b20026e6a51afef3fb9ebb1513a
dask
1
19,833
39
22
16
232
29
1
49
188
test_install_venv_project_directory
More granular control over PIPENV_VENV_IN_PROJECT variable. (#5026) * Allow PIPENV_VENV_IN_PROJECT to be read in as None, and ensure if it is set to False that it does not use .venv directory. * refactor based on PR feedback and add news fragment. * Review unit test coverage and add new tests. Remove unneccesary bits from other tests.
https://github.com/pypa/pipenv.git
def test_install_venv_project_directory(PipenvInstance): with PipenvInstance(chdir=True) as p: with temp_environ(), TemporaryDirectory( prefix="pipenv-", suffix="temp_workon_home" ) as workon_home: os.environ["WORKON_HOME"] = workon_home c = p.pipenv("install six") assert c.returncode == 0 venv_loc = None for line in c.stderr.splitlines(): if line.startswith("Virtualenv location:"): venv_loc = Path(line.split(":", 1)[-1].strip()) assert venv_loc is not None assert venv_loc.joinpath(".project").exists() @pytest.mark.cli @pytest.mark.deploy @pytest.mark.system
@pytest.mark.cli @pytest.mark.deploy @pytest.mark.system
129
test_install_basic.py
Python
tests/integration/test_install_basic.py
949ee95d6748e8777bed589f0d990aa4792b28f8
pipenv
4
83,182
50
11
15
172
18
0
60
158
test_json_get_subscribers_for_guest_user
docs: Consistently hyphenate “web-public”. In English, compound adjectives should essentially always be hyphenated. This makes them easier to parse, especially for users who might not recognize that the words “web public” go together as a phrase. Signed-off-by: Anders Kaseorg <[email protected]>
https://github.com/zulip/zulip.git
def test_json_get_subscribers_for_guest_user(self) -> None: guest_user = self.example_user("polonius") never_subscribed = gather_subscriptions_helper(guest_user, True).never_subscribed # A guest user can only see never subscribed streams that are web-public. # For Polonius, the only web-public stream that he is not subscribed at # this point is Rome. self.assert_length(never_subscribed, 1) web_public_stream_id = never_subscribed[0]["stream_id"] result = self.client_get(f"/json/streams/{web_public_stream_id}/members") self.assert_json_success(result) result_dict = result.json() self.assertIn("subscribers", result_dict) self.assertIsInstance(result_dict["subscribers"], list) self.assertGreater(len(result_dict["subscribers"]), 0)
98
test_subs.py
Python
zerver/tests/test_subs.py
90e202cd38d00945c81da4730d39e3f5c5b1e8b1
zulip
1
245,574
90
16
30
456
30
0
144
486
convert
[Fix] replace mmcv's function and modules imported with mmengine's (#8594) * use mmengine's load_state_dict and load_checkpoint * from mmengine import dump * from mmengine import FileClient dump list_from_file * remove redundant registry * update * update * update * replace _load_checkpoint with CheckpointLoad.load_checkpoint * changes according to mmcv #2216 * changes due to mmengine #447 * changes due mmengine #447 and mmcv #2217 * changes due mmengine #447 and mmcv #2217 * update * update * update
https://github.com/open-mmlab/mmdetection.git
def convert(src, dst, depth): # load arch_settings if depth not in arch_settings: raise ValueError('Only support ResNet-50 and ResNet-101 currently') block_nums = arch_settings[depth] # load caffe model caffe_model = load(src, encoding='latin1') blobs = caffe_model['blobs'] if 'blobs' in caffe_model else caffe_model # convert to pytorch style state_dict = OrderedDict() converted_names = set() convert_conv_fc(blobs, state_dict, 'conv1', 'conv1', converted_names) convert_bn(blobs, state_dict, 'res_conv1_bn', 'bn1', converted_names) for i in range(1, len(block_nums) + 1): for j in range(block_nums[i - 1]): if j == 0: convert_conv_fc(blobs, state_dict, f'res{i + 1}_{j}_branch1', f'layer{i}.{j}.downsample.0', converted_names) convert_bn(blobs, state_dict, f'res{i + 1}_{j}_branch1_bn', f'layer{i}.{j}.downsample.1', converted_names) for k, letter in enumerate(['a', 'b', 'c']): convert_conv_fc(blobs, state_dict, f'res{i + 1}_{j}_branch2{letter}', f'layer{i}.{j}.conv{k+1}', converted_names) convert_bn(blobs, state_dict, f'res{i + 1}_{j}_branch2{letter}_bn', f'layer{i}.{j}.bn{k + 1}', converted_names) # check if all layers are converted for key in blobs: if key not in converted_names: print(f'Not Convert: {key}') # save checkpoint checkpoint = dict() checkpoint['state_dict'] = state_dict torch.save(checkpoint, dst)
223
detectron2pytorch.py
Python
tools/model_converters/detectron2pytorch.py
d0695e68654ca242be54e655491aef8c959ac345
mmdetection
9
249,486
56
11
15
177
17
0
71
132
check_rust_lib_up_to_date
Check if Rust lib needs rebuilding. (#13759) This protects against the common mistake of failing to remember to rebuild Rust code after making changes.
https://github.com/matrix-org/synapse.git
def check_rust_lib_up_to_date() -> None: if not _dist_is_editable(): return synapse_dir = os.path.dirname(synapse.__file__) synapse_root = os.path.abspath(os.path.join(synapse_dir, "..")) # Double check we've not gone into site-packages... if os.path.basename(synapse_root) == "site-packages": return # ... and it looks like the root of a python project. if not os.path.exists("pyproject.toml"): return # Get the hash of all Rust source files hash = _hash_rust_files_in_directory(os.path.join(synapse_root, "rust", "src")) if hash != get_rust_file_digest(): raise Exception("Rust module outdated. Please rebuild using `poetry install`")
99
rust.py
Python
synapse/util/rust.py
ebfeac7c5ded851a2639911ec6adf9d0fcdb029a
synapse
5
176,497
7
9
118
33
6
0
7
12
generate_gml
Update black (#5438) * CI: sync up black dev requirements version with precommit * Run black Co-authored-by: Jarrod Millman <[email protected]>
https://github.com/networkx/networkx.git
def generate_gml(G, stringizer=None): r valid_keys = re.compile("^[A-Za-z][0-9A-Za-z_]*$")
285
gml.py
Python
networkx/readwrite/gml.py
f6755ffa00211b523c6c0bec5398bc6c3c43c8b1
networkx
10
270,714
5
7
2
26
4
0
5
19
_defun_call
Reformatting the codebase with black. PiperOrigin-RevId: 450093126
https://github.com/keras-team/keras.git
def _defun_call(self, inputs): return self._make_op(inputs)
15
base_layer.py
Python
keras/engine/base_layer.py
84afc5193d38057e2e2badf9c889ea87d80d8fbf
keras
1
291,724
22
10
11
118
13
0
27
72
skybell_mock
Upgrade pytest-aiohttp (#82475) * Upgrade pytest-aiohttp * Make sure executors, tasks and timers are closed Some test will trigger warnings on garbage collect, these warnings spills over into next test. Some test trigger tasks that raise errors on shutdown, these spill over into next test. This is to mimic older pytest-aiohttp and it's behaviour on test cleanup. Discussions on similar changes for pytest-aiohttp are here: https://github.com/pytest-dev/pytest-asyncio/pull/309 * Replace loop with event_loop * Make sure time is frozen for tests * Make sure the ConditionType is not async /home-assistant/homeassistant/helpers/template.py:2082: RuntimeWarning: coroutine 'AsyncMockMixin._execute_mock_call' was never awaited def wrapper(*args, **kwargs): Enable tracemalloc to get traceback where the object was allocated. See https://docs.pytest.org/en/stable/how-to/capture-warnings.html#resource-warnings for more info. * Increase litejet press tests with a factor 10 The times are simulated anyway, and we can't stop the normal event from occuring. * Use async handlers for aiohttp tests/components/motioneye/test_camera.py::test_get_still_image_from_camera tests/components/motioneye/test_camera.py::test_get_still_image_from_camera tests/components/motioneye/test_camera.py::test_get_stream_from_camera tests/components/motioneye/test_camera.py::test_get_stream_from_camera tests/components/motioneye/test_camera.py::test_camera_option_stream_url_template tests/components/motioneye/test_camera.py::test_camera_option_stream_url_template /Users/joakim/src/hass/home-assistant/venv/lib/python3.9/site-packages/aiohttp/web_urldispatcher.py:189: DeprecationWarning: Bare functions are deprecated, use async ones warnings.warn( * Switch to freezegun in modbus tests The tests allowed clock to tick in between steps * Make sure skybell object are fully mocked Old tests would trigger attempts to post to could services: ``` DEBUG:aioskybell:HTTP post https://cloud.myskybell.com/api/v3/login/ Request with headers: {'content-type': 'application/json', 'accept': '*/*', 'x-skybell-app-id': 'd2b542c7-a7e4-4e1e-b77d-2b76911c7c46', 'x-skybell-client-id': '1f36a3c0-6dee-4997-a6db-4e1c67338e57'} ``` * Fix sorting that broke after rebase
https://github.com/home-assistant/core.git
def skybell_mock(): mocked_skybell_device = AsyncMock(spec=SkybellDevice) mocked_skybell = AsyncMock(spec=Skybell) mocked_skybell.async_get_devices.return_value = [mocked_skybell_device] mocked_skybell.async_send_request.return_value = {"id": USER_ID} mocked_skybell.user_id = USER_ID with patch( "homeassistant.components.skybell.config_flow.Skybell", return_value=mocked_skybell, ), patch("homeassistant.components.skybell.Skybell", return_value=mocked_skybell): yield mocked_skybell
68
conftest.py
Python
tests/components/skybell/conftest.py
c576a68d336bc91fd82c299d9b3e5dfdc1c14960
core
1
293,022
57
15
16
193
21
0
113
278
async_update_group_state
Improve binary sensor group when member is unknown or unavailable (#67468)
https://github.com/home-assistant/core.git
def async_update_group_state(self) -> None: all_states = [self.hass.states.get(x) for x in self._entity_ids] # filtered_states are members currently in the state machine filtered_states: list[str] = [x.state for x in all_states if x is not None] # Set group as unavailable if all members are unavailable self._attr_available = any( state != STATE_UNAVAILABLE for state in filtered_states ) valid_state = self.mode( state not in (STATE_UNKNOWN, STATE_UNAVAILABLE) for state in filtered_states ) if not valid_state: # Set as unknown if any / all member is not unknown or unavailable self._attr_is_on = None else: # Set as ON if any / all member is ON states = list(map(lambda x: x == STATE_ON, filtered_states)) state = self.mode(states) self._attr_is_on = state
122
binary_sensor.py
Python
homeassistant/components/group/binary_sensor.py
c5dd5e18c0443b332dddfbbf4a8ff03374c5c070
core
7
304,515
13
9
6
53
4
0
13
56
_async_stop
Auto recover when the Bluetooth adapter stops responding (#77043)
https://github.com/home-assistant/core.git
async def _async_stop(self) -> None: if self._cancel_watchdog: self._cancel_watchdog() self._cancel_watchdog = None await self._async_stop_scanner()
29
scanner.py
Python
homeassistant/components/bluetooth/scanner.py
ced8278e3222501dde7d769ea4b57aae75f62438
core
2
320,177
100
14
32
384
35
0
169
528
decide_what_tags_to_keep
Fixes deleting images if the branch API returns an error code and makes the error code a failure
https://github.com/paperless-ngx/paperless-ngx.git
def decide_what_tags_to_keep(self): # Default to everything gets kept still super().decide_what_tags_to_keep() # Locate the feature branches feature_branches = {} for branch in self.branch_api.get_branches( repo=self.repo_name, ): if branch.name.startswith("feature-"): logger.debug(f"Found feature branch {branch.name}") feature_branches[branch.name] = branch logger.info(f"Located {len(feature_branches)} feature branches") if not len(feature_branches): # Our work here is done, delete nothing return # Filter to packages which are tagged with feature-* packages_tagged_feature: List[ContainerPackage] = [] for package in self.all_package_versions: if package.tag_matches("feature-"): packages_tagged_feature.append(package) # Map tags like "feature-xyz" to a ContainerPackage feature_pkgs_tags_to_versions: Dict[str, ContainerPackage] = {} for pkg in packages_tagged_feature: for tag in pkg.tags: feature_pkgs_tags_to_versions[tag] = pkg logger.info( f'Located {len(feature_pkgs_tags_to_versions)} versions of package {self.package_name} tagged "feature-"', ) # All the feature tags minus all the feature branches leaves us feature tags # with no corresponding branch self.tags_to_delete = list( set(feature_pkgs_tags_to_versions.keys()) - set(feature_branches.keys()), ) # All the tags minus the set of going to be deleted tags leaves us the # tags which will be kept around self.tags_to_keep = list( set(self.all_pkgs_tags_to_version.keys()) - set(self.tags_to_delete), ) logger.info( f"Located {len(self.tags_to_delete)} versions of package {self.package_name} to delete", )
199
cleanup-tags.py
Python
.github/scripts/cleanup-tags.py
9214b412556e994dcf32dd2c7e050169f5cace1c
paperless-ngx
8
134,379
6
6
8
22
4
0
6
20
is_multi_agent
[RLlib] AlgorithmConfig: Next steps (volume 01); Algos, RolloutWorker, PolicyMap, WorkerSet use AlgorithmConfig objects under the hood. (#29395)
https://github.com/ray-project/ray.git
def is_multi_agent(self) -> bool: return self._is_multi_agent
12
algorithm_config.py
Python
rllib/algorithms/algorithm_config.py
182744bbd151c166b8028355eae12a5da63fb3cc
ray
1
100,648
11
9
3
42
6
0
11
25
_skip_num
bugfix: extract - stop progress bar from going over max value
https://github.com/deepfakes/faceswap.git
def _skip_num(self) -> int: return self._args.extract_every_n if hasattr(self._args, "extract_every_n") else 1
25
extract.py
Python
scripts/extract.py
0d23714875f81ddabdbe8f4e40bef6e5f29eeb19
faceswap
2
169,398
43
14
16
204
17
0
67
237
_wrap_result
Bug fix using GroupBy.resample produces inconsistent behavior when calling it over empty df #47705 (#47672) * DOC #45443 edited the documentation of where/mask functions * DOC #45443 edited the documentation of where/mask functions * Update generic.py * Bug 43767 fix * fixing lines doc * visual indent * visual indent * indentation * grouby.resample bug * groubby.sample * syntax * syntax * syntax * syntax * what's new * flake8 error * pytest * blank line * editting resample * editting resample * syntax * syntax * syntax * space * space * space * inplace * spelling * test * test resampler * tests * tests * Update resample.py * Update resample.py * Update resample.py * Update v1.6.0.rst
https://github.com/pandas-dev/pandas.git
def _wrap_result(self, result): # GH 47705 obj = self.obj if ( isinstance(result, ABCDataFrame) and result.empty and not isinstance(result.index, PeriodIndex) ): result = result.set_index( _asfreq_compat(obj.index[:0], freq=self.freq), append=True ) if isinstance(result, ABCSeries) and self._selection is not None: result.name = self._selection if isinstance(result, ABCSeries) and result.empty: # When index is all NaT, result is empty but index is not result.index = _asfreq_compat(obj.index[:0], freq=self.freq) result.name = getattr(obj, "name", None) return result
132
resample.py
Python
pandas/core/resample.py
fba672389b74ca4afece56040ae079a1f2b71544
pandas
8
297,321
9
8
3
45
6
1
9
14
mock_expires_at
Google Assistant SDK integration (#82328) * Copy google_sheets to google_assistant_sdk This is to improve diff of the next commit with the actual implementation. Commands used: cp -r homeassistant/components/google_sheets/ homeassistant/components/google_assistant_sdk/ cp -r tests/components/google_sheets/ tests/components/google_assistant_sdk/ find homeassistant/components/google_assistant_sdk/ tests/components/google_assistant_sdk/ -type f | xargs sed -i \ -e 's@google_sheets@google_assistant_sdk@g' \ -e 's@Google Sheets@Google Assistant SDK@g' \ -e 's@tkdrob@tronikos@g' * Google Assistant SDK integration Allows sending commands and broadcast messages to Google Assistant. * Remove unnecessary async_entry_has_scopes check * Bump gassist-text to fix protobuf dependency
https://github.com/home-assistant/core.git
def mock_expires_at() -> int: return time.time() + 3600 @pytest.fixture(name="config_entry")
@pytest.fixture(name="config_entry")
15
conftest.py
Python
tests/components/google_assistant_sdk/conftest.py
5d316734659cc331658ea2f77a2985cd2c58d043
core
1
284,585
18
10
24
96
10
0
25
60
print_help
Bounty Hunter mood: 11 bugs fixed (#1853) * fix #1850 * fix #1831 * add extra check to Reddit API keys * ignore warning message to update praw api * improve OpenBB links * fix quick performance only on stocks class because I'm James bitch * fix quick performance only on stocks class because I'm James bitch * fix #1829 * fix #1821 * add messari to keys - fix #1819 * example of multiple oclumns to check on options/chains * minor improvement in xlabel re. #1814 * remove repeated command * fix #1698 * fix line too long * fix #1814 fr now * fix tests
https://github.com/OpenBB-finance/OpenBBTerminal.git
def print_help(self): has_ticker_start = "[unvl]" if not self.ticker else "" has_ticker_end = "[/unvl]" if not self.ticker else "" help_text = f console.print(text=help_text, menu="Stocks - Government")
42
gov_controller.py
Python
openbb_terminal/stocks/government/gov_controller.py
a6f7e111e68346aeab315985b3704c2710693b38
OpenBBTerminal
3
252,298
40
15
11
195
19
0
46
150
standalone_binaries
simplify ci build script, add MSIX installer and Microsoft store.
https://github.com/mitmproxy/mitmproxy.git
def standalone_binaries(): with archive(DIST_DIR / f"mitmproxy-{version()}-{operating_system()}") as f: _pyinstaller("standalone.spec") for tool in ["mitmproxy", "mitmdump", "mitmweb"]: executable = TEMP_DIR / "pyinstaller/dist" / tool if platform.system() == "Windows": executable = executable.with_suffix(".exe") # Test if it works at all O:-) print(f"> {executable} --version") subprocess.check_call([executable, "--version"]) f.add(str(executable), str(executable.name)) print(f"Packed {f.name}.")
91
build.py
Python
release/build.py
3872d33111d27430c4b5f1ae021e91e3522cc0e3
mitmproxy
3
144,311
16
9
15
79
12
0
18
64
_get_all_child_nodes
[Ray DAG] Implement experimental Ray DAG API for task/class (#22058)
https://github.com/ray-project/ray.git
def _get_all_child_nodes(self) -> Set["DAGNode"]: scanner = _PyObjScanner() children = set() for n in scanner.find_nodes([self._bound_args, self._bound_kwargs]): children.add(n) return children
47
dag_node.py
Python
python/ray/experimental/dag/dag_node.py
c065e3f69ec248383d98b45a8d1c00832ccfdd57
ray
2
285,873
33
17
11
171
18
0
40
154
refresh_datasets_on_menu
Forecasting Menu [Work in Progress] (#1933) * Gave forecasting memory * Fixed scripts, refactored * FIxed poetry lock * edge case check for forecast target * Improved combine and load functionality * Cleaned up translations * Fixed issue with covariates * Fixed issue checking covariates * Another covariates check fix * Ignored regr and linregr warnings * Fixed covariate issues * switched from forecasting to forecast * Finished transition to forecast * Can add entire dataset with one command * Improved combine description * Removed naming covariates * Created new installation * typo * Make plot show dates if available * Added better handling or users without the menu * Removed unused file * Fix * Better handling for nontraditional datasets * Fixed black and pylint * Fixed tests * Added darts install to main tests * Working on darts with CI * Added back test file * Made large tables print better * naive baseline * typo * Finished naive * no dollar on prediction * fixed positive MAPE bug * quick refactoring * Fixed two different args for same thing * added extra patience * linreg mape fix * info fix * Refactored API, bumped to Darts 0.21.0 * Added fixes * Increased verbosity for wrong column * Updated dependencies * Hid warnings * Fixed importing * Fixed tests * Fixed ugly seasonal plotting * Fixed forecast line color * Switched chart output to blue * Simplified lambda_price_prediction_color * fixed residuals * Chnage * Removed darts from CI per Chavi * Added fixes to tests * Added knnfix * Fixed issue where n!= o * Added changes * Added changes * Imrpoved forecast dash * Added Theo notebook * Added enhancements to dash * Added notebook * Added fix for jupyter lab * Added debug stuff * Change * Updated docs * Fixed formatting * Fixed formatting * Removed prints * Filtered some info * Added button to run model * Improved api * Added secret feautr (no peeking Martin) * Cleaned code * Fixed tests * Added test fixes * Added fixes * Fixes * FIxes for pres * Remove bad tests * Removed knn * Fixed issues with removing mc * doc for conda * Added forecast improvements * Added streamlit support * Fixed issues * fix expo with streamlit due to quantile() * fixed performance issues with streamlit for now.. * clean up historical forecast with new trainer * quick fix for regression trainer params * Added fixes * quick fix for other fix for regression trainer params * table formatting for timestamp * potential fix for inf in feature engineered datasets * Basic working in new format * dw * Trying * Fixed issues * Improved graphing * fixing trainer for LR and formatting * doge and linting * page break * automatic cleaning of datasets * automatic cleaning of datasets- fix * Fixed forecast dates * Made dashboard prettier * Added fixes * Added fixes * Added options * Fixed error * remove caching * adding in spinner * Added vairable n_predict in streamlit * Added mypy fix * renaming and range change * new index for n predict * check positive float for window size * Update _index.md * Update _index.md * Update _index.md * Update _index.md * Update _index.md * Update _index.md * Update _index.md * Update _index.md * Update _index.md * renaming * reorg files * Update _index.md * hidden which command for versions * Update _index.md * Update _index.md * which: ns parser * hugo for: which * hugo for: forecasting fix * formatting black * update stock controller test * Lay groundwork for better residual plotting * improved delete to allow for periods in title * improved automatic cleaning of inf * Added new API * Added new API * Added new API * formatting for black * Updated our testing CI * Reverted changes * Added forecast docs * Fixed mypy issues * Fixes tests * Did some refactoring, added a report * new api in streamlit * Added integrated tests * Update _index.md * improved loading in custom dataset * menu spacing * installer fixes * Added docs fixes * Adding comments to test if commit working * Fixed report * naming conventions * formatting * removing unused var * Made last report imporvements * Update README.md * Added fix * Switched to warning * Added fixes * Added fixes * Added fixes * Added fixes * Update economy av view test * Remove forgotten print statement * Update depencencies * Added verbosity to pytest * Added fixes * Fixed pylint * Fixed actions checkout * Added fixes Co-authored-by: colin99d <[email protected]> Co-authored-by: Colin Delahunty <[email protected]> Co-authored-by: James Simmons <[email protected]> Co-authored-by: minhhoang1023 <[email protected]> Co-authored-by: Theodore Aptekarev <[email protected]>
https://github.com/OpenBB-finance/OpenBBTerminal.git
def refresh_datasets_on_menu(self): self.list_dataset_cols = list() maxfile = max(len(file) for file in self.files) self.loaded_dataset_cols = "\n" for dataset, data in self.datasets.items(): self.loaded_dataset_cols += ( f" {dataset} {(maxfile - len(dataset)) * ' '}: " f"{', '.join(data.columns)}\n" ) for col in data.columns: self.list_dataset_cols.append(f"{dataset}.{col}")
72
forecast_controller.py
Python
openbb_terminal/forecast/forecast_controller.py
7fd72d9ee1e8847717195859bf6d608268a94e2f
OpenBBTerminal
4
198,798
31
14
10
127
11
0
37
135
apply_load
changes to just append loads of the same direction along with others made
https://github.com/sympy/sympy.git
def apply_load(self, location, magnitude, direction): magnitude = sympify(magnitude) direction = sympify(direction) if location not in self.node_labels: raise ValueError("Load must be applied at a known node") else: if location in list(self._loads): self._loads[location].append([magnitude, direction]) else: self._loads[location] = [[magnitude, direction]]
80
truss.py
Python
sympy/physics/continuum_mechanics/truss.py
3b6e792c521077d661ae212e7de414696e13ccb6
sympy
3
246,011
6
8
7
29
5
0
6
20
get_room_max_stream_ordering
Remove redundant `get_current_events_token` (#11643) * Push `get_room_{min,max_stream_ordering}` into StreamStore Both implementations of this are identical, so we may as well push it down and get rid of the abstract base class nonsense. * Remove redundant `StreamStore` class This is empty now * Remove redundant `get_current_events_token` This was an exact duplicate of `get_room_max_stream_ordering`, so let's get rid of it. * newsfile
https://github.com/matrix-org/synapse.git
def get_room_max_stream_ordering(self) -> int: return self._stream_id_gen.get_current_token()
16
stream.py
Python
synapse/storage/databases/main/stream.py
2359ee3864a065229c80e3ff58faa981edd24558
synapse
1
144,850
6
9
4
31
4
0
6
20
as_pydict
[Datasets] Expose `TableRow` as public API; minimize copies/type conversions on row-based ops. (#22305) This PR properly exposes `TableRow` as a public API (API docs + the "Public" tag), since it's already exposed to the user in our row-based ops. In addition, the following changes are made: 1. During row-based ops, we also choose a batch format that lines up with the current dataset format in order to eliminate unnecessary copies and type conversions. 2. `TableRow` now derives from `collections.abc.Mapping`, which lets `TableRow` better interop with code expecting a mapping, and includes a few helpful mixins so we only have to implement `__getitem__`, `__iter__`, and `__len__`.
https://github.com/ray-project/ray.git
def as_pydict(self) -> dict: return dict(self.items())
17
row.py
Python
python/ray/data/row.py
53c4c7b1be54fe67a54b34f3f0c076ad502e35cc
ray
1
58,977
29
16
18
183
28
1
36
165
install_protected_system_blocks
Adds ability to delete block types via the CLI (#6849) * Ensure system blocks are protected on destructive API calls * Enable deleting block types * Ensure Block Types are protected against destructive API actions * Ensure updating protected Block Types on update doesn't impact saving * ⚫ * isort * Suppress status errors * ⚫
https://github.com/PrefectHQ/prefect.git
async def install_protected_system_blocks(session): for block in [ prefect.blocks.system.JSON, prefect.blocks.system.DateTime, prefect.blocks.system.Secret, prefect.filesystems.LocalFileSystem, prefect.infrastructure.Process, ]: block_type = block._to_block_type() block_type.is_protected = True block_type = await models.block_types.create_block_type( session=session, block_type=block_type, override=True ) block_schema = await models.block_schemas.create_block_schema( session=session, block_schema=block._to_block_schema(block_type_id=block_type.id), override=True, ) @router.post("/install_system_block_types")
@router.post("/install_system_block_types")
112
block_types.py
Python
src/prefect/orion/api/block_types.py
1a3a3adf0bf4d83206f0367b98905a9db15cfec4
prefect
2
177,078
22
11
6
112
11
0
30
56
radius
Add weight distance metrics (#5305) Adds the weight keyword argument to allow users to compute weighted distance metrics e.g. diameter, eccentricity, periphery, etc. The kwarg works in the same fashion as the weight param for shortest paths - i.e. if a string, look up with edge attr by key, if callable, compute the weight via the function. Default is None, meaning return unweighted result which is the current behavior. Co-authored-by: Dan Schult <[email protected]> Co-authored-by: Ross Barnowski <[email protected]>
https://github.com/networkx/networkx.git
def radius(G, e=None, usebounds=False, weight=None): if usebounds is True and e is None and not G.is_directed(): return _extrema_bounding(G, compute="radius", weight=weight) if e is None: e = eccentricity(G, weight=weight) return min(e.values())
71
distance_measures.py
Python
networkx/algorithms/distance_measures.py
28f78cfa9a386620ee1179582fda1db5ffc59f84
networkx
5
21,767
6
6
3
22
4
0
6
20
is_dotted
Update tomlkit==0.9.2 Used: python -m invoke vendoring.update --package=tomlkit
https://github.com/pypa/pipenv.git
def is_dotted(self) -> bool: return self._dotted
12
items.py
Python
pipenv/vendor/tomlkit/items.py
8faa74cdc9da20cfdcc69f5ec29b91112c95b4c9
pipenv
1
260,074
27
12
11
99
6
0
33
98
_get_prediction_method
API Rename base_estimator in CalibratedClassifierCV (#22054) Co-authored-by: Kevin Roice <[email protected]> Co-authored-by: Guillaume Lemaitre <[email protected]> Co-authored-by: Thomas J. Fan <[email protected]>
https://github.com/scikit-learn/scikit-learn.git
def _get_prediction_method(clf): if hasattr(clf, "decision_function"): method = getattr(clf, "decision_function") return method, "decision_function" elif hasattr(clf, "predict_proba"): method = getattr(clf, "predict_proba") return method, "predict_proba" else: raise RuntimeError( "'estimator' has no 'decision_function' or 'predict_proba' method." )
53
calibration.py
Python
sklearn/calibration.py
effdd6e215c67f2ae8ed1e378ea1661e936059a4
scikit-learn
3
174,928
9
13
11
79
11
1
9
40
generate_metadata
Use data-dist-info-metadata (PEP 658) to decouple resolution from downloading (#11111) Co-authored-by: Tzu-ping Chung <[email protected]>
https://github.com/pypa/pip.git
def generate_metadata(self) -> bytes: return dedent( f ).encode("utf-8") @pytest.fixture(scope="function")
@pytest.fixture(scope="function")
19
test_download.py
Python
tests/functional/test_download.py
bad03ef931d9b3ff4f9e75f35f9c41f45839e2a1
pip
1
176,579
79
16
95
326
30
1
115
310
weighted_projected_graph
Fixes #5403: Errors on non-distinct bipartite node sets (#5442) * is_bipartite_node_set errors if nodes not distinct * weighted_projected_graph errors if nodeset too big * update release_dev.rst * Use sets instead of lists for collecting flowfuncs in tests. (#5589) Co-authored-by: Ross Barnowski <[email protected]>
https://github.com/networkx/networkx.git
def weighted_projected_graph(B, nodes, ratio=False): r if B.is_directed(): pred = B.pred G = nx.DiGraph() else: pred = B.adj G = nx.Graph() G.graph.update(B.graph) G.add_nodes_from((n, B.nodes[n]) for n in nodes) n_top = float(len(B) - len(nodes)) if n_top < 1: raise NetworkXAlgorithmError( f"the size of the nodes to project onto ({len(nodes)}) is >= the graph size ({len(B)}).\n" "They are either not a valid bipartite partition or contain duplicates" ) for u in nodes: unbrs = set(B[u]) nbrs2 = {n for nbr in unbrs for n in B[nbr]} - {u} for v in nbrs2: vnbrs = set(pred[v]) common = unbrs & vnbrs if not ratio: weight = len(common) else: weight = len(common) / n_top G.add_edge(u, v, weight=weight) return G @not_implemented_for("multigraph")
@not_implemented_for("multigraph")
188
projection.py
Python
networkx/algorithms/bipartite/projection.py
bc7ace58c872d527475c09345f89579ff82e4c5d
networkx
9
104,814
43
14
9
136
14
0
56
101
xdirname
Add support for metadata files to `imagefolder` (#4069) * Add support for metadata files to `imagefolder` * Fix imagefolder drop_labels test * Replace csv with jsonl * Add test * Correct resolution for nested metadata files * Allow None as JSON Lines value * Add comments * Count path segments * Address comments * Improve test * Update src/datasets/packaged_modules/imagefolder/imagefolder.py Co-authored-by: Quentin Lhoest <[email protected]> * test e2e imagefolder with metadata * add test for zip archives * fix test * add some debug logging to know which files are ignored * add test for bad/malformed metadata file * revert use of posix path to fix windows tests * style * Refactor tests for packaged modules Text and Csv Co-authored-by: Quentin Lhoest <[email protected]> Co-authored-by: Quentin Lhoest <[email protected]>
https://github.com/huggingface/datasets.git
def xdirname(a): a, *b = str(a).split("::") if is_local_path(a): a = os.path.dirname(Path(a).as_posix()) else: a = posixpath.dirname(a) # if we end up at the root of the protocol, we get for example a = 'http:' # so we have to fix it by adding the '//' that was removed: if a.endswith(":"): a += "//" return "::".join([a] + b)
75
streaming_download_manager.py
Python
src/datasets/utils/streaming_download_manager.py
7017b0965f0a0cae603e7143de242c3425ecef91
datasets
3
247,306
36
11
12
157
9
0
51
164
test_add_alias
Add type hints to `tests/rest/client` (#12108) * Add type hints to `tests/rest/client` * newsfile * fix imports * add `test_account.py` * Remove one type hint in `test_report_event.py` * change `on_create_room` to `async` * update new functions in `test_third_party_rules.py` * Add `test_filter.py` * add `test_rooms.py` * change to `assertEquals` to `assertEqual` * lint
https://github.com/matrix-org/synapse.git
def test_add_alias(self) -> None: # Create an additional alias. second_alias = "#second:test" self._set_alias_via_directory(second_alias) # Add the canonical alias. self._set_canonical_alias({"alias": self.alias, "alt_aliases": [self.alias]}) # Then add the second alias. self._set_canonical_alias( {"alias": self.alias, "alt_aliases": [self.alias, second_alias]} ) # Canonical alias now exists! res = self._get_canonical_alias() self.assertEqual( res, {"alias": self.alias, "alt_aliases": [self.alias, second_alias]} )
90
test_rooms.py
Python
tests/rest/client/test_rooms.py
2ffaf30803f93273a4d8a65c9e6c3110c8433488
synapse
1
275,982
5
6
2
18
4
0
5
19
objects_to_serialize
Reformatting the codebase with black. PiperOrigin-RevId: 450093126
https://github.com/keras-team/keras.git
def objects_to_serialize(self, serialization_cache): raise NotImplementedError
10
base_serialization.py
Python
keras/saving/saved_model/base_serialization.py
84afc5193d38057e2e2badf9c889ea87d80d8fbf
keras
1
224,044
23
13
5
118
14
0
26
44
build_page
Remove spaces at the ends of docstrings, normalize quotes
https://github.com/mkdocs/mkdocs.git
def build_page(title, path, config, md_src=''): files = Files([File(path, config['docs_dir'], config['site_dir'], config['use_directory_urls'])]) page = Page(title, list(files)[0], config) # Fake page.read_source() page.markdown, page.meta = meta.get_data(md_src) return page, files
74
build_tests.py
Python
mkdocs/tests/build_tests.py
e7f07cc82ab2be920ab426ba07456d8b2592714d
mkdocs
1
249,813
45
13
26
138
18
0
58
154
to_synapse_error
Remove redundant types from comments. (#14412) Remove type hints from comments which have been added as Python type hints. This helps avoid drift between comments and reality, as well as removing redundant information. Also adds some missing type hints which were simple to fill in.
https://github.com/matrix-org/synapse.git
def to_synapse_error(self) -> SynapseError: # try to parse the body as json, to get better errcode/msg, but # default to M_UNKNOWN with the HTTP status as the error text try: j = json_decoder.decode(self.response.decode("utf-8")) except ValueError: j = {} if not isinstance(j, dict): j = {} errcode = j.pop("errcode", Codes.UNKNOWN) errmsg = j.pop("error", self.msg) return ProxiedRequestError(self.code, errmsg, errcode, j)
82
errors.py
Python
synapse/api/errors.py
d8cc86eff484b6f570f55a5badb337080c6e4dcd
synapse
3
269,919
9
10
4
45
6
0
10
42
_implements_test_batch_hooks
Reformatting the codebase with black. PiperOrigin-RevId: 450093126
https://github.com/keras-team/keras.git
def _implements_test_batch_hooks(self): return not generic_utils.is_default( self.on_test_batch_begin ) or not generic_utils.is_default(self.on_test_batch_end)
26
callbacks.py
Python
keras/callbacks.py
84afc5193d38057e2e2badf9c889ea87d80d8fbf
keras
2
269,930
11
9
4
51
6
0
11
43
on_predict_begin
Reformatting the codebase with black. PiperOrigin-RevId: 450093126
https://github.com/keras-team/keras.git
def on_predict_begin(self, logs=None): logs = self._process_logs(logs) for callback in self.callbacks: callback.on_predict_begin(logs)
31
callbacks.py
Python
keras/callbacks.py
84afc5193d38057e2e2badf9c889ea87d80d8fbf
keras
2
133,184
10
10
4
52
8
0
12
44
map
[CI] Format Python code with Black (#21975) See #21316 and #21311 for the motivation behind these changes.
https://github.com/ray-project/ray.git
def map(self, func, iterable, chunksize=None): return self._map_async( func, iterable, chunksize=chunksize, unpack_args=False ).get()
35
pool.py
Python
python/ray/util/multiprocessing/pool.py
7f1bacc7dc9caf6d0ec042e39499bbf1d9a7d065
ray
1
129,882
30
14
13
119
14
0
37
159
_get_current_node_resource_key
[CI] Format Python code with Black (#21975) See #21316 and #21311 for the motivation behind these changes.
https://github.com/ray-project/ray.git
def _get_current_node_resource_key(self) -> str: current_node_id = ray.get_runtime_context().node_id.hex() for node in ray.nodes(): if node["NodeID"] == current_node_id: # Found the node. for key in node["Resources"].keys(): if key.startswith("node:"): return key else: raise ValueError("Cannot find the node dictionary for current node.")
67
job_manager.py
Python
dashboard/modules/job/job_manager.py
7f1bacc7dc9caf6d0ec042e39499bbf1d9a7d065
ray
5
256,261
41
9
21
193
20
0
42
257
_get_state_dict
Apply black formatting (#2115) * Testing black on ui/ * Applying black on docstores * Add latest docstring and tutorial changes * Create a single GH action for Black and docs to reduce commit noise to the minimum, slightly refactor the OpenAPI action too * Remove comments * Relax constraints on pydoc-markdown * Split temporary black from the docs. Pydoc-markdown was obsolete and needs a separate PR to upgrade * Fix a couple of bugs * Add a type: ignore that was missing somehow * Give path to black * Apply Black * Apply Black * Relocate a couple of type: ignore * Update documentation * Make Linux CI run after applying Black * Triggering Black * Apply Black * Remove dependency, does not work well * Remove manually double trailing commas * Update documentation Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
https://github.com/deepset-ai/haystack.git
def _get_state_dict(self): state_dict = { "evaluate_every": self.evaluate_every, "n_gpu": self.n_gpu, "grad_acc_steps": self.grad_acc_steps, "device": self.device, "local_rank": self.local_rank, "early_stopping": self.early_stopping, "epochs": self.epochs, "checkpoint_on_sigterm": self.checkpoint_on_sigterm, "checkpoint_root_dir": self.checkpoint_root_dir, "checkpoint_every": self.checkpoint_every, "checkpoints_to_keep": self.checkpoints_to_keep, "from_epoch": self.from_epoch, "from_step": self.from_step, "global_step": self.global_step, "log_learning_rate": self.log_learning_rate, "log_loss_every": self.log_loss_every, "disable_tqdm": self.disable_tqdm, } return state_dict
114
base.py
Python
haystack/modeling/training/base.py
a59bca366174d9c692fa19750c24d65f47660ef7
haystack
1
281,704
10
9
31
54
9
0
10
31
print_help
Custom data context (#1193) * Add first iteration of custom context * Add sample data + improve plot * Change `head` to `show` with sorting and limit. Add "-x" to plot and dynamic update of completer * generate random time series for test csv * Make columns lower case. Check if date is in columns and convert to timestamp. Improve plotting for dates * Add qa to custom * Add pred to custom * Hugooooo * Testing * dang whitespace Co-authored-by: Colin Delahunty <[email protected]> Co-authored-by: didierlopes.eth <[email protected]>
https://github.com/OpenBB-finance/OpenBBTerminal.git
def print_help(self): help_text = f console.print(text=help_text, menu="Custom - Quantitative Analysis")
22
qa_controller.py
Python
gamestonk_terminal/custom/quantitative_analysis/qa_controller.py
6a66f3f3ed934e0615ff4ba283ee67fcc43d3656
OpenBBTerminal
1
136,073
22
8
5
42
5
1
23
48
_arrow_extension_scalars_are_subclassable
[Datasets] [Arrow 7+ Support - 3/N] Add support for Arrow 8, 9, 10, and nightly. (#29999) This PR adds support for Arrow 8, 9, 10, and nightly in Ray, and is the third PR in a set of stacked PRs making up this mono-PR for Arrow 7+ support (#29161), and is stacked on top of a PR fixing task cancellation in Ray Core (#29984) and a PR adding support for Arrow 7 (#29993). The last two commits are the relevant commits for review. Summary of Changes This PR: - For Arrow 9+, add allow_bucket_creation=true to S3 URIs for the Ray Core Storage API and for the Datasets S3 write API ([Datasets] In Arrow 9+, creating S3 buckets requires explicit opt-in. #29815). - For Arrow 9+, create an ExtensionScalar subclass for tensor extension types that returns an ndarray view from .as_py() ([Datasets] For Arrow 8+, tensor column element access returns an ExtensionScalar. #29816). - For Arrow 8.*, we manually convert the ExtensionScalar to an ndarray for tensor extension types, since the ExtensionScalar type exists but isn't subclassable in Arrow 8 ([Datasets] For Arrow 8+, tensor column element access returns an ExtensionScalar. #29816). - For Arrow 10+, we match on other potential error messages when encountering permission issues when interacting with S3 ([Datasets] In Arrow 10+, S3 errors raised due to permission issues can vary beyond our current pattern matching #29994). - adds CI jobs for Arrow 8, 9, 10, and nightly - removes the pyarrow version upper bound
https://github.com/ray-project/ray.git
def _arrow_extension_scalars_are_subclassable(): # TODO(Clark): Remove utility once we only support Arrow 9.0.0+. return ( PYARROW_VERSION is None or PYARROW_VERSION >= MIN_PYARROW_VERSION_SCALAR_SUBCLASS ) @PublicAPI(stability="beta")
@PublicAPI(stability="beta")
15
arrow.py
Python
python/ray/air/util/tensor_extensions/arrow.py
06d5dc36e19f16b7e1f8d6e2e872a81624f6a42c
ray
2
181,661
32
12
15
338
18
0
78
123
test_transform
Revert "Deployed 7ccda9a with MkDocs version: 1.3.0" This reverts commit bd9629c40e01241766197119b581a99409b07068.
https://github.com/EpistasisLab/tpot.git
def test_transform(): input = np.array(((0, 1, 2, 3, 4, 5), (0, 1, 2, 3, 4, 5))).transpose() ohe = OneHotEncoder() ohe.fit(input) test_data = np.array(((0, 1, 2, 6), (0, 1, 6, 7))).transpose() output = ohe.transform(test_data).todense() assert np.sum(output) == 5 input = np.array(((0, 1, 2, 3, 4, 5), (0, 1, 2, 3, 4, 5))).transpose() ips = scipy.sparse.csr_matrix(input) ohe = OneHotEncoder() ohe.fit(ips) test_data = np.array(((0, 1, 2, 6), (0, 1, 6, 7))).transpose() tds = scipy.sparse.csr_matrix(test_data) output = ohe.transform(tds).todense() assert np.sum(output) == 3
233
one_hot_encoder_tests.py
Python
tests/one_hot_encoder_tests.py
388616b6247ca4ea8de4e2f340d6206aee523541
tpot
1
258,797
97
10
19
230
24
0
152
251
test_sample_y_shapes
BUG Fix covariance and stdev shape in GPR with normalize_y (#22199) Co-authored-by: Guillaume Lemaitre <[email protected]> Co-authored-by: Nakamura-Zimmerer, Tenavi (ARC-AF) <[email protected]>
https://github.com/scikit-learn/scikit-learn.git
def test_sample_y_shapes(normalize_y, n_targets): rng = np.random.RandomState(1234) n_features, n_samples_train = 6, 9 # Number of spatial locations to predict at n_samples_X_test = 7 # Number of sample predictions per test point n_samples_y_test = 5 y_train_shape = (n_samples_train,) if n_targets is not None: y_train_shape = y_train_shape + (n_targets,) # By convention single-output data is squeezed upon prediction if n_targets is not None and n_targets > 1: y_test_shape = (n_samples_X_test, n_targets, n_samples_y_test) else: y_test_shape = (n_samples_X_test, n_samples_y_test) X_train = rng.randn(n_samples_train, n_features) X_test = rng.randn(n_samples_X_test, n_features) y_train = rng.randn(*y_train_shape) model = GaussianProcessRegressor(normalize_y=normalize_y) # FIXME: before fitting, the estimator does not have information regarding # the number of targets and default to 1. This is inconsistent with the shape # provided after `fit`. This assert should be made once the following issue # is fixed: # https://github.com/scikit-learn/scikit-learn/issues/22430 # y_samples = model.sample_y(X_test, n_samples=n_samples_y_test) # assert y_samples.shape == y_test_shape model.fit(X_train, y_train) y_samples = model.sample_y(X_test, n_samples=n_samples_y_test) assert y_samples.shape == y_test_shape
142
test_gpr.py
Python
sklearn/gaussian_process/tests/test_gpr.py
3786daf7dc5c301478d489b0756f90d0ac5d010f
scikit-learn
4
135,927
51
15
19
112
7
0
82
399
_check_if_correct_nn_framework_installed
[RLlib] AlgorithmConfigs: Make None a valid value for methods to set properties; Use new `NotProvided` singleton, instead, to indicate no changes wanted on that property. (#30020)
https://github.com/ray-project/ray.git
def _check_if_correct_nn_framework_installed(self, _tf1, _tf, _torch): if self.framework_str in {"tf", "tf2"}: if not (_tf1 or _tf): raise ImportError( ( "TensorFlow was specified as the framework to use (via `config." "framework([tf|tf2])`)! However, no installation was " "found. You can install TensorFlow via `pip install tensorflow`" ) ) elif self.framework_str == "torch": if not _torch: raise ImportError( ( "PyTorch was specified as the framework to use (via `config." "framework('torch')`)! However, no installation was found. You " "can install PyTorch via `pip install torch`." ) )
60
algorithm_config.py
Python
rllib/algorithms/algorithm_config.py
087548031bcf22dd73364b58acb70e61a49f2427
ray
6
103,055
72
17
65
496
33
0
106
357
test_fish_integration
Add fish pipestatus integration tests and changelog entries
https://github.com/kovidgoyal/kitty.git
def test_fish_integration(self): fish_prompt, right_prompt = 'left>', '<right' completions_dir = os.path.join(kitty_base_dir, 'shell-integration', 'fish', 'vendor_completions.d') with self.run_shell( shell='fish', rc=f) as pty: q = fish_prompt + ' ' * (pty.screen.columns - len(fish_prompt) - len(right_prompt)) + right_prompt pty.wait_till(lambda: pty.screen_contents().count(right_prompt) == 1) self.ae(pty.screen_contents(), q) # XDG_DATA_DIRS pty.send_cmd_to_child('set -q XDG_DATA_DIRS; or echo ok') pty.wait_till(lambda: pty.screen_contents().count(right_prompt) == 2) self.ae(str(pty.screen.line(1)), 'ok') # completion and prompt marking pty.send_cmd_to_child('clear') pty.send_cmd_to_child('_test_comp_path') pty.wait_till(lambda: pty.screen_contents().count(right_prompt) == 2) q = '\n'.join(str(pty.screen.line(i)) for i in range(1, pty.screen.cursor.y)) self.ae(q, 'ok') self.ae(pty.last_cmd_output(), q) # resize and redraw (fish_handle_reflow) pty.write_to_child(r'echo $COLUMNS') pty.set_window_size(rows=20, columns=40) q = fish_prompt + 'echo $COLUMNS' + ' ' * (40 - len(fish_prompt) - len(right_prompt) - len('echo $COLUMNS')) + right_prompt pty.process_input_from_child()
744
shell_integration.py
Python
kitty_tests/shell_integration.py
e2f16ff62451eccb4fc901d30042ceca23940188
kitty
3
3,813
11
10
4
57
8
0
13
41
test_jobs_empty
🎉 🎉 Source FB Marketing: performance and reliability fixes (#9805) * Facebook Marketing performance improvement * add comments and little refactoring * fix integration tests with the new config * improve job status handling, limit concurrency to 10 * fix campaign jobs, refactor manager * big refactoring of async jobs, support random order of slices * update source _read_incremental to hook new state logic * fix issues with timeout * remove debugging and clean up, improve retry logic * merge changes from #8234 * fix call super _read_increment * generalize batch execution, add use_batch flag * improve coverage, do some refactoring of spec * update test, remove overrides of source * add split by AdSet * add smaller insights * fix end_date < start_date case * add account_id to PK * add notes * fix new streams * fix reversed incremental stream * update spec.json for SAT * upgrade CDK and bump version Co-authored-by: Dmytro Rezchykov <[email protected]> Co-authored-by: Eugene Kulak <[email protected]>
https://github.com/airbytehq/airbyte.git
def test_jobs_empty(self, api): manager = InsightAsyncJobManager(api=api, jobs=[]) jobs = list(manager.completed_jobs()) assert not jobs
34
test_async_job_manager.py
Python
airbyte-integrations/connectors/source-facebook-marketing/unit_tests/test_async_job_manager.py
a3aae8017a0a40ff2006e2567f71dccb04c997a5
airbyte
1
288,032
8
8
3
35
5
0
8
22
statflag
Refactor apcupsd to use config flow (#64809) * Add Config Flow to APCUPSd integration and remove YAML support. * Hide the binary sensor if user does not select STATFLAG resource. * Add tests for config flows. * Simplify config flow code. * Spell fix. * Fix pylint warnings. * Simplify the code for config flow. * First attempt to implement import flows to suppport legacy YAML configurations. * Remove unnecessary log calls. * Wrap synchronous update call with `hass.async_add_executor_job`. * Import the YAML configurations when sensor platform is set up. * Move the logger call since the variables are not properly set up. * Add codeowner. * Fix name field of manifest.json. * Fix linting issue. * Fix incorrect dependency due to incorrect rebase. * Update codeowner and config flows via hassfest. * Postpone the deprecation warning to 2022.7. * Import future annotations for init file. * Add an newline at the end to make prettier happy. * Update github id. * Add type hints for return types of steps in config flow. * Move the deprecation date for YAML config to 2022.12. * Update according to reviews. * Use async_forward_entry_setups. * Add helper properties to `APCUPSdData` class. * Add device_info for binary sensor. * Simplify config flow. * Remove options flow strings. * update the tests according to the changes. * Add `entity_registry_enabled_default` to entities and use imported CONF_RESOURCES to disable entities instead of skipping them. * Update according to reviews. * Do not use model of the UPS as the title for the integration. Instead, simply use "APCUPSd" as the integration title and let the device info serve as title for each device instead. * Change schema to be a global variable. * Add more comments. * Rewrite the tests for config flows. * Fix enabled_by_default. * Show friendly titles in the integration. * Add import check in `async_setup_platform` to avoid importing in sensor platform setup. * Add import check in `async_setup_platform` to avoid importing in sensor platform setup. * Update comments in test files. * Use parametrize instead of manually iterating different test cases. * Swap the order of the platform constants. * Avoid using broad exceptions. * Set up device info via `_attr_device_info`. * Remove unrelated test in `test_config_flow`. * Use `DeviceInfo` instead of dict to assign to `_attr_device_info`. * Add english translation. * Add `async_create_issue` for deprecated YAML configuration. * Enable UPS status by default since it could show "online, charging, on battery etc" which is meaningful for all users. * Apply suggestions from code review * Apply suggestion * Apply suggestion Co-authored-by: Martin Hjelmare <[email protected]>
https://github.com/home-assistant/core.git
def statflag(self) -> str | None: return self.status.get("STATFLAG")
19
__init__.py
Python
homeassistant/components/apcupsd/__init__.py
52307708c843b947a2d631f2fe7ddaa8bd9a90d7
core
1
276,924
6
7
2
31
4
1
6
11
enable_interactive_logging
Reformatting the codebase with black. PiperOrigin-RevId: 450093126
https://github.com/keras-team/keras.git
def enable_interactive_logging(): INTERACTIVE_LOGGING.enable = True @keras_export("keras.utils.disable_interactive_logging")
@keras_export("keras.utils.disable_interactive_logging")
10
io_utils.py
Python
keras/utils/io_utils.py
84afc5193d38057e2e2badf9c889ea87d80d8fbf
keras
1
269,372
9
7
2
33
4
1
9
15
preprocess_input
Reformatting the codebase with black. PiperOrigin-RevId: 450093126
https://github.com/keras-team/keras.git
def preprocess_input(x, data_format=None): # pylint: disable=unused-argument return x @keras_export("keras.applications.efficientnet_v2.decode_predictions")
@keras_export("keras.applications.efficientnet_v2.decode_predictions")
12
efficientnet_v2.py
Python
keras/applications/efficientnet_v2.py
84afc5193d38057e2e2badf9c889ea87d80d8fbf
keras
1
38,439
48
14
17
248
19
0
76
177
is_config_or_test
Update codeparrot data preprocessing (#16944) * add new preprocessing arguments * add new filters * add new filters to readme * fix config and test count, update function names and docstrings * reformat code * update readme * Update readme * rename config_test filter Co-authored-by: Leandro von Werra <[email protected]> * rename few_assignments filter Co-authored-by: Leandro von Werra <[email protected]> * rename tokenizer in arguments Co-authored-by: Leandro von Werra <[email protected]> * rename functions and add limit_line argument for config_test filter * update threshold for config_test filter Co-authored-by: Leandro von Werra <[email protected]> Co-authored-by: Loubna ben allal <[email protected]>
https://github.com/huggingface/transformers.git
def is_config_or_test(example, scan_width=5, coeff=0.05): keywords = ["unit tests", "test file", "configuration file"] lines = example["content"].splitlines() count_config = 0 count_test = 0 # first test for _, line in zip(range(scan_width), lines): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test nlines = example["content"].count("\n") threshold = int(coeff * nlines) for line in lines: count_config += line.lower().count("config") count_test += line.lower().count("test") if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False}
145
preprocessing.py
Python
examples/research_projects/codeparrot/scripts/preprocessing.py
e730e1256732b5dfeae2bdd427beacc3fbc20e2a
transformers
7
308,879
5
6
2
17
2
0
5
12
_async_set_value
Add device configuration entities to flux_led (#62786) Co-authored-by: Chris Talkington <[email protected]>
https://github.com/home-assistant/core.git
async def _async_set_value(self) -> None:
8
number.py
Python
homeassistant/components/flux_led/number.py
e222e1b6f05b630bef5aed73e307ca5072b6f286
core
1
209,511
83
14
20
202
14
0
132
343
_parse_multi_byte
E275 - Missing whitespace after keyword (#3711) Co-authored-by: Alexander Aring <[email protected]> Co-authored-by: Anmol Sarma <[email protected]> Co-authored-by: antoine.torre <[email protected]> Co-authored-by: Antoine Vacher <[email protected]> Co-authored-by: Arnaud Ebalard <[email protected]> Co-authored-by: atlowl <[email protected]> Co-authored-by: Brian Bienvenu <[email protected]> Co-authored-by: Chris Packham <[email protected]> Co-authored-by: CQ <[email protected]> Co-authored-by: Daniel Collins <[email protected]> Co-authored-by: Federico Maggi <[email protected]> Co-authored-by: Florian Maury <[email protected]> Co-authored-by: _Frky <[email protected]> Co-authored-by: g-mahieux <[email protected]> Co-authored-by: gpotter2 <[email protected]> Co-authored-by: Guillaume Valadon <[email protected]> Co-authored-by: Hao Zheng <[email protected]> Co-authored-by: Haresh Khandelwal <[email protected]> Co-authored-by: Harri Hämäläinen <[email protected]> Co-authored-by: hecke <[email protected]> Co-authored-by: Jan Romann <[email protected]> Co-authored-by: Jan Sebechlebsky <[email protected]> Co-authored-by: jdiog0 <[email protected]> Co-authored-by: jockque <[email protected]> Co-authored-by: Julien Bedel <[email protected]> Co-authored-by: Keith Scott <[email protected]> Co-authored-by: Kfir Gollan <[email protected]> Co-authored-by: Lars Munch <[email protected]> Co-authored-by: ldp77 <[email protected]> Co-authored-by: Leonard Crestez <[email protected]> Co-authored-by: Marcel Patzlaff <[email protected]> Co-authored-by: Martijn Thé <[email protected]> Co-authored-by: Martine Lenders <[email protected]> Co-authored-by: Michael Farrell <[email protected]> Co-authored-by: Michał Mirosław <[email protected]> Co-authored-by: mkaliszan <[email protected]> Co-authored-by: mtury <[email protected]> Co-authored-by: Neale Ranns <[email protected]> Co-authored-by: Octavian Toader <[email protected]> Co-authored-by: Peter Eisenlohr <[email protected]> Co-authored-by: Phil <[email protected]> Co-authored-by: Pierre Lalet <[email protected]> Co-authored-by: Pierre Lorinquer <[email protected]> Co-authored-by: piersoh <[email protected]> Co-authored-by: plorinquer <[email protected]> Co-authored-by: pvinci <[email protected]> Co-authored-by: Rahul Jadhav <[email protected]> Co-authored-by: Robin Jarry <[email protected]> Co-authored-by: romain-perez <[email protected]> Co-authored-by: rperez <rperez@debian> Co-authored-by: Sabrina Dubroca <[email protected]> Co-authored-by: Sebastian Baar <[email protected]> Co-authored-by: sebastien mainand <[email protected]> Co-authored-by: smehner1 <[email protected]> Co-authored-by: speakinghedge <[email protected]> Co-authored-by: Steven Van Acker <[email protected]> Co-authored-by: Thomas Faivre <[email protected]> Co-authored-by: Tran Tien Dat <[email protected]> Co-authored-by: Wael Mahlous <[email protected]> Co-authored-by: waeva <[email protected]> Co-authored-by: Alexander Aring <[email protected]> Co-authored-by: Anmol Sarma <[email protected]> Co-authored-by: antoine.torre <[email protected]> Co-authored-by: Antoine Vacher <[email protected]> Co-authored-by: Arnaud Ebalard <[email protected]> Co-authored-by: atlowl <[email protected]> Co-authored-by: Brian Bienvenu <[email protected]> Co-authored-by: Chris Packham <[email protected]> Co-authored-by: CQ <[email protected]> Co-authored-by: Daniel Collins <[email protected]> Co-authored-by: Federico Maggi <[email protected]> Co-authored-by: Florian Maury <[email protected]> Co-authored-by: _Frky <[email protected]> Co-authored-by: g-mahieux <[email protected]> Co-authored-by: gpotter2 <[email protected]> Co-authored-by: Guillaume Valadon <[email protected]> Co-authored-by: Hao Zheng <[email protected]> Co-authored-by: Haresh Khandelwal <[email protected]> Co-authored-by: Harri Hämäläinen <[email protected]> Co-authored-by: hecke <[email protected]> Co-authored-by: Jan Romann <[email protected]> Co-authored-by: Jan Sebechlebsky <[email protected]> Co-authored-by: jdiog0 <[email protected]> Co-authored-by: jockque <[email protected]> Co-authored-by: Julien Bedel <[email protected]> Co-authored-by: Keith Scott <[email protected]> Co-authored-by: Kfir Gollan <[email protected]> Co-authored-by: Lars Munch <[email protected]> Co-authored-by: ldp77 <[email protected]> Co-authored-by: Leonard Crestez <[email protected]> Co-authored-by: Marcel Patzlaff <[email protected]> Co-authored-by: Martijn Thé <[email protected]> Co-authored-by: Martine Lenders <[email protected]> Co-authored-by: Michael Farrell <[email protected]> Co-authored-by: Michał Mirosław <[email protected]> Co-authored-by: mkaliszan <[email protected]> Co-authored-by: mtury <[email protected]> Co-authored-by: Neale Ranns <[email protected]> Co-authored-by: Octavian Toader <[email protected]> Co-authored-by: Peter Eisenlohr <[email protected]> Co-authored-by: Phil <[email protected]> Co-authored-by: Pierre Lalet <[email protected]> Co-authored-by: Pierre Lorinquer <[email protected]> Co-authored-by: piersoh <[email protected]> Co-authored-by: pvinci <[email protected]> Co-authored-by: Rahul Jadhav <[email protected]> Co-authored-by: Robin Jarry <[email protected]> Co-authored-by: romain-perez <[email protected]> Co-authored-by: rperez <rperez@debian> Co-authored-by: Sabrina Dubroca <[email protected]> Co-authored-by: Sebastian Baar <[email protected]> Co-authored-by: sebastien mainand <[email protected]> Co-authored-by: smehner1 <[email protected]> Co-authored-by: Steven Van Acker <[email protected]> Co-authored-by: Thomas Faivre <[email protected]> Co-authored-by: Tran Tien Dat <[email protected]> Co-authored-by: Wael Mahlous <[email protected]> Co-authored-by: waeva <[email protected]>
https://github.com/secdev/scapy.git
def _parse_multi_byte(self, s): # type: (str) -> int assert len(s) >= 2 tmp_len = len(s) value = 0 i = 1 byte = orb(s[i]) # For CPU sake, stops at an arbitrary large number! max_value = 1 << 64 # As long as the MSG is set, an another byte must be read while byte & 0x80: value += (byte ^ 0x80) << (7 * (i - 1)) if value > max_value: raise error.Scapy_Exception( 'out-of-bound value: the string encodes a value that is too large (>2^{{64}}): {}'.format(value) # noqa: E501 ) i += 1 assert i < tmp_len, 'EINVAL: x: out-of-bound read: the string ends before the AbstractUVarIntField!' # noqa: E501 byte = orb(s[i]) value += byte << (7 * (i - 1)) value += self._max_value assert value >= 0 return value
125
http2.py
Python
scapy/contrib/http2.py
08b1f9d67c8e716fd44036a027bdc90dcb9fcfdf
scapy
3
300,199
47
11
20
255
19
0
99
193
test_service_calls_with_all_entities
Add ws66i core integration (#56094) * Add ws66i core integration * Remove all ws66i translations * Update ws66i unit tests to meet minimum code coverage * Update ws66i based on @bdraco review * General improvements after 2nd PR review * Disable entities if amp shutoff, set default source names, set 30sec polling * Add _attr_ and change async_on_unload * Improve entity generation * Implement coordinator * Made options fields required, retry connection on failed attempts, use ZoneStatus for attributes * Refactor WS66i entity properties, raise HomeAssistantError on restore service if no snapshot * Update to pyws66i v1.1 * Add quality scale of silver to manifest * Update config_flow test
https://github.com/home-assistant/core.git
async def test_service_calls_with_all_entities(hass): _ = await _setup_ws66i_with_options(hass, MockWs66i()) # Changing media player to new state await _call_media_player_service( hass, SERVICE_VOLUME_SET, {"entity_id": ZONE_1_ID, "volume_level": 0.0} ) await _call_media_player_service( hass, SERVICE_SELECT_SOURCE, {"entity_id": ZONE_1_ID, "source": "one"} ) # Saving existing values await _call_ws66i_service(hass, SERVICE_SNAPSHOT, {"entity_id": "all"}) # Changing media player to new state await _call_media_player_service( hass, SERVICE_VOLUME_SET, {"entity_id": ZONE_1_ID, "volume_level": 1.0} ) await _call_media_player_service( hass, SERVICE_SELECT_SOURCE, {"entity_id": ZONE_1_ID, "source": "three"} ) # await coordinator.async_refresh() # await hass.async_block_till_done() # Restoring media player to its previous state await _call_ws66i_service(hass, SERVICE_RESTORE, {"entity_id": "all"}) await hass.async_block_till_done() state = hass.states.get(ZONE_1_ID) assert state.attributes[ATTR_MEDIA_VOLUME_LEVEL] == 0.0 assert state.attributes[ATTR_INPUT_SOURCE] == "one"
151
test_media_player.py
Python
tests/components/ws66i/test_media_player.py
5e737bfe4fbc5a724f5fdf04ea9319c2224cb114
core
1
110,678
75
13
15
206
18
0
96
385
to_jshtml
Deprecate attributes and expire deprecation in animation
https://github.com/matplotlib/matplotlib.git
def to_jshtml(self, fps=None, embed_frames=True, default_mode=None): if fps is None and hasattr(self, '_interval'): # Convert interval in ms to frames per second fps = 1000 / self._interval # If we're not given a default mode, choose one base on the value of # the _repeat attribute if default_mode is None: default_mode = 'loop' if getattr(self, '_repeat', False) else 'once' if not hasattr(self, "_html_representation"): # Can't open a NamedTemporaryFile twice on Windows, so use a # TemporaryDirectory instead. with TemporaryDirectory() as tmpdir: path = Path(tmpdir, "temp.html") writer = HTMLWriter(fps=fps, embed_frames=embed_frames, default_mode=default_mode) self.save(str(path), writer=writer) self._html_representation = path.read_text() return self._html_representation
122
animation.py
Python
lib/matplotlib/animation.py
ec60128011c1009313e0c24b1dfc89313b2d1b59
matplotlib
6
60,417
27
12
7
91
7
0
37
66
CheckForBadCharacters
Balanced joint maximum mean discrepancy for deep transfer learning
https://github.com/jindongwang/transferlearning.git
def CheckForBadCharacters(filename, lines, error): for linenum, line in enumerate(lines): if u'\ufffd' in line: error(filename, linenum, 'readability/utf8', 5, 'Line contains invalid UTF-8 (or Unicode replacement character).') if '\0' in line: error(filename, linenum, 'readability/nul', 5, 'Line contains NUL byte.')
55
cpp_lint.py
Python
code/deep/BJMMD/caffe/scripts/cpp_lint.py
cc4d0564756ca067516f71718a3d135996525909
transferlearning
4
69,222
8
10
13
33
6
0
9
2
get_actual_gle_dict
feat: Asset Capitalization - manual selection of entry type - GLE cleanup with smaller functions - GLE considering periodical inventory - test cases
https://github.com/frappe/erpnext.git
def get_actual_gle_dict(name): return dict( frappe.db.sql( , name, ) )
20
test_asset_capitalization.py
Python
erpnext/assets/doctype/asset_capitalization/test_asset_capitalization.py
58d430fe3ee62e93ad8d16a08bb42156a25b7d41
erpnext
1
106,001
9
9
3
38
7
0
10
31
extract
Fix description of streaming in the docs (#5313) * fix streaming description in stream section * fix audio docs * change docstrings for streaming download manager * Apply suggestions from code review * all urls -> URL(s), correct return types Co-authored-by: Quentin Lhoest <[email protected]> Co-authored-by: Albert Villanova del Moral <[email protected]>
https://github.com/huggingface/datasets.git
def extract(self, url_or_urls): urlpaths = map_nested(self._extract, url_or_urls, map_tuple=True) return urlpaths
24
streaming_download_manager.py
Python
src/datasets/download/streaming_download_manager.py
832780ac40e3bfb0e15336cc9eca894c1da4158f
datasets
1
26,744
25
15
19
111
11
0
39
221
validate_image_file
Save images to product media from external URLs (#9329) * Save images to product media from external URLs
https://github.com/saleor/saleor.git
def validate_image_file(file, field_name, error_class) -> None: if not file: raise ValidationError( { field_name: ValidationError( "File is required.", code=error_class.REQUIRED ) } ) if not is_image_mimetype(file.content_type): raise ValidationError( { field_name: ValidationError( "Invalid file type.", code=error_class.INVALID ) } ) _validate_image_format(file, field_name, error_class)
69
__init__.py
Python
saleor/graphql/core/utils/__init__.py
5763cb1ad22b84e29a270b9b0b365f32074b753d
saleor
3
216,133
88
15
34
319
11
0
122
500
test__compress_ids_multiple_module_functions
handle compressing multiple mod.fun combos under an id
https://github.com/saltstack/salt.git
def test__compress_ids_multiple_module_functions(): # raw data entering the outputter data = { "local": { "cmd_|-try_this_|-echo 'hi'_|-run": { "__id__": "try_this", "__run_num__": 1, "__sls__": "wayne", "changes": {"pid": 32615, "retcode": 0, "stderr": "", "stdout": "hi"}, "comment": 'Command "echo ' "'hi'\" run", "duration": 8.218, "name": "echo 'hi'", "result": True, "start_time": "23:43:25.715842", }, "test_|-try_this_|-asdf_|-nop": { "__id__": "try_this", "__run_num__": 0, "__sls__": "wayne", "changes": {}, "comment": "Success!", "duration": 0.906, "name": "asdf", "result": True, "start_time": "23:43:25.714010", }, } } # check output text for formatting opts = copy.deepcopy(highstate.__opts__) opts["state_compress_ids"] = True with patch("salt.output.highstate.__opts__", opts, create=True): actual_output = highstate.output(data) # if we only cared about the ID/SLS/Result combo, this would be 4 not 2 assert "Succeeded: 2 (changed=1)" in actual_output assert "Failed: 0" in actual_output assert "Total states run: 2" in actual_output
165
test_highstate.py
Python
tests/pytests/unit/output/test_highstate.py
b677222e9e5031afc2e3962247af5d4adfc91cbe
salt
1
241,549
47
14
12
145
14
0
54
181
reset
Reset the total fit-validation batch progress on epoch (#11244)
https://github.com/Lightning-AI/lightning.git
def reset(self) -> None: if self.restarting: self.batch_progress.reset_on_restart() self.scheduler_progress.reset_on_restart() self.batch_loop.optimizer_loop.optim_progress.reset_on_restart() else: self.batch_progress.reset_on_run() self.scheduler_progress.reset_on_run() self.batch_loop.optimizer_loop.optim_progress.reset_on_run() # when the epoch starts, the total val batch progress should be reset as it's supposed to count the batches # seen per epoch, this is useful for tracking when validation is run multiple times per epoch self.val_loop.epoch_loop.batch_progress.total.reset() self._outputs = []
84
training_epoch_loop.py
Python
pytorch_lightning/loops/epoch/training_epoch_loop.py
e9009d60588306b48a2931bf304a493c468c6740
lightning
2
214,475
4
8
2
24
4
0
4
18
__len__
feat: :sparkles: initial implementation of JsonlCorpora and Datasets
https://github.com/flairNLP/flair.git
def __len__(self): return len(self.sentences)
13
sequence_labeling.py
Python
flair/datasets/sequence_labeling.py
4b61c1f4b2e2d67fbda22eb01af78ffa23a15ab7
flair
1
281,141
27
12
15
137
19
0
32
165
test_coin_api_load_df_for_ta
Crypto menu refactor (#1119) * enabled some crypto commands in dd to be called independent of source loaded * support for coin_map_df in all dd functions + load ta and plot chart refactor * updated tests and removed coingecko scrapping where possible * removed ref of command from hugo * updated pycoingecko version * refactoring load * refactored load to fetch prices; pred can run independent of source now * load by default usd on cp/cg and usdt on cb/bin * updated to rich for formatting and updated dependencies * fixed changes requested * update docs * revert discord requirements * removed absolute from calculate change for price * fixing pr issues * fix loading issue when similar coins exist, move coins to home, fill n/a * update docs for coins * adds load to ta and pred menu
https://github.com/OpenBB-finance/OpenBBTerminal.git
def test_coin_api_load_df_for_ta(self, mock_load): with open( "tests/gamestonk_terminal/cryptocurrency/json/test_cryptocurrency_helpers/btc_usd_test_data.json", encoding="utf8", ) as f: sample_return = json.load(f) mock_load.return_value = sample_return mock_return, vs = load_ta_data( coin_map_df=self.coin_map_df, source="cg", currency="usd", days=30, ) self.assertTrue(mock_return.shape == (31, 4)) self.assertTrue(vs == "usd")
81
test_cryptocurrency_helpers.py
Python
tests/gamestonk_terminal/cryptocurrency/test_cryptocurrency_helpers.py
ea964109d654394cc0a5237e6ec5510ba6404097
OpenBBTerminal
1
315,206
18
12
8
100
11
0
21
85
async_step_zeroconf
Add config flow for Bose SoundTouch (#72967) Co-authored-by: Paulus Schoutsen <[email protected]>
https://github.com/home-assistant/core.git
async def async_step_zeroconf(self, discovery_info): self.host = discovery_info.host try: await self._async_get_device_id() except RequestException: return self.async_abort(reason="cannot_connect") self.context["title_placeholders"] = {"name": self.name} return await self.async_step_zeroconf_confirm()
56
config_flow.py
Python
homeassistant/components/soundtouch/config_flow.py
273e9b287f482f5d1f0718ebf09a56c49c66513d
core
2
19,887
11
9
8
37
4
0
11
25
_requirement_name
check point progress on only bringing in pip==22.0.4 (#4966) * vendor in pip==22.0.4 * updating vendor packaging version * update pipdeptree to fix pipenv graph with new version of pip. * Vendoring of pip-shims 0.7.0 * Vendoring of requirementslib 1.6.3 * Update pip index safety restrictions patch for pip==22.0.4 * Update patches * exclude pyptoject.toml from black to see if that helps. * Move this part of the hash collection back to the top (like prior implementation) because it affects the outcome of this test now in pip 22.0.4
https://github.com/pypa/pipenv.git
def _requirement_name(self) -> str: return str(self.req) if self.req else "unknown package"
21
exceptions.py
Python
pipenv/patched/notpip/_internal/exceptions.py
f3166e673fe8d40277b804d35d77dcdb760fc3b3
pipenv
2
288,291
28
11
9
94
11
0
32
71
is_valid_config_entry
Log config_flow errors for waze_travel_time (#79352)
https://github.com/home-assistant/core.git
def is_valid_config_entry(hass, origin, destination, region): origin = find_coordinates(hass, origin) destination = find_coordinates(hass, destination) try: WazeRouteCalculator(origin, destination, region).calc_all_routes_info() except WRCError as error: _LOGGER.error("Error trying to validate entry: %s", error) return False return True
59
helpers.py
Python
homeassistant/components/waze_travel_time/helpers.py
31ddf6cc31aaaf08a09cd6afa9eb830450d1817d
core
2
281,055
10
9
3
45
9
0
10
31
call_dashboards
Refactored Reports (#1123) * First working program * added Voila support * Attempt to fix requirements * Revert depen * Another dependency test * Fixed errors * Small changes * Fixed help * Added port warning * Fixed continued terminal use * Last change, we good fam * Added placeholder so we can have stored * Didier changes * Fixed exit * Fixed issue
https://github.com/OpenBB-finance/OpenBBTerminal.git
def call_dashboards(self, _): from gamestonk_terminal.jupyter.dashboards import dashboards_controller self.queue = dashboards_controller.menu(self.queue)
28
jupyter_controller.py
Python
gamestonk_terminal/jupyter/jupyter_controller.py
eb0dde4603468f64c5d99c57bd272816062e6a23
OpenBBTerminal
1
90,379
97
13
43
560
18
0
165
707
test_order_by
feat(api): Sort sessions by release timestamp [TET-7] (#34699) This PR adds the ability to sort by release timestamp in sessions v2 over metrics Essentially it: Runs a pre-flight query to Release model according to ordering of release timestamp whether ASC or DESC and gets the resulting release groups. It also adds environment filters to the pre flight query if environment filters are sent in the query param of the request Fetches those groups and uses them as a filter in subsequent metrics queries Formats the output so if groups are present in the Release model but not in the metrics dataset, those groups still exist but are nulled out in the output. Otherwise, it shows the results of the groups in the metrics dataset. Orders the output according to resulting groups from preflight query
https://github.com/getsentry/sentry.git
def test_order_by(self): # Step 1: Create 3 releases release1b = self.create_release(version="1B") release1c = self.create_release(version="1C") release1d = self.create_release(version="1D") # Step 2: Create crash free rate for each of those releases # Release 1c -> 66.7% Crash free rate for _ in range(0, 2): self.store_session(make_session(self.project, release=release1c.version)) self.store_session(make_session(self.project, release=release1c.version, status="crashed")) # Release 1b -> 33.3% Crash free rate for _ in range(0, 2): self.store_session( make_session(self.project, release=release1b.version, status="crashed") ) self.store_session(make_session(self.project, release=release1b.version)) # Create Sessions in each of these releases # Release 1d -> 80% Crash free rate for _ in range(0, 4): self.store_session(make_session(self.project, release=release1d.version)) self.store_session(make_session(self.project, release=release1d.version, status="crashed")) # Step 3: Make request response = self.do_request( { "project": self.project.id, # project without users "statsPeriod": "1d", "interval": "1d", "field": ["crash_free_rate(session)"], "groupBy": ["release"], "orderBy": "-release.timestamp", "per_page": 3, } ) # Step 4: Validate Results assert response.data["groups"] == [ { "by": {"release": "1D"}, "totals": {"crash_free_rate(session)": 0.8}, "series": {"crash_free_rate(session)": [0.8]}, }, { "by": {"release": "1C"}, "totals": {"crash_free_rate(session)": 0.6666666666666667}, "series": {"crash_free_rate(session)": [0.6666666666666667]}, }, { "by": {"release": "1B"}, "totals": {"crash_free_rate(session)": 0.33333333333333337}, "series": {"crash_free_rate(session)": [0.33333333333333337]}, }, ]
334
test_organization_sessions.py
Python
tests/snuba/api/endpoints/test_organization_sessions.py
af66ec160dda84b9d417295a5586a530f422511d
sentry
4
283,018
15
11
26
75
11
0
16
44
print_help
Add TA and test coverage for the Forex context (#1532) * Add technical analysis to forex menu * Remove "stock" and "$" from the common ta charts * Add tests for forex context * Update forex custom reset functions * Remove -t arg (ticker) insertion into other_args in forex controller * Remove excluded command from forex ta help string * Update test data Co-authored-by: didierlopes.eth <[email protected]>
https://github.com/OpenBB-finance/OpenBBTerminal.git
def print_help(self): currency_str = f" {self.ticker} (from {self.start.strftime('%Y-%m-%d')})" help_text = f console.print(text=help_text, menu="Forex - Technical Analysis")
26
ta_controller.py
Python
gamestonk_terminal/forex/technical_analysis/ta_controller.py
ede6ec8b1a0dda6c85e81084922c7be80765cd43
OpenBBTerminal
1
195,194
26
9
16
149
11
0
31
195
r2c2_prompt
Patch 8322 (#4709) * add dafetymix teacher * safety_mix teacher * safety_mix teacher pos and neg teachers * add tests for teacher * add license info * improvement * add task list * add task list and lint * add init.py * adding some patch to director * seeker changes * th * 3 * jing * changes * z and r * remove .opts * fix docs * add contrractions * lint Co-authored-by: Dexter Ju <[email protected]> Co-authored-by: Jing Xu <[email protected]>
https://github.com/facebookresearch/ParlAI.git
def r2c2_prompt(self): return { 'sdm': CONST.IS_SEARCH_REQUIRED, 'mdm': CONST.IS_MEMORY_REQUIRED, 'sgm': CONST.GENERATE_QUERY, 'mgm': CONST.GENERATE_MEMORY, 'mkm': CONST.ACCESS_MEMORY, 'ckm': CONST.EXTRACT_ENTITY, 'skm': CONST.GENERATE_KNOWLEDGE, 'mrm': '', 'crm': '', 'srm': '', 'vrm': '', 'grm': '', 'orm': '', }[self.value]
80
module.py
Python
projects/bb3/agents/module.py
b1acb681207559da56a787ba96e16f0e23697d92
ParlAI
1
260,765
15
7
3
33
5
0
15
43
partial_fit
MAINT Add parameter validation to `HashingVectorizer`. (#24181) Co-authored-by: jeremie du boisberranger <[email protected]>
https://github.com/scikit-learn/scikit-learn.git
def partial_fit(self, X, y=None): # TODO: only validate during the first call self._validate_params() return self
19
text.py
Python
sklearn/feature_extraction/text.py
7c835d550c1dcaf44938b1c285db017a773d7dba
scikit-learn
1
43,671
13
10
4
75
11
0
14
30
task_group
Map and Partial DAG authoring interface for Dynamic Task Mapping (#19965) * Make DAGNode a proper Abstract Base Class * Prevent mapping an already mapped Task/TaskGroup Also prevent calls like .partial(...).partial(...). It is uncertain whether these kinds of repeated partial/map calls have utility, so let's disable them entirely for now to simplify implementation. We can always add them if they are proven useful. Co-authored-by: Tzu-ping Chung <[email protected]>
https://github.com/apache/airflow.git
def task_group(python_callable=None, *tg_args, **tg_kwargs): if callable(python_callable): return TaskGroupDecorator(function=python_callable, kwargs=tg_kwargs) return cast("Callable[[T], T]", functools.partial(TaskGroupDecorator, kwargs=tg_kwargs))
47
task_group.py
Python
airflow/decorators/task_group.py
e9226139c2727a4754d734f19ec625c4d23028b3
airflow
2