ast_errors stringlengths 0 3.2k | d_id int64 44 121k | id int64 70 338k | n_whitespaces int64 3 14k | path stringlengths 8 134 | n_words int64 4 4.82k | n_identifiers int64 1 131 | random_cut stringlengths 16 15.8k | commit_message stringlengths 2 15.3k | fun_name stringlengths 1 84 | commit_id stringlengths 40 40 | repo stringlengths 3 28 | file_name stringlengths 5 79 | ast_levels int64 6 31 | nloc int64 1 548 | url stringlengths 31 59 | complexity int64 1 66 | token_counts int64 6 2.13k | n_ast_errors int64 0 28 | vocab_size int64 4 1.11k | n_ast_nodes int64 15 19.2k | language stringclasses 1
value | documentation dict | code stringlengths 101 62.2k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
70,593 | 244,941 | 110 | tests/test_datasets/test_pipelines/utils.py | 23 | 11 | def check_result_same(results, pipeline_results, check_keys):
for key in check_keys:
if results.get(key, None) is None:
continue
if isinstance(results[key], (BitmapMasks, PolygonMasks)):
assert_allclose(pipeline_results[key].to_ndarray(),
resu... | Refactor Autoaugment | check_result_same | 6146a83cb898110ba0170f956903b74741a6ac37 | mmdetection | utils.py | 14 | 9 | https://github.com/open-mmlab/mmdetection.git | 4 | 77 | 0 | 22 | 117 | Python | {
"docstring": "Check whether the ``pipeline_results`` is the same with the predefined\n ``results``.\n\n Args:\n results (dict): Predefined results which should be the standard\n output of the transform pipeline.\n pipeline_results (dict): Results processed by the transform\n ... | def check_result_same(results, pipeline_results, check_keys):
for key in check_keys:
if results.get(key, None) is None:
continue
if isinstance(results[key], (BitmapMasks, PolygonMasks)):
assert_allclose(pipeline_results[key].to_ndarray(),
resu... | |
38,186 | 159,308 | 65 | scripts/release.py | 29 | 11 | def get_rasa_sdk_version() -> Text:
dependencies_filename = "pyproject.toml"
toml_data = toml.load(project_root() / dependencies_filename)
try:
| add changelog for 3.0.6 release (#10771)
* add changelog
* update poetry.lock | get_rasa_sdk_version | 40d5139b3ec136b82e28cdc80d99076b9e6b1e6a | rasa | release.py | 13 | 9 | https://github.com/RasaHQ/rasa.git | 2 | 58 | 0 | 27 | 110 | Python | {
"docstring": "Find out what the referenced version of the Rasa SDK is.",
"language": "en",
"n_whitespaces": 10,
"n_words": 11,
"vocab_size": 10
} | def get_rasa_sdk_version() -> Text:
dependencies_filename = "pyproject.toml"
toml_data = toml.load(project_root() / dependencies_filename)
try:
sdk_version = toml_data["tool"]["poetry"]["dependencies"]["rasa-sdk"]
return sdk_version[1:].strip()
except AttributeError:
raise ... | |
73,365 | 250,287 | 464 | tests/handlers/test_e2e_room_keys.py | 68 | 12 | def test_update_omitted_version(self) -> None:
version = self.get_success(
self.handler.create_version(
self.local_user,
{
"algorithm": "m.megolm_backup.v1",
"auth_data": "first_version_auth_data",
},
... | Add missing type hints to tests.handlers. (#14680)
And do not allow untyped defs in tests.handlers. | test_update_omitted_version | 652d1669c5a103b1c20478770c4aaf18849c09a3 | synapse | test_e2e_room_keys.py | 13 | 33 | https://github.com/matrix-org/synapse.git | 1 | 122 | 0 | 47 | 214 | Python | {
"docstring": "Check that the update succeeds if the version is missing from the body",
"language": "en",
"n_whitespaces": 12,
"n_words": 13,
"vocab_size": 11
} | def test_update_omitted_version(self) -> None:
version = self.get_success(
self.handler.create_version(
self.local_user,
{
"algorithm": "m.megolm_backup.v1",
"auth_data": "first_version_auth_data",
},
... | |
47,773 | 196,273 | 285 | sympy/geometry/plane.py | 104 | 25 | def parameter_value(self, other, u, v=None):
from sympy.geometry.point import Point
if not isinstance(other, GeometryEntity):
other = Point(other, dim=self.ambient_dimension)
if not isinstance(other, Point):
raise ValueError("o | Updated import locations | parameter_value | 498015021131af4dbb07eb110e5badaba8250c7b | sympy | plane.py | 13 | 20 | https://github.com/sympy/sympy.git | 9 | 184 | 0 | 73 | 286 | Python | {
"docstring": "Return the parameter(s) corresponding to the given point.\n\n Examples\n ========\n\n >>> from sympy import pi, Plane\n >>> from sympy.abc import t, u, v\n >>> p = Plane((2, 0, 0), (0, 0, 1), (0, 1, 0))\n\n By default, the parameter value returned defines a po... | def parameter_value(self, other, u, v=None):
from sympy.geometry.point import Point
if not isinstance(other, GeometryEntity):
other = Point(other, dim=self.ambient_dimension)
if not isinstance(other, Point):
raise ValueError("other must be a point")
if ot... | |
@register | 53,089 | 211,415 | 48 | ppdet/modeling/architectures/pose3d_metro.py | 31 | 9 | def orthographic_projection(X, camera):
camera = camera.reshape((-1, 1, 3))
X_trans = X[:, :, :2] + camera[:, :, 1:]
shape = paddle.shape(X_trans)
X_2d = (camera[:, :, 0] * X_trans.reshape((shape[0], -1))).reshape(shape)
return X_2d
| pose3d metro modeling (#6612)
* pose3d metro modeling
* delete extra comments | orthographic_projection | d4e34fe165c09db65fd00113708be1b711ac957c | PaddleDetection | pose3d_metro.py | 14 | 6 | https://github.com/PaddlePaddle/PaddleDetection.git | 1 | 86 | 1 | 25 | 137 | Python | {
"docstring": "Perform orthographic projection of 3D points X using the camera parameters\n Args:\n X: size = [B, N, 3]\n camera: size = [B, 3]\n Returns:\n Projected 2D points -- size = [B, N, 2]\n ",
"language": "en",
"n_whitespaces": 63,
"n_words": 33,
"vocab_size": 24
} | def orthographic_projection(X, camera):
camera = camera.reshape((-1, 1, 3))
X_trans = X[:, :, :2] + camera[:, :, 1:]
shape = paddle.shape(X_trans)
X_2d = (camera[:, :, 0] * X_trans.reshape((shape[0], -1))).reshape(shape)
return X_2d
@register |
30,003 | 133,390 | 53 | python/ray/util/sgd/torch/worker_group.py | 14 | 8 | def _load_state_id(self, state_id):
remote_calls = [
worker.load_state_stream.remote(state_id) for worker in self.remote_workers
]
| [CI] Format Python code with Black (#21975)
See #21316 and #21311 for the motivation behind these changes. | _load_state_id | 7f1bacc7dc9caf6d0ec042e39499bbf1d9a7d065 | ray | worker_group.py | 10 | 5 | https://github.com/ray-project/ray.git | 2 | 28 | 0 | 13 | 45 | Python | {
"docstring": "Loads the object with id `state_id` to all workers.",
"language": "en",
"n_whitespaces": 8,
"n_words": 9,
"vocab_size": 9
} | def _load_state_id(self, state_id):
remote_calls = [
worker.load_state_stream.remote(state_id) for worker in self.remote_workers
]
return remote_calls
| |
48,851 | 198,278 | 788 | sympy/geometry/line.py | 210 | 30 | def __new__(cls, p1, pt=None, angle=None, **kwargs):
p1 = Point(p1, dim=2)
if pt is not None and angle is None:
try:
p2 = Point(pt, dim=2)
except (NotImplementedError, TypeError, ValueError):
raise ValueError(filldedent())
if p1 == p2:
... | Use sympify less | __new__ | 2a1afca9477eb781f16d5d6b63fa37abed7740a3 | sympy | line.py | 23 | 41 | https://github.com/sympy/sympy.git | 17 | 367 | 0 | 113 | 557 | Python | {
"docstring": "\n The 2nd argument was not a valid Point; if\n it was meant to be an angle it should be\n given with keyword \"angle\".",
"language": "en",
"n_whitespaces": 80,
"n_words": 23,
"vocab_size": 20
} | def __new__(cls, p1, pt=None, angle=None, **kwargs):
p1 = Point(p1, dim=2)
if pt is not None and angle is None:
try:
p2 = Point(pt, dim=2)
except (NotImplementedError, TypeError, ValueError):
raise ValueError(filldedent())
if p1 == p2:
... | |
78,575 | 266,772 | 770 | test/lib/ansible_test/_internal/delegation.py | 231 | 76 | def delegate_command(args, host_state, exclude, require): # type: (EnvironmentConfig, HostState, t.List[str], t.List[str]) -> None
con = host_state.controller_profile.get_origin_controller_connection()
working_directory = host_state.controller_profile.get_working_directory()
host_delegation = not isin... | ansible-test - Code cleanup and refactoring. (#77169)
* Remove unnecessary PyCharm ignores.
* Ignore intentional undefined attribute usage.
* Add missing type hints. Fix existing type hints.
* Fix docstrings and comments.
* Use function to register completion handler.
* Pass strings to display functions.
* Fix C... | delegate_command | a06fa496d3f837cca3c437ab6e9858525633d147 | ansible | delegation.py | 17 | 57 | https://github.com/ansible/ansible.git | 16 | 487 | 0 | 154 | 803 | Python | {
"docstring": "Delegate execution based on the provided host state.",
"language": "en",
"n_whitespaces": 7,
"n_words": 8,
"vocab_size": 8
} | def delegate_command(args, host_state, exclude, require): # type: (EnvironmentConfig, HostState, t.List[str], t.List[str]) -> None
con = host_state.controller_profile.get_origin_controller_connection()
working_directory = host_state.controller_profile.get_working_directory()
host_delegation = not isin... | |
36,629 | 156,292 | 132 | dask/dataframe/io/tests/test_parquet.py | 75 | 19 | def test_in_predicate_requires_an_iterable(tmp_path, engine, filter_value):
path = tmp_path / "gh_8720_pandas.parquet"
df = pd.DataFrame(
{"A": [1, 2, 3, 4], "B": [1, 1, 2, 2]},
)
df.to_parquet(path, engine=engine)
with pytest.raises(TypeError, match="Value of 'in' filter"):
dd.... | Check that values for the `in` predicate in `read_parquet` are correct (#8846)
As reported in #8720, the _flatten_filters function required the value
to be hashable. This implicitly required the value to be a tuple,
although lists and sets would also be appropriate since they support the
'in' operation. _flatten_fi... | test_in_predicate_requires_an_iterable | 9000abdd43772a82dcbf7999c5126b571d698d8a | dask | test_parquet.py | 11 | 13 | https://github.com/dask/dask.git | 1 | 137 | 0 | 52 | 218 | Python | {
"docstring": "Regression test for https://github.com/dask/dask/issues/8720",
"language": "en",
"n_whitespaces": 3,
"n_words": 4,
"vocab_size": 4
} | def test_in_predicate_requires_an_iterable(tmp_path, engine, filter_value):
path = tmp_path / "gh_8720_pandas.parquet"
df = pd.DataFrame(
{"A": [1, 2, 3, 4], "B": [1, 1, 2, 2]},
)
df.to_parquet(path, engine=engine)
with pytest.raises(TypeError, match="Value of 'in' filter"):
dd.... | |
78,571 | 266,768 | 65 | test/lib/ansible_test/_internal/config.py | 25 | 10 | def only_targets(self, target_type): # type: (t.Type[THostConfig]) -> t.List[THostConfig]
if not self.targets:
raise Exception('There must be one or more targets.')
assert type_guard(self.targets, targ | ansible-test - Code cleanup and refactoring. (#77169)
* Remove unnecessary PyCharm ignores.
* Ignore intentional undefined attribute usage.
* Add missing type hints. Fix existing type hints.
* Fix docstrings and comments.
* Use function to register completion handler.
* Pass strings to display functions.
* Fix C... | only_targets | a06fa496d3f837cca3c437ab6e9858525633d147 | ansible | config.py | 10 | 5 | https://github.com/ansible/ansible.git | 2 | 44 | 0 | 25 | 72 | Python | {
"docstring": "\n Return a list of target host configurations.\n Requires that there are one or more targets, all the specified type.\n ",
"language": "en",
"n_whitespaces": 41,
"n_words": 19,
"vocab_size": 19
} | def only_targets(self, target_type): # type: (t.Type[THostConfig]) -> t.List[THostConfig]
if not self.targets:
raise Exception('There must be one or more targets.')
assert type_guard(self.targets, target_type)
return t.cast(t.List[THostConfig], self.targets)
| |
22,584 | 107,070 | 75 | lib/matplotlib/widgets.py | 19 | 11 | def _get_animated_artists(self):
return tuple([a for ax_ in self.ax.get_figure().get_axes()
for a in ax_.g | Fix z_order | _get_animated_artists | 334cc617b8ed3b6b4ec6cb64ff16a040ef454149 | matplotlib | widgets.py | 14 | 4 | https://github.com/matplotlib/matplotlib.git | 5 | 48 | 0 | 15 | 78 | Python | {
"docstring": "\n Convenience method to get all animated artists of a figure, except\n those already present in self.artists. 'z_order' is ignored.\n ",
"language": "en",
"n_whitespaces": 41,
"n_words": 19,
"vocab_size": 19
} | def _get_animated_artists(self):
return tuple([a for ax_ in self.ax.get_figure().get_axes()
for a in ax_.get_children()
if a.get_animated() and a not in self.artists])
| |
23,558 | 109,376 | 131 | lib/mpl_toolkits/axisartist/axislines.py | 35 | 15 | def new_gridlines(self, ax):
g | Get rcParams from mpl | new_gridlines | 438d30b227b1fef7e8733578f851e76a8e360f24 | matplotlib | axislines.py | 11 | 9 | https://github.com/matplotlib/matplotlib.git | 1 | 69 | 0 | 31 | 114 | Python | {
"docstring": "\n Create and return a new GridlineCollection instance.\n\n *which* : \"major\" or \"minor\"\n *axis* : \"both\", \"x\" or \"y\"\n\n ",
"language": "en",
"n_whitespaces": 47,
"n_words": 18,
"vocab_size": 16
} | def new_gridlines(self, ax):
gridlines = GridlinesCollection(
None, transform=ax.transData, colors=mpl.rcParams['grid.color'],
linestyles=mpl.rcParams['grid.linestyle'],
linewidths=mpl.rcParams['grid.linewidth'])
ax._set_artist_props(gridlines)
gridli... | |
3,409 | 20,522 | 73 | pipenv/patched/notpip/_vendor/pygments/util.py | 26 | 9 | def duplicates_removed(it, already_seen=()):
lst = []
seen = set()
for i in it:
if i in seen or i in already_seen:
continue
lst.append(i)
seen. | check point progress on only bringing in pip==22.0.4 (#4966)
* vendor in pip==22.0.4
* updating vendor packaging version
* update pipdeptree to fix pipenv graph with new version of pip.
* Vendoring of pip-shims 0.7.0
* Vendoring of requirementslib 1.6.3
* Update pip index safety restrictions patch for p... | duplicates_removed | f3166e673fe8d40277b804d35d77dcdb760fc3b3 | pipenv | util.py | 9 | 14 | https://github.com/pypa/pipenv.git | 4 | 49 | 0 | 19 | 82 | Python | {
"docstring": "\n Returns a list with duplicates removed from the iterable `it`.\n\n Order is preserved.\n ",
"language": "en",
"n_whitespaces": 23,
"n_words": 13,
"vocab_size": 13
} | def duplicates_removed(it, already_seen=()):
lst = []
seen = set()
for i in it:
if i in seen or i in already_seen:
continue
lst.append(i)
seen.add(i)
return lst
| |
38,490 | 160,110 | 40 | numpy/ma/extras.py | 11 | 11 | def masked_all(shape, dtype=float):
a = masked_array(np.empty(shape, dtype),
| DOC: fix data type of parameter shape (#21251)
`np.ma.masked_all` uses `np.empty` under the hood, so the parameter
description for shape in `masked_all` should be updated to match
that of `np.empty`.
Relevant issue: #21203 | masked_all | 119bf865b15747bea815ec3ced10e2bbc1ba8de1 | numpy | extras.py | 13 | 4 | https://github.com/numpy/numpy.git | 1 | 39 | 0 | 10 | 61 | Python | {
"docstring": "\n Empty masked array with all elements masked.\n\n Return an empty masked array of the given shape and dtype, where all the\n data are masked.\n\n Parameters\n ----------\n shape : int or tuple of ints\n Shape of the required MaskedArray, e.g., ``(2, 3)`` or ``2``.\n dtype... | def masked_all(shape, dtype=float):
a = masked_array(np.empty(shape, dtype),
mask=np.ones(shape, make_mask_descr(dtype)))
return a
| |
@not_implemented_for("directed") | 42,034 | 176,676 | 333 | networkx/algorithms/centrality/closeness.py | 125 | 25 | def closeness_centrality(G, u=None, distance=None, wf_improved=True):
r
if G.is_directed():
G = G.reverse() # create a reversed graph view
if distance is not None:
# use Dijkstra's algorithm with specified attribute as edge weight
path_length = functools.partial(
nx.sin... | added example to closeness.py (#5645)
* added example on closeness
* docstring improvement | closeness_centrality | 58b63cb57cd1747c23611ee0b46991a5be2db751 | networkx | closeness.py | 17 | 118 | https://github.com/networkx/networkx.git | 9 | 186 | 1 | 79 | 294 | Python | {
"docstring": "Compute closeness centrality for nodes.\n\n Closeness centrality [1]_ of a node `u` is the reciprocal of the\n average shortest path distance to `u` over all `n-1` reachable nodes.\n\n .. math::\n\n C(u) = \\frac{n - 1}{\\sum_{v=1}^{n-1} d(v, u)},\n\n where `d(v, u)` is the shortest... | def closeness_centrality(G, u=None, distance=None, wf_improved=True):
r
if G.is_directed():
G = G.reverse() # create a reversed graph view
if distance is not None:
# use Dijkstra's algorithm with specified attribute as edge weight
path_length = functools.partial(
nx.sin... |
72,870 | 249,367 | 131 | synapse/storage/databases/main/event_push_actions.py | 38 | 11 | async def _remove_old_push_actions_that_have_rotated(self) -> None:
# We want to clear out anything that is older than a day that *has* already
# been rotated.
rotated_upto_stream_ordering = await self.db_pool.simple_select_one_onecol(
table="event_push_summary_stream_order... | Clarifications for event push action processing. (#13485)
* Clarifies comments.
* Fixes an erroneous comment (about return type) added in #13455
(ec24813220f9d54108924dc04aecd24555277b99).
* Clarifies the name of a variable.
* Simplifies logic of pulling out the latest join for the requesting user. | _remove_old_push_actions_that_have_rotated | 46bd7f4ed9020bbed459c03a11c26d7f7c3093b0 | synapse | event_push_actions.py | 11 | 18 | https://github.com/matrix-org/synapse.git | 3 | 64 | 0 | 34 | 72 | Python | {
"docstring": "Clear out old push actions that have been summarised.",
"language": "en",
"n_whitespaces": 8,
"n_words": 9,
"vocab_size": 9
} | async def _remove_old_push_actions_that_have_rotated(self) -> None:
# We want to clear out anything that is older than a day that *has* already
# been rotated.
rotated_upto_stream_ordering = await self.db_pool.simple_select_one_onecol(
table="event_push_summary_stream_order... | |
43,394 | 181,606 | 25 | tests/export_tests.py | 11 | 4 | def test_indent():
multiline_string =
indented_multiline_string =
| Revert "Deployed 7ccda9a with MkDocs version: 1.3.0"
This reverts commit bd9629c40e01241766197119b581a99409b07068. | test_indent | 388616b6247ca4ea8de4e2f340d6206aee523541 | tpot | export_tests.py | 8 | 10 | https://github.com/EpistasisLab/tpot.git | 1 | 20 | 0 | 9 | 39 | Python | {
"docstring": "Assert that indenting a multiline string by 4 spaces prepends 4 spaces before each new line.test\ntest1\ntest2\ntest3 test\n test1\n test2\n test3",
"language": "en",
"n_whitespaces": 31,
"n_words": 23,
"vocab_size": 18
} | def test_indent():
multiline_string =
indented_multiline_string =
assert indented_multiline_string == _indent(multiline_string, 4)
| |
74,702 | 254,598 | 27 | onnx/tools/update_model_dims.py | 19 | 6 | def update_inputs_outputs_dims(model, input_dims, output_dims): # type: (ModelProto, Dict[Text, List[Any]], Dict[Text, List[Any]]) -> ModelProto
dim_param_set = set() # type: Set[Text]
| six: remove all references (#3926)
* six: remove all references
ONNX has dropped python 2 support and does not need this anymore.
Signed-off-by: Chris Hua <hua.christopher@gmail.com>
* six: use bytes for binary_type
I misread the changes; the correct migration is binary_type -> bytes in python3.
Signed-of... | update_inputs_outputs_dims | 2e70f6769ca9b9d0e859fbbd6854f3abc478897b | onnx | update_model_dims.py | 8 | 19 | https://github.com/onnx/onnx.git | 5 | 139 | 0 | 16 | 28 | Python | {
"docstring": "\n This function updates the dimension sizes of the model's inputs and outputs to the values\n provided in input_dims and output_dims. if the dim value provided is negative, a unique dim_param\n will be set for that dimension.\n\n Example. if we have the following shape for... | def update_inputs_outputs_dims(model, input_dims, output_dims): # type: (ModelProto, Dict[Text, List[Any]], Dict[Text, List[Any]]) -> ModelProto
dim_param_set = set() # type: Set[Text]
| |
25,277 | 114,801 | 118 | mindsdb/integrations/lightwood_handler/lightwood_handler/utils.py | 43 | 10 | def get_aliased_columns(aliased_columns, model_alias, targets, mode=None):
for col in targets:
if mode == 'input':
if str(col.parts[0]) != model_alias and col.alias is not None:
aliased_columns[aliased_columns.index(col.parts[-1])] = str(col.alias)
if mode == 'outpu... | add utils file | get_aliased_columns | 0c2fc2e6f9d32e8b6785890cdfd7a2bf320b4273 | mindsdb | utils.py | 17 | 9 | https://github.com/mindsdb/mindsdb.git | 8 | 109 | 0 | 28 | 173 | Python | {
"docstring": " This method assumes mdb_sql will alert if there are two columns with the same alias ",
"language": "en",
"n_whitespaces": 16,
"n_words": 15,
"vocab_size": 15
} | def get_aliased_columns(aliased_columns, model_alias, targets, mode=None):
for col in targets:
if mode == 'input':
if str(col.parts[0]) != model_alias and col.alias is not None:
aliased_columns[aliased_columns.index(col.parts[-1])] = str(col.alias)
if mode == 'outpu... | |
78,195 | 265,786 | 172 | netbox/utilities/urls.py | 66 | 17 | def get_model_urls(app_label, model_name):
paths = []
# Retrieve reg | #9072: Implement a mechanism for dynamically registering model detail views | get_model_urls | 0d7851ed9de2792ea6d9ed223c315c235290ddd7 | netbox | urls.py | 16 | 14 | https://github.com/netbox-community/netbox.git | 4 | 88 | 0 | 45 | 172 | Python | {
"docstring": "\n Return a list of URL paths for detail views registered to the given model.\n\n Args:\n app_label: App/plugin name\n model_name: Model name\n ",
"language": "en",
"n_whitespaces": 45,
"n_words": 21,
"vocab_size": 20
} | def get_model_urls(app_label, model_name):
paths = []
# Retrieve registered views for this model
try:
views = registry['views'][app_label][model_name]
except KeyError:
# No views have been registered for this model
views = []
for view in views:
# Import the vie... | |
4,186 | 22,110 | 31 | pipenv/patched/pip/_vendor/requests/sessions.py | 11 | 6 | def head(self, url, **kwargs):
r
kwargs.setdefault("allow_redirects", False)
return self.request | Rename notpip to pip. Vendor in pip-22.2.1 and latest requirementslib and vistir. | head | cd5a9683be69c86c8f3adcd13385a9bc5db198ec | pipenv | sessions.py | 8 | 9 | https://github.com/pypa/pipenv.git | 1 | 32 | 0 | 10 | 52 | Python | {
"docstring": "Sends a HEAD request. Returns :class:`Response` object.\n\n :param url: URL for the new :class:`Request` object.\n :param \\*\\*kwargs: Optional arguments that ``request`` takes.\n :rtype: requests.Response\n ",
"language": "en",
"n_whitespaces": 52,
"n_words": 24,
... | def head(self, url, **kwargs):
r
kwargs.setdefault("allow_redirects", False)
return self.request("HEAD", url, **kwargs)
| |
10,774 | 53,293 | 199 | src/prefect/agent.py | 43 | 24 | async def get_and_submit_flow_runs(self) -> List[FlowRun]:
if not self.started:
raise RuntimeError("Agent is not started. Use `async with OrionAgent()...`")
self.logger.debug("Checking for flow runs...")
submittable_runs = await self.client.read_flow_runs(
sort... | Improve `prefect orion start` output | get_and_submit_flow_runs | 9efee44dbee3a326e3e754139f4ea0d721849561 | prefect | agent.py | 12 | 19 | https://github.com/PrefectHQ/prefect.git | 3 | 93 | 0 | 38 | 160 | Python | {
"docstring": "\n Queries for scheduled flow runs and submits them for execution in parallel\n ",
"language": "en",
"n_whitespaces": 27,
"n_words": 12,
"vocab_size": 11
} | async def get_and_submit_flow_runs(self) -> List[FlowRun]:
if not self.started:
raise RuntimeError("Agent is not started. Use `async with OrionAgent()...`")
self.logger.debug("Checking for flow runs...")
submittable_runs = await self.client.read_flow_runs(
sort... | |
56,962 | 223,543 | 84 | python3.10.4/Lib/email/_header_value_parser.py | 35 | 9 | def get_quoted_string(value):
quoted_string = QuotedString()
if value and value[0] in CFWS_LEADER:
token, value = get_cfws(value)
quoted_string.append(token)
token, value = get_bare_quoted_string(value)
quoted_string.append(token)
if value and value[0] in CFWS_LEADER:
to... | add python 3.10.4 for windows | get_quoted_string | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | _header_value_parser.py | 10 | 11 | https://github.com/XX-net/XX-Net.git | 5 | 77 | 0 | 17 | 127 | Python | {
"docstring": "quoted-string = [CFWS] <bare-quoted-string> [CFWS]\n\n 'bare-quoted-string' is an intermediate class defined by this\n parser and not by the RFC grammar. It is the quoted string\n without any attached CFWS.\n ",
"language": "en",
"n_whitespaces": 42,
"n_words": 29,
"vocab_size": 2... | def get_quoted_string(value):
quoted_string = QuotedString()
if value and value[0] in CFWS_LEADER:
token, value = get_cfws(value)
quoted_string.append(token)
token, value = get_bare_quoted_string(value)
quoted_string.append(token)
if value and value[0] in CFWS_LEADER:
to... | |
491 | 3,584 | 96 | airbyte-integrations/bases/source-acceptance-test/source_acceptance_test/tests/test_core.py | 31 | 13 | def test_defined_keyword_exist_in_schema(self, keyword, discovered_catalog):
schemas_errors = []
for stream_name, stream in discovered_catalog.items():
check_result = find_keyword_schema(stream.json_schema, key=keyword)
if check_result:
schemas_errors | SAT: check for not allowed keywords `allOf`, `not` in connectors schema (#9851)
* test_defined_keyword_exist_in_schema added
Signed-off-by: Sergey Chvalyuk <grubberr@gmail.com> | test_defined_keyword_exist_in_schema | c8ee3f834120aa365acaa90b5eb583ac52c476ca | airbyte | test_core.py | 11 | 7 | https://github.com/airbytehq/airbyte.git | 3 | 52 | 0 | 28 | 91 | Python | {
"docstring": "Checking for the presence of not allowed keywords within each json schema",
"language": "en",
"n_whitespaces": 11,
"n_words": 12,
"vocab_size": 12
} | def test_defined_keyword_exist_in_schema(self, keyword, discovered_catalog):
schemas_errors = []
for stream_name, stream in discovered_catalog.items():
check_result = find_keyword_schema(stream.json_schema, key=keyword)
if check_result:
schemas_errors.app... | |
50,851 | 204,719 | 123 | django/core/management/utils.py | 39 | 15 | def normalize_path_patterns(patterns):
patterns = [os.path.normcase(p) for p in patterns]
dir_suffixes = {"%s*" % path_sep for path_sep in {"/", os.sep}}
norm_patterns = []
for pattern in patterns:
for dir_suffix in dir_suffixes:
if pattern.endswith(dir_suffix):
... | Refs #33476 -- Reformatted code with Black. | normalize_path_patterns | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | utils.py | 18 | 12 | https://github.com/django/django.git | 6 | 86 | 0 | 29 | 141 | Python | {
"docstring": "Normalize an iterable of glob style patterns based on OS.",
"language": "en",
"n_whitespaces": 9,
"n_words": 10,
"vocab_size": 10
} | def normalize_path_patterns(patterns):
patterns = [os.path.normcase(p) for p in patterns]
dir_suffixes = {"%s*" % path_sep for path_sep in {"/", os.sep}}
norm_patterns = []
for pattern in patterns:
for dir_suffix in dir_suffixes:
if pattern.endswith(dir_suffix):
... | |
186 | 1,383 | 20 | packages/syft/src/syft/core/node/common/node_service/success_resp_message.py | 6 | 3 | def get_protobuf_schema() -> GeneratedProtocolMessageType:
return ErrorResponseMessage_PB
| ADD new Syft ErrorResponseMessage | get_protobuf_schema | 14892f3e25065f85fcca953eac681f50880c0c48 | PySyft | success_resp_message.py | 6 | 14 | https://github.com/OpenMined/PySyft.git | 1 | 9 | 0 | 6 | 18 | Python | {
"docstring": "Return the type of protobuf object which stores a class of this type\n As a part of serialization and deserialization, we need the ability to\n lookup the protobuf object type directly from the object type. This\n static method allows us to do this.\n Importantly, this meth... | def get_protobuf_schema() -> GeneratedProtocolMessageType:
return ErrorResponseMessage_PB
| |
72,295 | 248,467 | 287 | tests/rest/media/test_media_retention.py | 59 | 17 | def test_remote_media_cache_retention(self) -> None:
# Advance 31 days (in seconds)
self.reactor.advance(31 * 24 * 60 * 60)
# Check that media has been correctly purged.
# Local media should be unaffected.
# Remote media accessed <30 days ago should still exist.
... | Add config options for media retention (#12732) | test_remote_media_cache_retention | 2fc787c341ff540e5880932f116498ec0ed7a2c2 | synapse | test_media_retention.py | 14 | 20 | https://github.com/matrix-org/synapse.git | 1 | 106 | 0 | 47 | 158 | Python | {
"docstring": "\n Tests that entries from the remote media cache that have not been accessed\n recently is purged, while local media is unaffected.\n ",
"language": "en",
"n_whitespaces": 43,
"n_words": 21,
"vocab_size": 18
} | def test_remote_media_cache_retention(self) -> None:
# Advance 31 days (in seconds)
self.reactor.advance(31 * 24 * 60 * 60)
# Check that media has been correctly purged.
# Local media should be unaffected.
# Remote media accessed <30 days ago should still exist.
... | |
52,781 | 209,791 | 232 | scapy/arch/windows/__init__.py | 54 | 11 | def setmodulation(self, modu):
# type: (int) -> bool
# According to https://nmap.org/npcap/guide/npcap-devguide.html#npcap-feature-dot11 # noqa: E501
self._check_npcap_requirement()
_modus = {
0: "dsss",
1: "fhss",
2: "irbaseband",
... | [Hinty] Core typing: windows (#3684)
* Core typing: windows
Co-authored-by: Pierre <pierre@droids-corp.org> | setmodulation | a2b7a28faff1db058dd22ce097a268e0ad5d1d33 | scapy | __init__.py | 10 | 17 | https://github.com/secdev/scapy.git | 2 | 92 | 0 | 50 | 159 | Python | {
"docstring": "Set the interface modulation. It can be:\n - 0: dsss\n - 1: fhss\n - 2: irbaseband\n - 3: ofdm\n - 4: hrdss\n - 5: erp\n - 6: ht\n - 7: vht\n - 8: ihv\n - 9: mimo-ofdm\n - 10: mimo-ofdm\n ... | def setmodulation(self, modu):
# type: (int) -> bool
# According to https://nmap.org/npcap/guide/npcap-devguide.html#npcap-feature-dot11 # noqa: E501
self._check_npcap_requirement()
_modus = {
0: "dsss",
1: "fhss",
2: "irbaseband",
... | |
53,959 | 215,415 | 22 | salt/transport/rabbitmq.py | 8 | 7 | def _on_connection_error(self, connection, exception):
log.error("Failed to connect", exc_info=True)
| Start to add base class defs | _on_connection_error | ab4803984bce4a4de7cc10910e7310c4babf557e | salt | rabbitmq.py | 8 | 2 | https://github.com/saltstack/salt.git | 1 | 20 | 0 | 8 | 34 | Python | {
"docstring": "\n Invoked by pika when connection on connection error\n :param connection:\n :param exception:\n :return:\n ",
"language": "en",
"n_whitespaces": 49,
"n_words": 13,
"vocab_size": 11
} | def _on_connection_error(self, connection, exception):
log.error("Failed to connect", exc_info=True)
| |
40,202 | 168,095 | 38 | pandas/core/indexes/multi.py | 17 | 4 | def is_monotonic_decreasing(self) -> bool:
| DOC: Add numpydoc SS06 validation (#47885) | is_monotonic_decreasing | 62a69beddbedde349891378992c902c0b9341a9f | pandas | multi.py | 9 | 5 | https://github.com/pandas-dev/pandas.git | 1 | 17 | 0 | 15 | 32 | Python | {
"docstring": "\n Return a boolean if the values are equal or decreasing.\n ",
"language": "en",
"n_whitespaces": 25,
"n_words": 10,
"vocab_size": 10
} | def is_monotonic_decreasing(self) -> bool:
# monotonic decreasing if and only if reverse is monotonic increasing
return self[::-1].is_monotonic_increasing
| |
19,669 | 99,591 | 201 | tests/sentry/notifications/test_notifications.py | 70 | 28 | def test_sends_assignment_notification(self):
url = f"/api/0/issues/{self.group.id}/"
with self.tasks():
response = self.client.put(url, format="json", data={"assignedTo": self.user.username})
assert response.status_code == 200, response.content
msg = mail.outbox[0... | fix(notifications): Use `metrics_key` (#34572) | test_sends_assignment_notification | 1730c481f1a8a71446326fa1ff72e10663016385 | sentry | test_notifications.py | 15 | 15 | https://github.com/getsentry/sentry.git | 1 | 114 | 0 | 50 | 230 | Python | {
"docstring": "\n Test that an email AND Slack notification are sent with\n the expected values when an issue is assigned.\n ",
"language": "en",
"n_whitespaces": 40,
"n_words": 18,
"vocab_size": 17
} | def test_sends_assignment_notification(self):
url = f"/api/0/issues/{self.group.id}/"
with self.tasks():
response = self.client.put(url, format="json", data={"assignedTo": self.user.username})
assert response.status_code == 200, response.content
msg = mail.outbox[0... | |
50,164 | 202,903 | 473 | django/db/migrations/questioner.py | 119 | 18 | def _ask_default(self, default=''):
self.prompt_output.write('Please enter the default value as valid Python.')
if default:
self.prompt_output.write(
f"Accept the default '{default}' by pressing 'Enter' or "
f"provide another value."
)
... | Refs #29026 -- Allowed customizing InteractiveMigrationQuestioner's prompt destination.
Previously, the questioner did not obey the value of stdout provided
to the command. | _ask_default | 0ab58c120939093fea90822f376e1866fc714d1f | django | questioner.py | 17 | 30 | https://github.com/django/django.git | 9 | 158 | 0 | 91 | 288 | Python | {
"docstring": "\n Prompt for a default value.\n\n The ``default`` argument allows providing a custom default value (as a\n string) which will be shown to the user and used as the return value\n if the user doesn't provide any other input.\n ",
"language": "en",
"n_whitespaces":... | def _ask_default(self, default=''):
self.prompt_output.write('Please enter the default value as valid Python.')
if default:
self.prompt_output.write(
f"Accept the default '{default}' by pressing 'Enter' or "
f"provide another value."
)
... | |
45,596 | 186,689 | 37 | certbot-apache/certbot_apache/_internal/parser.py | 9 | 5 | def reset_modules(self) -> None:
self.modules = {}
self.update_modules | Add typing to certbot.apache (#9071)
* Add typing to certbot.apache
Co-authored-by: Adrien Ferrand <ferrand.ad@gmail.com> | reset_modules | 7d9e9a49005de7961e84d2a7c608db57dbab3046 | certbot | parser.py | 7 | 6 | https://github.com/certbot/certbot.git | 1 | 24 | 0 | 9 | 44 | Python | {
"docstring": "Reset the loaded modules list. This is called from cleanup to clear\n temporarily loaded modules.",
"language": "en",
"n_whitespaces": 21,
"n_words": 15,
"vocab_size": 14
} | def reset_modules(self) -> None:
self.modules = {}
self.update_modules()
self.parse_modules()
| |
30,112 | 133,800 | 259 | rllib/agents/ppo/tests/test_ddppo.py | 56 | 28 | def test_ddppo_compilation(self):
| [CI] Format Python code with Black (#21975)
See #21316 and #21311 for the motivation behind these changes. | test_ddppo_compilation | 7f1bacc7dc9caf6d0ec042e39499bbf1d9a7d065 | ray | test_ddppo.py | 15 | 15 | https://github.com/ray-project/ray.git | 4 | 121 | 0 | 45 | 202 | Python | {
"docstring": "Test whether a DDPPOTrainer can be built with both frameworks.",
"language": "en",
"n_whitespaces": 9,
"n_words": 10,
"vocab_size": 10
} | def test_ddppo_compilation(self):
config = ppo.ddppo.DEFAULT_CONFIG.copy()
config["num_gpus_per_worker"] = 0
num_iterations = 2
for _ in framework_iterator(config, frameworks="torch"):
trainer = ppo.ddppo.DDPPOTrainer(config=config, env="CartPole-v0")
fo... | |
@keras_export('keras.metrics.categorical_crossentropy',
'keras.losses.categorical_crossentropy')
@tf.__internal__.dispatch.add_dispatch_support | 79,731 | 268,862 | 56 | keras/losses.py | 30 | 19 | def log_cosh(y_true, y_pred):
y_pred = tf.convert_to_tensor(y_pred)
y_true = tf.cast(y_true, y_pred.dtype)
def _logcosh(x):
return x + tf.math.softplus(- | Fix keras docstrings
PiperOrigin-RevId: 424275818 | log_cosh | 89879e2c76e86c685e44c47a6cdb82f7e645c142 | keras | losses.py | 13 | 5 | https://github.com/keras-team/keras.git | 1 | 47 | 1 | 26 | 157 | Python | {
"docstring": "Logarithm of the hyperbolic cosine of the prediction error.\n\n `log(cosh(x))` is approximately equal to `(x ** 2) / 2` for small `x` and\n to `abs(x) - log(2)` for large `x`. This means that 'logcosh' works mostly\n like the mean squared error, but will not be so strongly affected by the\n occasi... | def log_cosh(y_true, y_pred):
y_pred = tf.convert_to_tensor(y_pred)
y_true = tf.cast(y_true, y_pred.dtype)
def _logcosh(x):
return x + tf.math.softplus(-2. * x) - tf.cast(
tf.math.log(2.), x.dtype)
return backend.mean(_logcosh(y_pred - y_true), axis=-1)
@keras_export('keras.metrics.categorica... |
80,821 | 271,597 | 38 | keras/engine/training.py | 13 | 4 | def make_predict_function(self, force=False):
if self.predict_function is not None and not force:
return self.predict_function
| Reformatting the codebase with black.
PiperOrigin-RevId: 450093126 | make_predict_function | 84afc5193d38057e2e2badf9c889ea87d80d8fbf | keras | training.py | 8 | 17 | https://github.com/keras-team/keras.git | 6 | 83 | 0 | 11 | 41 | Python | {
"docstring": "Creates a function that executes one step of inference.\n\n This method can be overridden to support custom inference logic.\n This method is called by `Model.predict` and `Model.predict_on_batch`.\n\n Typically, this method directly controls `tf.function` and\n `tf.distrib... | def make_predict_function(self, force=False):
if self.predict_function is not None and not force:
return self.predict_function
| |
76,436 | 260,715 | 31 | sklearn/feature_selection/_rfe.py | 10 | 7 | def fit(self, X, y, **fit_params):
| MAINT Add parameter validation to RFE and RFECV. (#24137)
Co-authored-by: Thomas J. Fan <thomasjpfan@gmail.com> | fit | c9d4e1f86e6d8c58441b1aa01d0a79f25cf3a999 | scikit-learn | _rfe.py | 8 | 3 | https://github.com/scikit-learn/scikit-learn.git | 1 | 30 | 0 | 9 | 47 | Python | {
"docstring": "Fit the RFE model and then the underlying estimator on the selected features.\n\n Parameters\n ----------\n X : {array-like, sparse matrix} of shape (n_samples, n_features)\n The training input samples.\n\n y : array-like of shape (n_samples,)\n The ta... | def fit(self, X, y, **fit_params):
self._validate_params()
return self._fit(X, y, **fit_params)
| |
11,321 | 55,468 | 52 | tests/cli/test_storage.py | 31 | 8 | def test_invalid_number_selection_fails():
number_string = "99999999"
result = get_first_menu_and_fail(number_string)
lines = result.stdout.splitlines()
# Strange string addition are due to coloring, I believe
assert lines[-1] == f"\x1b[31mInvalid selection {number_string}\x1b[0m"
assert re... | basic tests for storage cli | test_invalid_number_selection_fails | 21b8eed6887646c8c2a752961a84c855dd4fed22 | prefect | test_storage.py | 9 | 6 | https://github.com/PrefectHQ/prefect.git | 1 | 38 | 0 | 27 | 74 | Python | {
"docstring": "\n We need to make sure that if we give an invalid number that the CLI\n will exit.\n ",
"language": "en",
"n_whitespaces": 27,
"n_words": 17,
"vocab_size": 16
} | def test_invalid_number_selection_fails():
number_string = "99999999"
result = get_first_menu_and_fail(number_string)
lines = result.stdout.splitlines()
# Strange string addition are due to coloring, I believe
assert lines[-1] == f"\x1b[31mInvalid selection {number_string}\x1b[0m"
assert re... | |
48,569 | 197,470 | 16 | sympy/utilities/misc.py | 10 | 9 | def filldedent(s, w=70, **kwargs):
return '\n' + fill(de | Pass keyword arguments to filldedent() through to fill() | filldedent | 2047f4855577845b1b99e0926b887d313725a6e7 | sympy | misc.py | 14 | 2 | https://github.com/sympy/sympy.git | 1 | 38 | 0 | 10 | 66 | Python | {
"docstring": "\n Strips leading and trailing empty lines from a copy of ``s``, then dedents,\n fills and returns it.\n\n Empty line stripping serves to deal with docstrings like this one that\n start with a newline after the initial triple quote, inserting an empty\n line at the beginning of the stri... | def filldedent(s, w=70, **kwargs):
return '\n' + fill(dedent(str(s)).strip('\n'), width=w, **kwargs)
| |
38,732 | 160,788 | 1,357 | numpy/lib/arraysetops.py | 528 | 56 | def in1d(ar1, ar2, assume_unique=False, invert=False, kind=None):
# Ravel both arrays, behavior for the first array could be different
ar1 = np.asarray(ar1).ravel()
ar2 = np.asarray(ar2).ravel()
# Ensure that iteration through object arrays yields size-1 arrays
if ar2.dtype == object:
... | MAINT: change kind names for in1d
- Switch dictionary->table, mergesort->sort | in1d | 4ed458f16d9dd64554ccf49e315c5b8fb577d4cd | numpy | arraysetops.py | 16 | 80 | https://github.com/numpy/numpy.git | 24 | 598 | 0 | 272 | 978 | Python | {
"docstring": "\n Test whether each element of a 1-D array is also present in a second array.\n\n Returns a boolean array the same length as `ar1` that is True\n where an element of `ar1` is in `ar2` and False otherwise.\n\n We recommend using :func:`isin` instead of `in1d` for new code.\n\n Parameter... | def in1d(ar1, ar2, assume_unique=False, invert=False, kind=None):
# Ravel both arrays, behavior for the first array could be different
ar1 = np.asarray(ar1).ravel()
ar2 = np.asarray(ar2).ravel()
# Ensure that iteration through object arrays yields size-1 arrays
if ar2.dtype == object:
... | |
71,096 | 246,202 | 124 | tests/rest/admin/test_user.py | 50 | 14 | def test_devices(self) -> None:
# Login in as the user
self._get_token()
# Check that we don't see a new device in our devices list
channel = self.make_request(
"GET", "devices", b"{}", access_token=self.other_user_tok
)
self.assertEqual(HTTPStatus.O... | Add type hints to `tests/rest/admin` (#11851) | test_devices | 901b264c0c88f39cbfb8b2229e0dc57968882658 | synapse | test_user.py | 11 | 8 | https://github.com/matrix-org/synapse.git | 1 | 66 | 0 | 42 | 111 | Python | {
"docstring": "Tests that logging in as a user doesn't create a new device for them.",
"language": "en",
"n_whitespaces": 13,
"n_words": 14,
"vocab_size": 13
} | def test_devices(self) -> None:
# Login in as the user
self._get_token()
# Check that we don't see a new device in our devices list
channel = self.make_request(
"GET", "devices", b"{}", access_token=self.other_user_tok
)
self.assertEqual(HTTPStatus.O... | |
2,845 | 14,346 | 51 | pydantic/_internal/_validation_functions.py | 17 | 4 | def in_ipython() -> bool:
try:
eval('__IPYTHON__')
except NameError:
| Switching to `pydantic_core` (#4516)
* working on core schema generation
* adapting main.py
* getting tests to run
* fix tests
* disable pyright, fix mypy
* moving to class-based model generation
* working on validators
* change how models are created
* start fixing test_main.py
* fixing myp... | in_ipython | 594effa279668bd955e98f1cd5c036b37d3bbd40 | pydantic | _validation_functions.py | 10 | 10 | https://github.com/pydantic/pydantic.git | 2 | 22 | 0 | 16 | 44 | Python | {
"docstring": "\n Check whether we're in an ipython environment, including jupyter notebooks.\n ",
"language": "en",
"n_whitespaces": 17,
"n_words": 10,
"vocab_size": 10
} | def in_ipython() -> bool:
try:
eval('__IPYTHON__')
except NameError:
return False
else: # pragma: no cover
return True
| |
76,072 | 260,097 | 51 | sklearn/utils/tests/test_param_validation.py | 28 | 8 | def test_stroptions_deprecated_internal_subset():
| MNT Param validation: do not expose internal values in error msg (#23459)
* allow to not expose internal valid params in error msg
* ensure deprecated and internal do not overlap
* deprecated and internal must be subsets of options
* black | test_stroptions_deprecated_internal_subset | 122876e9ab1ab494b4bf0ca3360d5a1527caf2e7 | scikit-learn | test_param_validation.py | 12 | 5 | https://github.com/scikit-learn/scikit-learn.git | 1 | 65 | 0 | 17 | 124 | Python | {
"docstring": "Check that the deprecated and internal parameters must be subsets of options.",
"language": "en",
"n_whitespaces": 11,
"n_words": 12,
"vocab_size": 12
} | def test_stroptions_deprecated_internal_subset():
with pytest.raises(ValueError, match="deprecated options must be a subset"):
StrOptions({"a", "b", "c"}, deprecated={"a", "d"})
with pytest.raises(ValueError, match="internal options must be a subset"):
StrOptions({"a", "b", "c"}, internal=... | |
52,110 | 207,805 | 114 | tests/admin_views/tests.py | 29 | 15 | def test_change_view_without_object_change_permission(self):
change_url = reverse("admin9:admin_views_article_change", args=(self.a1.pk,))
self.client.force_login(self.viewuser)
response = self.client.get(change_url)
self.assertEqual(response.context["title"], "View article")
... | Refs #33476 -- Reformatted code with Black. | test_change_view_without_object_change_permission | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | tests.py | 12 | 11 | https://github.com/django/django.git | 1 | 81 | 0 | 27 | 138 | Python | {
"docstring": "\n The object should be read-only if the user has permission to view it\n and change objects of that type but not to change the current object.\n ",
"language": "en",
"n_whitespaces": 48,
"n_words": 26,
"vocab_size": 23
} | def test_change_view_without_object_change_permission(self):
change_url = reverse("admin9:admin_views_article_change", args=(self.a1.pk,))
self.client.force_login(self.viewuser)
response = self.client.get(change_url)
self.assertEqual(response.context["title"], "View article")
... | |
39,358 | 163,015 | 1,142 | pandas/__init__.py | 355 | 25 | def __getattr__(name):
import warnings
if name in __deprecated_num_index_names:
warnings.warn(
f"pandas.{name} is deprecated "
"and will be removed from pandas in a future version. "
"Use pandas.Index with the appropriate dtype instead.",
FutureWarning,
... | API: hide NumericIndex from public top-level namespace in favor of pd.Index (#44819) | __getattr__ | abd74363b3f911cbba79cfd4d78fb17521b39928 | pandas | __init__.py | 13 | 55 | https://github.com/pandas-dev/pandas.git | 6 | 187 | 0 | 229 | 793 | Python | {
"docstring": "\npandas - a powerful data analysis and manipulation library for Python\n=====================================================================\n\n**pandas** is a Python package providing fast, flexible, and expressive data\nstructures designed to make working with \"relational\" or \"labeled\" data bo... | def __getattr__(name):
import warnings
if name in __deprecated_num_index_names:
warnings.warn(
f"pandas.{name} is deprecated "
"and will be removed from pandas in a future version. "
"Use pandas.Index with the appropriate dtype instead.",
FutureWarning,
... | |
54,169 | 215,777 | 79 | salt/modules/file.py | 33 | 11 | def readlink(path, canonicalize=False):
path = os.path.expanduser(path)
if not os.path.isabs(path):
raise SaltInvocationError("Path to link must be absolute.")
if not os.path.islink(path):
raise SaltInvocationError("A valid link was not specified | Use salt.utils.path.readlink to handle junctions on Windows | readlink | 6680407756aac9ee0eaf150f3a69622b658f7321 | salt | file.py | 12 | 10 | https://github.com/saltstack/salt.git | 4 | 77 | 0 | 26 | 130 | Python | {
"docstring": "\n .. versionadded:: 2014.1.0\n\n Return the path that a symlink points to\n If canonicalize is set to True, then it return the final target\n\n CLI Example:\n\n .. code-block:: bash\n\n salt '*' file.readlink /path/to/link\n ",
"language": "en",
"n_whitespaces": 58,
"n_... | def readlink(path, canonicalize=False):
path = os.path.expanduser(path)
if not os.path.isabs(path):
raise SaltInvocationError("Path to link must be absolute.")
if not os.path.islink(path):
raise SaltInvocationError("A valid link was not specified.")
if canonicalize:
retur... | |
@contextlib.contextmanager | 48,128 | 196,714 | 18 | sympy/testing/pytest.py | 6 | 5 | def warns_deprecated_sympy():
with warns(SymPyDeprecationWarning):
yield
@contextlib.contextmanager | Note that warns_deprecated_sympy should not be used outside of the test suite | warns_deprecated_sympy | fdaacf84861404f7857081bc85850a781a78979f | sympy | pytest.py | 9 | 3 | https://github.com/sympy/sympy.git | 1 | 12 | 1 | 6 | 35 | Python | {
"docstring": "\n Shorthand for ``warns(SymPyDeprecationWarning)``\n\n This is the recommended way to test that ``SymPyDeprecationWarning`` is\n emitted for deprecated features in SymPy. To test for other warnings use\n ``warns``. To suppress warnings without asserting that they are emitted\n use ``ig... | def warns_deprecated_sympy():
with warns(SymPyDeprecationWarning):
yield
@contextlib.contextmanager |
38,751 | 160,832 | 83 | numpy/testing/_private/utils.py | 25 | 11 | def assert_array_equal(x, y, err_msg='', verbose=True, *, strict=False):
__tracebackhide__ = | ENH: Add strict parameter to assert_array_equal. (#21595)
Fixes #9542
Co-authored-by: Bas van Beek <43369155+BvB93@users.noreply.github.com> | assert_array_equal | cafec60a5e28af98fb8798049edd7942720d2d74 | numpy | utils.py | 9 | 5 | https://github.com/numpy/numpy.git | 1 | 51 | 0 | 24 | 77 | Python | {
"docstring": "\n Raises an AssertionError if two array_like objects are not equal.\n\n Given two array_like objects, check that the shape is equal and all\n elements of these objects are equal (but see the Notes for the special\n handling of a scalar). An exception is raised at shape mismatch or\n co... | def assert_array_equal(x, y, err_msg='', verbose=True, *, strict=False):
__tracebackhide__ = True # Hide traceback for py.test
assert_array_compare(operator.__eq__, x, y, err_msg=err_msg,
verbose=verbose, header='Arrays are not equal',
strict=strict)
| |
24,328 | 110,849 | 64 | lib/matplotlib/transforms.py | 11 | 11 | def update_from_data_x(self, x, ignore=None):
x = np.ravel(x)
self.update | [Doc] Fix ndarray-links for arguments | update_from_data_x | f16da868d016363c4cd734b2abd6535230b094df | matplotlib | transforms.py | 13 | 4 | https://github.com/matplotlib/matplotlib.git | 1 | 50 | 0 | 11 | 77 | Python | {
"docstring": "\n Update the x-bounds of the `Bbox` based on the passed in data. After\n updating, the bounds will have positive *width*, and *x0* will be the\n minimal value.\n\n Parameters\n ----------\n x : `~numpy.ndarray`\n Array of x-values.\n\n ignor... | def update_from_data_x(self, x, ignore=None):
x = np.ravel(x)
self.update_from_data_xy(np.column_stack([x, np.ones(x.size)]),
ignore=ignore, updatey=False)
| |
50,119 | 202,417 | 477 | tests/csrf_tests/tests.py | 157 | 15 | def test_https_malformed_referer(self):
malformed_referer_msg = "Referer checking failed - Referer is malformed."
req = self._get_POST_request_with_token()
req._is_secure_override = True
req.META["HTTP_REFERER"] = "http://http://www.example.com/"
mw = CsrfViewMiddleware(... | Refs #33476 -- Reformatted code with Black. | test_https_malformed_referer | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | tests.py | 9 | 33 | https://github.com/django/django.git | 1 | 292 | 0 | 65 | 472 | Python | {
"docstring": "\n A POST HTTPS request with a bad referer is rejected.\n ",
"language": "en",
"n_whitespaces": 25,
"n_words": 10,
"vocab_size": 10
} | def test_https_malformed_referer(self):
malformed_referer_msg = "Referer checking failed - Referer is malformed."
req = self._get_POST_request_with_token()
req._is_secure_override = True
req.META["HTTP_REFERER"] = "http://http://www.example.com/"
mw = CsrfViewMiddleware(... | |
21,477 | 102,162 | 32 | tools/test/test_gen_backend_stubs.py | 11 | 6 | def test_missing_cpp_namespace(self) -> None:
y | Revert "Revert D32498569: allow external backend codegen to toggle whether to generate out= and inplace kernels" (#69950)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/69950
This reverts commit f6cad53443704dfe5a20cc62bee14d91e3bffcaa.
Test Plan: Imported from OSS
Reviewed By: albanD
Diff... | test_missing_cpp_namespace | bb5b4cceb6f737448eaaa6817cd773b6f4b0e77d | pytorch | test_gen_backend_stubs.py | 8 | 7 | https://github.com/pytorch/pytorch.git | 1 | 26 | 0 | 10 | 47 | Python | {
"docstring": "\\\nbackend: XLA\nsupported:\n- absYou must provide a value for \"cpp_namespace\"",
"language": "en",
"n_whitespaces": 8,
"n_words": 12,
"vocab_size": 12
} | def test_missing_cpp_namespace(self) -> None:
yaml_str =
output_error = self.get_errors_from_gen_backend_stubs(yaml_str)
self.assertExpectedInline(output_error, )
| |
42,410 | 177,506 | 16 | networkx/algorithms/shortest_paths/weighted.py | 10 | 6 | def single_source_dijkstra_path_length(G, source, cutoff=None, weight="weight"):
return multi_source_dijkstra_path_length(G, {source}, cutoff=cutoff, weight=weight)
| Hide edges with a weight of None in A*. (#5945)
* Hide edges with a weight of None in A*.
This matches the Dijkstra's weight interface.
* Update Dijkstra's and A* docs for weights of None.
* Add tests for A* with weight of None.
* Add another test for A* with a weight function.
* Document that None indi... | single_source_dijkstra_path_length | d82815dba6c8ddce19cd49f700298dc82a58f066 | networkx | weighted.py | 8 | 2 | https://github.com/networkx/networkx.git | 1 | 33 | 0 | 10 | 50 | Python | {
"docstring": "Find shortest weighted path lengths in G from a source node.\n\n Compute the shortest path length between source and all other\n reachable nodes for a weighted graph.\n\n Parameters\n ----------\n G : NetworkX graph\n\n source : node label\n Starting node for path\n\n cutof... | def single_source_dijkstra_path_length(G, source, cutoff=None, weight="weight"):
return multi_source_dijkstra_path_length(G, {source}, cutoff=cutoff, weight=weight)
| |
83,834 | 281,536 | 51 | gamestonk_terminal/stocks/behavioural_analysis/ba_controller.py | 24 | 11 | def print_help(self):
has_ticker_start = "" if self.ticker else "[unvl]"
has_ticker_end = "" if self.ticker else "[/unvl]"
help_text = f
c | Terminal Wide Rich (#1161)
* My idea for how we handle Rich moving forward
* remove independent consoles
* FIxed pylint issues
* add a few vars
* Switched print to console
* More transitions
* Changed more prints
* Replaced all prints
* Fixing tabulate
* Finished replace tabulate
* Finish... | print_help | 82747072c511beb1b2672846ae2ee4aec53eb562 | OpenBBTerminal | ba_controller.py | 12 | 40 | https://github.com/OpenBB-finance/OpenBBTerminal.git | 3 | 39 | 0 | 18 | 119 | Python | {
"docstring": "[cmds]\n load load a specific stock ticker for analysis\n\n[param]Ticker: [/param]{self.ticker.upper() or None}\n{has_ticker_start}\n[src][Finbrain][/src]\n headlines sentiment from 15+ major news headlines\n[src][Finnhub][/src]\n stats sentiment stats including comparis... | def print_help(self):
has_ticker_start = "" if self.ticker else "[unvl]"
has_ticker_end = "" if self.ticker else "[/unvl]"
help_text = f
console.print(text=help_text, menu="Stocks - Behavioural Analysis")
| |
77,511 | 263,916 | 117 | PyInstaller/lib/modulegraph/util.py | 75 | 9 | def iterate_instructions(code_object):
# The arg extension the EXTENDED_ARG opcode represents is automatically handled by get_instructions() but the
# instruction is left in. Get rid of it to make subsequent parsing easie | Make modulegraph EXTENDED_ARG opcode aware (#7055).
Fix :class:`AssertionError` during build when analysing a ``.pyc`` file
containing more that 255 variable names followed by an import statement all in
the same namespace.
Fixes #7055. | iterate_instructions | 299ff39d1260cf80241b81ed39bbb89b78709583 | pyinstaller | util.py | 12 | 6 | https://github.com/pyinstaller/pyinstaller.git | 5 | 47 | 0 | 56 | 83 | Python | {
"docstring": "Delivers the byte-code instructions as a continuous stream.\n\n Yields `dis.Instruction`. After each code-block (`co_code`), `None` is\n yielded to mark the end of the block and to interrupt the steam.\n ",
"language": "en",
"n_whitespaces": 38,
"n_words": 29,
"vocab_size": 25
} | def iterate_instructions(code_object):
# The arg extension the EXTENDED_ARG opcode represents is automatically handled by get_instructions() but the
# instruction is left in. Get rid of it to make subsequent parsing easier/safer.
yield from (i for i in get_instructions(code_object) if i.opname != "EXTE... | |
28,451 | 127,486 | 367 | python/ray/tune/tuner.py | 61 | 14 | def fit(self) -> ResultGrid:
if not self._is_ray_client:
try:
return self._local_tuner.fit()
except Exception as e:
| [docs/air] Fix up some minor docstrings (#28361) | fit | 5c500f6308dce526c50a5230029fb4909b492a35 | ray | tuner.py | 18 | 41 | https://github.com/ray-project/ray.git | 4 | 93 | 0 | 37 | 175 | Python | {
"docstring": "Executes hyperparameter tuning job as configured and returns result.\n\n Failure handling:\n For the kind of exception that happens during the execution of a trial,\n one may inspect it together with stacktrace through the returned result grid.\n See ``ResultGrid`` for refe... | def fit(self) -> ResultGrid:
if not self._is_ray_client:
try:
return self._local_tuner.fit()
except Exception as e:
raise TuneError(
f"Tune run failed. "
f'Please use tuner = Tuner.restore("'
... | |
109,870 | 311,203 | 27 | homeassistant/components/tradfri/base_class.py | 6 | 4 | def _handle_coordinator_update(self) -> None:
sel | Implement coordinator class for Tradfri integration (#64166)
* Initial commit coordinator
* More coordinator implementation
* More coordinator implementation
* Allow integration reload
* Move API calls to try/catch block
* Move back fixture
* Remove coordinator test file
* Ensure unchanged file
... | _handle_coordinator_update | 9d404b749a0aa0d0527e145c911559df5ecf2afd | core | base_class.py | 9 | 8 | https://github.com/home-assistant/core.git | 1 | 20 | 0 | 6 | 38 | Python | {
"docstring": "\n Handle updated data from the coordinator.\n\n Tests fails without this method.\n ",
"language": "en",
"n_whitespaces": 33,
"n_words": 11,
"vocab_size": 11
} | def _handle_coordinator_update(self) -> None:
self._refresh()
super()._handle_coordinator_update()
| |
36,562 | 156,109 | 195 | dask/dataframe/core.py | 67 | 35 | def _skew_1d(self, column, bias=True, nan_policy="propagate"):
# import depends on scipy, not installed by default
from dask.array import stats as da_stats
if pd.Int64Dtype.is_dtype(column._meta_nonempty):
column = column.astype("f8")
if not np.issubdtype(column.dt... | absolufy-imports - No relative - PEP8 (#8796)
Conversation in https://github.com/dask/distributed/issues/5889 | _skew_1d | cccb9d8d8e33a891396b1275c2448c352ef40c27 | dask | core.py | 11 | 15 | https://github.com/dask/dask.git | 3 | 164 | 0 | 55 | 252 | Python | {
"docstring": "1D version of the skew calculation.\n\n Uses the array version from da.stats in case we are passing in a single series\n ",
"language": "en",
"n_whitespaces": 35,
"n_words": 21,
"vocab_size": 18
} | def _skew_1d(self, column, bias=True, nan_policy="propagate"):
# import depends on scipy, not installed by default
from dask.array import stats as da_stats
if pd.Int64Dtype.is_dtype(column._meta_nonempty):
column = column.astype("f8")
if not np.issubdtype(column.dt... | |
89,989 | 290,876 | 34 | tests/components/sensor/test_init.py | 14 | 8 | def test_device_classes_aligned():
for device_class in NumberDeviceClass:
assert hasattr(SensorDeviceClass, device_class.name)
assert getattr(SensorDeviceClass, device_class.name).value == | Align number and sensor device classes (#81909)
* Align number and sensor device classes
* Add tests
* Tweak tests | test_device_classes_aligned | b6586d5c34bf7ea5c30fbb1b62c438078ea14f39 | core | test_init.py | 12 | 4 | https://github.com/home-assistant/core.git | 2 | 34 | 0 | 13 | 54 | Python | {
"docstring": "Make sure all number device classes are also available in SensorDeviceClass.",
"language": "en",
"n_whitespaces": 10,
"n_words": 11,
"vocab_size": 11
} | def test_device_classes_aligned():
for device_class in NumberDeviceClass:
assert hasattr(SensorDeviceClass, device_class.name)
assert getattr(SensorDeviceClass, device_class.name).value == device_class.value
| |
42,316 | 177,252 | 220 | networkx/algorithms/operators/all.py | 73 | 23 | def intersection_all(graphs):
R = None
for i, G in enumerate(graphs):
G_nodes_set = set(G.nodes)
G_edges_set = set(G.edges(keys=True) if G.is_multigraph() else G.edges())
if i == 0:
# create new graph
R = G.__class__()
node_intersection = G_nodes... | Make all.py generator friendly (#5984)
* Make compose_all generator friendly
* Make disjoint_union_all and intersection_all generator friendly
* Refactor disjoint_union_all to yield relabeled graphs
* Make union_all generator friendly
* Fix intersection_all
* Fix union_all signature
* Allow passing a... | intersection_all | 50ff08de69c6e9541cd6c029bede5dabf56cfe73 | networkx | all.py | 14 | 20 | https://github.com/networkx/networkx.git | 6 | 132 | 0 | 53 | 223 | Python | {
"docstring": "Returns a new graph that contains only the nodes and the edges that exist in\n all graphs.\n\n Parameters\n ----------\n graphs : iterable\n Iterable of NetworkX graphs\n\n Returns\n -------\n R : A new graph with the same type as the first graph in list\n\n Raises\n -... | def intersection_all(graphs):
R = None
for i, G in enumerate(graphs):
G_nodes_set = set(G.nodes)
G_edges_set = set(G.edges(keys=True) if G.is_multigraph() else G.edges())
if i == 0:
# create new graph
R = G.__class__()
node_intersection = G_nodes... | |
@functools.lru_cache(maxsize=None) | 2,983 | 19,466 | 62 | pipenv/patched/notpip/_internal/locations/__init__.py | 28 | 19 | def _looks_like_red_hat_scheme() -> bool: | Vendor in pip 21.2.4 release (from pip 21.2.2 prior). (#5009)
* Vendor in pip 21.2.4 release (from pip 21.2.2 prior).
* Add news fragment for pip 21.2.4 vendor update.
* Add potentially missing LICENSE files | _looks_like_red_hat_scheme | 7e33fcae4384563b4c927fd44318c29dd524a097 | pipenv | __init__.py | 13 | 16 | https://github.com/pypa/pipenv.git | 2 | 52 | 1 | 25 | 137 | Python | {
"docstring": "Red Hat patches ``sys.prefix`` and ``sys.exec_prefix``.\n\n Red Hat's ``00251-change-user-install-location.patch`` changes the install\n command's ``prefix`` and ``exec_prefix`` to append ``\"/local\"``. This is\n (fortunately?) done quite unconditionally, so we create a default command\n ... | def _looks_like_red_hat_scheme() -> bool:
from distutils.command.install import install
from distutils.dist import Distribution
cmd: Any = install(Distribution())
cmd.finalize_options()
return (
cmd.exec_prefix == f"{os.path.normpath(sys.exec_prefix)}/local"
and cmd.prefix == f... |
8,737 | 45,910 | 579 | airflow/providers/databricks/operators/databricks_sql.py | 184 | 33 | def _create_sql_query(self) -> str:
escaper = ParamEscaper()
maybe_with = ""
if self._encryption is not None or self._credential is not None:
maybe_encryption = ""
if self._encryption is not None:
maybe_encryption = self._generate_options("ENCRYPTION", esc... | Add new options to DatabricksCopyIntoOperator (#22076)
This includes:
* `encryption` - to specify encryption options for a given location
* `credential` - to specify authentication options for a given location
* `validate` - to control validation of schema & data | _create_sql_query | 401419432082d222b823e4f2a66f21e5cc3ab28d | airflow | databricks_sql.py | 18 | 40 | https://github.com/apache/airflow.git | 13 | 257 | 0 | 105 | 522 | Python | {
"docstring": "COPY INTO {self._table_name}\nFROM {location}\nFILEFORMAT = {self._file_format}\n{validation}{files_or_pattern}{format_options}{copy_options}\n",
"language": "en",
"n_whitespaces": 5,
"n_words": 9,
"vocab_size": 9
} | def _create_sql_query(self) -> str:
escaper = ParamEscaper()
maybe_with = ""
if self._encryption is not None or self._credential is not None:
maybe_encryption = ""
if self._encryption is not None:
maybe_encryption = self._generate_options("ENCRYPTION", esc... | |
81,887 | 277,193 | 80 | keras/wrappers/scikit_learn.py | 23 | 12 | def predict(self, x, **kwargs):
| Reformatting the codebase with black.
PiperOrigin-RevId: 450093126 | predict | 84afc5193d38057e2e2badf9c889ea87d80d8fbf | keras | scikit_learn.py | 13 | 7 | https://github.com/keras-team/keras.git | 2 | 69 | 0 | 19 | 109 | Python | {
"docstring": "Returns the class predictions for the given test data.\n\n Args:\n x: array-like, shape `(n_samples, n_features)`\n Test samples where `n_samples` is the number of samples\n and `n_features` is the number of features.\n **kwargs: dictionary ar... | def predict(self, x, **kwargs):
proba = self.model.predict(x, **kwargs)
if proba.shape[-1] > 1:
classes = proba.argmax(axis=-1)
else:
classes = (proba > 0.5).astype("int32")
return self.classes_[classes]
| |
9,744 | 49,439 | 340 | modules/image/text_recognition/ppocrv3_det_ch/processor.py | 106 | 23 | def resize_image_type0(self, img):
limit_side_len = self.max_side_len
h, w, _ = img.shape
# limit the max side
if max(h, w) > limit_side_len:
if h > w:
ratio = float(limit_side_len) / h
else:
ratio = float(limit_side_len) ... | add module | resize_image_type0 | 9b3119dfb63c4cbb7acfb9f1f1c09ac24e6d68d2 | PaddleHub | processor.py | 14 | 24 | https://github.com/PaddlePaddle/PaddleHub.git | 6 | 190 | 0 | 66 | 302 | Python | {
"docstring": "\n resize image to a size multiple of 32 which is required by the network\n args:\n img(array): array with shape [h, w, c]\n return(tuple):\n img, (ratio_h, ratio_w)\n ",
"language": "en",
"n_whitespaces": 77,
"n_words": 26,
"vocab_size": 26
} | def resize_image_type0(self, img):
limit_side_len = self.max_side_len
h, w, _ = img.shape
# limit the max side
if max(h, w) > limit_side_len:
if h > w:
ratio = float(limit_side_len) / h
else:
ratio = float(limit_side_len) ... | |
80,796 | 271,560 | 169 | keras/engine/training.py | 48 | 9 | def metrics(self):
metrics = []
if self._is_compiled:
# TODO(omalleyt): Track `LossesContainer` and `MetricsContainer` objects
# so that attr names are not l | Reformatting the codebase with black.
PiperOrigin-RevId: 450093126 | metrics | 84afc5193d38057e2e2badf9c889ea87d80d8fbf | keras | training.py | 12 | 10 | https://github.com/keras-team/keras.git | 5 | 64 | 0 | 36 | 107 | Python | {
"docstring": "Returns the model's metrics added using `compile()`, `add_metric()` APIs.\n\n Note: Metrics passed to `compile()` are available only after a `keras.Model`\n has been trained/evaluated on actual data.\n\n Examples:\n\n >>> inputs = tf.keras.layers.Input(shape=(3,))\n ... | def metrics(self):
metrics = []
if self._is_compiled:
# TODO(omalleyt): Track `LossesContainer` and `MetricsContainer` objects
# so that attr names are not load-bearing.
if self.compiled_loss is not None:
metrics += self.compiled_loss.metrics
... | |
50,661 | 204,162 | 33 | django/contrib/messages/api.py | 14 | 5 | def set_level(request, level):
if not hasattr(request, "_messages"):
retur | Refs #33476 -- Reformatted code with Black. | set_level | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | api.py | 9 | 5 | https://github.com/django/django.git | 2 | 28 | 0 | 13 | 48 | Python | {
"docstring": "\n Set the minimum level of messages to be recorded, and return ``True`` if\n the level was recorded successfully.\n\n If set to ``None``, use the default level (see the get_level() function).\n ",
"language": "en",
"n_whitespaces": 43,
"n_words": 30,
"vocab_size": 24
} | def set_level(request, level):
if not hasattr(request, "_messages"):
return False
request._messages.level = level
return True
| |
42,282 | 177,139 | 226 | networkx/algorithms/lowest_common_ancestors.py | 92 | 15 | def all_pairs_lowest_common_ancestor(G, pairs=None):
if not nx.is_directed_acyclic_graph(G):
raise nx.NetworkXError | Replace LCA with naive implementations (#5883)
* WIP: Replace functions to evaluate tests.
* Raise prompt exceptions by wrapping generator.
* Fix erroneous ground-truth self-ancestor in tests.
* Move pair creation outside of generator and validate.
* Convert input with fromkeys to preserve order and rm dup... | all_pairs_lowest_common_ancestor | 19c1454d3dfa70a893ea67f2d78515658e8c08e5 | networkx | lowest_common_ancestors.py | 19 | 17 | https://github.com/networkx/networkx.git | 6 | 100 | 0 | 67 | 171 | Python | {
"docstring": "Return the lowest common ancestor of all pairs or the provided pairs\n\n Parameters\n ----------\n G : NetworkX directed graph\n\n pairs : iterable of pairs of nodes, optional (default: all pairs)\n The pairs of nodes of interest.\n If None, will find the LCA of all pairs of ... | def all_pairs_lowest_common_ancestor(G, pairs=None):
if not nx.is_directed_acyclic_graph(G):
raise nx.NetworkXError("LCA only defined on directed acyclic graphs.")
if len(G) == 0:
raise nx.NetworkXPointlessConcept("LCA meaningless on null graphs.")
if pairs is None:
pairs = com... | |
11,176 | 54,884 | 21 | tests/test_task_runners.py | 7 | 4 | async def test_is_pickleable_after_start(self, task_runner):
task_runner.client_kwargs["set_as_default"] = True | Introduces testing module and task runner standard test suite (PrefectHQ/orion#1655) | test_is_pickleable_after_start | 7895fb5013aa98955bb31be96953ac13979a5b94 | prefect | test_task_runners.py | 8 | 6 | https://github.com/PrefectHQ/prefect.git | 1 | 50 | 0 | 7 | 30 | Python | {
"docstring": "\n The task_runner must be picklable as it is attached to `PrefectFuture` objects\n Reimplemented to set Dask client as default to allow unpickling\n ",
"language": "en",
"n_whitespaces": 44,
"n_words": 22,
"vocab_size": 19
} | async def test_is_pickleable_after_start(self, task_runner):
task_runner.client_kwargs["set_as_default"] = True | |
55,971 | 220,349 | 26 | python3.10.4/Lib/asyncio/base_events.py | 5 | 5 | def _add_callback_signalsafe(self, handle):
self._add_callback(handle)
self._write_to_self()
| add python 3.10.4 for windows | _add_callback_signalsafe | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | base_events.py | 7 | 3 | https://github.com/XX-net/XX-Net.git | 1 | 19 | 0 | 5 | 34 | Python | {
"docstring": "Like _add_callback() but called from a signal handler.",
"language": "en",
"n_whitespaces": 7,
"n_words": 8,
"vocab_size": 8
} | def _add_callback_signalsafe(self, handle):
self._add_callback(handle)
self._write_to_self()
| |
5,324 | 30,115 | 53 | spotdl/utils/ffmpeg.py | 24 | 8 | def get_local_ffmpeg() -> Optional[Path]:
| v4 init | get_local_ffmpeg | fa2ad657482aca9dc628e6d7062b8badf2706bb6 | spotify-downloader | ffmpeg.py | 12 | 10 | https://github.com/spotDL/spotify-downloader.git | 3 | 43 | 0 | 21 | 79 | Python | {
"docstring": "\n Get local ffmpeg binary path or None if not found.\n ",
"language": "en",
"n_whitespaces": 17,
"n_words": 10,
"vocab_size": 10
} | def get_local_ffmpeg() -> Optional[Path]:
ffmpeg_path = Path(
get_spotdl_path(), "ffmpeg" + ".exe" if platform.system() == "Windows" else ""
)
if ffmpeg_path.is_file():
return ffmpeg_path
return None
| |
54,474 | 216,231 | 184 | salt/modules/cmdmod.py | 88 | 20 | def _render_cmd(cmd, cwd, template, saltenv=None, pillarenv=None, pillar_override=None):
if saltenv is None:
saltenv = __opts__.get("saltenv", "base")
if not template:
return (cmd, cwd)
# render the path as a template using path_template_engine as the engine
if template not in salt... | fixes salt bug 61507 | _render_cmd | 21d3f4bc9eb7b9fb1118c59073595a9e9ee836bd | salt | cmdmod.py | 12 | 23 | https://github.com/saltstack/salt.git | 7 | 155 | 0 | 60 | 225 | Python | {
"docstring": "\n If template is a valid template engine, process the cmd and cwd through\n that engine.\n ",
"language": "en",
"n_whitespaces": 25,
"n_words": 15,
"vocab_size": 14
} | def _render_cmd(cmd, cwd, template, saltenv=None, pillarenv=None, pillar_override=None):
if saltenv is None:
saltenv = __opts__.get("saltenv", "base")
if not template:
return (cmd, cwd)
# render the path as a template using path_template_engine as the engine
if template not in salt... | |
71,631 | 247,373 | 56 | tests/rest/media/v1/test_html_preview.py | 21 | 8 | def test_windows_1252(self) -> None:
html = b
tree = decode_body(html, "http://example.com/test.html")
og = parse | Add type hints to `tests/rest` (#12146)
* Add type hints to `tests/rest`
* newsfile
* change import from `SigningKey` | test_windows_1252 | 7e91107be1a4287873266e588a3c5b415279f4c8 | synapse | test_html_preview.py | 10 | 13 | https://github.com/matrix-org/synapse.git | 1 | 44 | 0 | 18 | 83 | Python | {
"docstring": "A body which uses cp1252, but doesn't declare that.\n <html>\n <head><title>\\xf3</title></head>\n <body>\n Some text.\n </body>\n </html>\n ",
"language": "en",
"n_whitespaces": 65,
"n_words": 16,
"vocab_size": 16
} | def test_windows_1252(self) -> None:
html = b
tree = decode_body(html, "http://example.com/test.html")
og = parse_html_to_open_graph(tree, "http://example.com/test.html")
self.assertEqual(og, {"og:title": "ó", "og:description": "Some text."})
| |
99,466 | 300,606 | 58 | homeassistant/helpers/template.py | 17 | 10 | def square_root(value, default=_SENTINEL):
try:
return m | Fail template functions when no default specified (#71687) | square_root | 4885331509eeffe50f42d76b234996467b06170f | core | template.py | 13 | 7 | https://github.com/home-assistant/core.git | 3 | 42 | 0 | 15 | 70 | Python | {
"docstring": "Filter and function to get square root of the value.",
"language": "en",
"n_whitespaces": 9,
"n_words": 10,
"vocab_size": 10
} | def square_root(value, default=_SENTINEL):
try:
return math.sqrt(float(value))
except (ValueError, TypeError):
if default is _SENTINEL:
raise_no_default("sqrt", value)
return default
| |
90,842 | 291,738 | 73 | tests/test_core.py | 19 | 9 | async def test_track_task_functions(event_loop):
hass = ha.HomeAssistant()
try:
assert hass._track_task
hass.async_stop_track_tasks()
assert not hass._track_task
hass.async_track_tasks()
assert hass._track_task
finally:
await hass.async_stop()
| Upgrade pytest-aiohttp (#82475)
* Upgrade pytest-aiohttp
* Make sure executors, tasks and timers are closed
Some test will trigger warnings on garbage collect, these warnings
spills over into next test.
Some test trigger tasks that raise errors on shutdown, these spill
over into next test.
This is to mim... | test_track_task_functions | c576a68d336bc91fd82c299d9b3e5dfdc1c14960 | core | test_core.py | 11 | 10 | https://github.com/home-assistant/core.git | 2 | 46 | 0 | 15 | 83 | Python | {
"docstring": "Test function to start/stop track task and initial state.",
"language": "en",
"n_whitespaces": 8,
"n_words": 9,
"vocab_size": 9
} | async def test_track_task_functions(event_loop):
hass = ha.HomeAssistant()
try:
assert hass._track_task
hass.async_stop_track_tasks()
assert not hass._track_task
hass.async_track_tasks()
assert hass._track_task
finally:
await hass.async_stop()
| |
92,344 | 293,281 | 46 | homeassistant/components/kaleidescape/media_player.py | 14 | 8 | def media_position_updated_at(self) -> datetime | None:
| Add Kaleidescape integration (#67711) | media_position_updated_at | ea82f2e293f43d3e5be103a64b68d088c4b65545 | core | media_player.py | 9 | 5 | https://github.com/home-assistant/core.git | 2 | 27 | 0 | 13 | 46 | Python | {
"docstring": "When was the position of the current playing media valid.",
"language": "en",
"n_whitespaces": 9,
"n_words": 10,
"vocab_size": 9
} | def media_position_updated_at(self) -> datetime | None:
if self._device.movie.play_status in KALEIDESCAPE_PLAYING_STATES:
return utcnow()
return None
| |
56,101 | 220,720 | 100 | python3.10.4/Lib/asyncio/sslproto.py | 23 | 12 | def write(self, data):
if n | add python 3.10.4 for windows | write | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | sslproto.py | 15 | 7 | https://github.com/XX-net/XX-Net.git | 3 | 44 | 0 | 21 | 84 | Python | {
"docstring": "Write some data bytes to the transport.\n\n This does not block; it buffers the data and arranges for it\n to be sent out asynchronously.\n ",
"language": "en",
"n_whitespaces": 45,
"n_words": 24,
"vocab_size": 20
} | def write(self, data):
if not isinstance(data, (bytes, bytearray, memoryview)):
raise TypeError(f"data: expecting a bytes-like instance, "
f"got {type(data).__name__}")
if not data:
return
self._ssl_protocol._write_appdata(data)
| |
46,085 | 189,484 | 357 | manim/mobject/svg/text_mobject.py | 73 | 22 | def _extract_color_tags(self):
tags = re.finditer(
r'<color\s+col="([^"]+)"(\s+offset="([^"]+)")?>(.+?)</color>',
self.original_text,
re.S,
)
colormap = []
for tag in tags:
start = self._count_real_chars(self.original_text[: tag.s... | Hide more private methods from the docs. (#2468)
* hide privs from text_mobject.py
* hide privs from tex_mobject.py
* hide privs from code_mobject.py
* hide privs from svg_mobject.py
* remove SVGPath and utils from __init__.py
* don't import string_to_numbers
* hide privs from geometry.py
* hide p... | _extract_color_tags | 902e7eb4f0147b5882a613b67467e38a1d47f01e | manim | text_mobject.py | 14 | 24 | https://github.com/ManimCommunity/manim.git | 6 | 188 | 0 | 58 | 296 | Python | {
"docstring": "Used to determine which parts (if any) of the string should be formatted\n with a custom color.\n\n Removes the ``<color>`` tag, as it is not part of Pango's markup and would cause an error.\n\n Note: Using the ``<color>`` tags is deprecated. As soon as the legacy syntax is gone, ... | def _extract_color_tags(self):
tags = re.finditer(
r'<color\s+col="([^"]+)"(\s+offset="([^"]+)")?>(.+?)</color>',
self.original_text,
re.S,
)
colormap = []
for tag in tags:
start = self._count_real_chars(self.original_text[: tag.s... | |
39,268 | 162,729 | 888 | frequency_response.py | 239 | 44 | def write_readme(self, file_path, parametric_eq_peqs=None, fixed_band_eq_peq=None):
file_path = os.path.abspath(file_path)
dir_path = os.path.dirname(file_path)
model = self.name
# Write model
s = '# {}\n'.format(model)
s += 'See [usage instructions](https://git... | Added PEQ configs to CLI and function interfaces. Improved default value handling for PEQ parameters and added more predefined configs. Removed legacy PEQ optimization. Fixed readme write. Improved shelf filter initialization. Added plot method to PEQ. Notebook for comparing old and new optimizers. Bug fixes. | write_readme | 9120cdffe618c6c2ff16fe6a311b6a1367efdbc8 | AutoEq | frequency_response.py | 18 | 44 | https://github.com/jaakkopasanen/AutoEq.git | 11 | 312 | 0 | 135 | 638 | Python | {
"docstring": "Writes README.md with picture and Equalizer APO settings.",
"language": "en",
"n_whitespaces": 7,
"n_words": 8,
"vocab_size": 8
} | def write_readme(self, file_path, parametric_eq_peqs=None, fixed_band_eq_peq=None):
file_path = os.path.abspath(file_path)
dir_path = os.path.dirname(file_path)
model = self.name
# Write model
s = '# {}\n'.format(model)
s += 'See [usage instructions](https://git... | |
74,967 | 256,950 | 158 | haystack/telemetry.py | 59 | 13 | def _get_execution_environment():
if os.environ.get("CI", "False").lower() == "true":
execution_env | Add basic telemetry features (#2314)
* add basic telemetry features
* change pipeline_config to _component_config
* Update Documentation & Code Style
* add super().__init__() calls to error classes
* make posthog mock work with python 3.7
* Update Documentation & Code Style
* update link to docs web ... | _get_execution_environment | ac5617e757e9ace6f30b7291686d9dbbc339f433 | haystack | telemetry.py | 15 | 17 | https://github.com/deepset-ai/haystack.git | 7 | 94 | 0 | 36 | 180 | Python | {
"docstring": "\n Identifies the execution environment that Haystack is running in.\n Options are: colab notebook, kubernetes, CPU/GPU docker container, test environment, jupyter notebook, python script\n ",
"language": "en",
"n_whitespaces": 33,
"n_words": 23,
"vocab_size": 22
} | def _get_execution_environment():
if os.environ.get("CI", "False").lower() == "true":
execution_env = "ci"
elif "google.colab" in sys.modules:
execution_env = "colab"
elif "KUBERNETES_SERVICE_HOST" in os.environ:
execution_env = "kubernetes"
elif HAYSTACK_DOCKER_CONTAINER in... | |
12,258 | 60,716 | 58 | .venv/lib/python3.8/site-packages/pip/_internal/index/collector.py | 30 | 17 | def _ensure_html_response(url, session):
# type: (str, PipSession) -> None
scheme, netloc, path, query, fragment = urllib.parse.urlsplit(url)
if scheme not in {'http', 'https'}:
raise _NotHTTP()
resp = session.head(url, allow_redirects=True)
raise_for_status(resp)
_ensure_html | upd; format | _ensure_html_response | f638f5d0e6c8ebed0e69a6584bc7f003ec646580 | transferlearning | collector.py | 9 | 7 | https://github.com/jindongwang/transferlearning.git | 2 | 60 | 0 | 29 | 98 | Python | {
"docstring": "Send a HEAD request to the URL, and ensure the response contains HTML.\n\n Raises `_NotHTTP` if the URL is not available for a HEAD request, or\n `_NotHTML` if the content type is not text/html.\n ",
"language": "en",
"n_whitespaces": 43,
"n_words": 34,
"vocab_size": 26
} | def _ensure_html_response(url, session):
# type: (str, PipSession) -> None
scheme, netloc, path, query, fragment = urllib.parse.urlsplit(url)
if scheme not in {'http', 'https'}:
raise _NotHTTP()
resp = session.head(url, allow_redirects=True)
raise_for_status(resp)
_ensure_html_hea... | |
4,051 | 21,808 | 576 | pipenv/vendor/tomlkit/parser.py | 120 | 23 | def _parse_item(self) -> Optional[Tuple[Optional[Key], Item]]:
self.mark()
with self._state as state:
while True:
c = self._current
if c | Update tomlkit==0.9.2
Used:
python -m invoke vendoring.update --package=tomlkit | _parse_item | 8faa74cdc9da20cfdcc69f5ec29b91112c95b4c9 | pipenv | parser.py | 19 | 25 | https://github.com/pypa/pipenv.git | 7 | 144 | 0 | 77 | 253 | Python | {
"docstring": "\n Attempts to parse the next item and returns it, along with its key\n if the item is value-like.\n ",
"language": "en",
"n_whitespaces": 40,
"n_words": 18,
"vocab_size": 16
} | def _parse_item(self) -> Optional[Tuple[Optional[Key], Item]]:
self.mark()
with self._state as state:
while True:
c = self._current
if c == "\n":
# Found a newline; Return all whitespace found up to this point.
... | |
49,705 | 200,570 | 296 | sympy/tensor/tensor.py | 55 | 12 | def _xreplace(self, rule):
if self in rule:
return rule[self], True
elif rule:
rule = self._dedupe_indices_in_rule(rule)
args = []
changed = False
for a in self.args:
_xreplace = getattr(a, '_xreplace', None)
... | Add TensMul._xreplace to dedupe dummy indices
This apparently should not be done.
https://github.com/sympy/sympy/pull/24333#issuecomment-1333783127 | _xreplace | 49222e1e3645ca6948277b8e1a9b526daeb0948d | sympy | tensor.py | 15 | 19 | https://github.com/sympy/sympy.git | 6 | 113 | 0 | 39 | 182 | Python | {
"docstring": "\n Helper for xreplace. Tracks whether a replacement actually occurred.\n\n Given that the rule has entries {old:new, ...}, this handles the fact\n that if a dummy index in new is the same as an index in self, the\n dummy index in new must be renamed.\n ",
"languag... | def _xreplace(self, rule):
if self in rule:
return rule[self], True
elif rule:
rule = self._dedupe_indices_in_rule(rule)
args = []
changed = False
for a in self.args:
_xreplace = getattr(a, '_xreplace', None)
... | |
104,430 | 305,646 | 30 | homeassistant/components/mpd/media_player.py | 9 | 6 | async def async_turn_on(self) -> None:
await self._client.play()
await self._update_playlists(no_throttle=True)
| Improve entity type hints [m] (#77816) | async_turn_on | 6355e682fa4aeb526570597d919ad1fb76755b9a | core | media_player.py | 9 | 4 | https://github.com/home-assistant/core.git | 1 | 25 | 0 | 8 | 46 | Python | {
"docstring": "Service to send the MPD the command to start playing.",
"language": "en",
"n_whitespaces": 9,
"n_words": 10,
"vocab_size": 8
} | async def async_turn_on(self) -> None:
await self._client.play()
await self._update_playlists(no_throttle=True)
| |
80,442 | 270,345 | 228 | keras/distribute/distributed_training_utils_v1.py | 122 | 18 | def _build_network_on_replica(model, mode, inputs=None, targets=None):
# Need to do imports here since we run into a circular dependency error.
from keras import models # pylint: disable=g-import-not-at-top
from keras.engine import sequential # pylint: disable=g-import-not-at-top
# We rely on th... | Reformatting the codebase with black.
PiperOrigin-RevId: 450093126 | _build_network_on_replica | 84afc5193d38057e2e2badf9c889ea87d80d8fbf | keras | distributed_training_utils_v1.py | 13 | 33 | https://github.com/keras-team/keras.git | 6 | 188 | 0 | 88 | 133 | Python | {
"docstring": "Build an updated model on replicas.\n\n We create a new Keras model while sharing the variables from the old graph.\n Building a new sub-graph is required since the original keras model creates\n placeholders for the input and the output that are not accessible till we\n call iterator.get_... | def _build_network_on_replica(model, mode, inputs=None, targets=None):
# Need to do imports here since we run into a circular dependency error.
from keras import models # pylint: disable=g-import-not-at-top
from keras.engine import sequential # pylint: disable=g-import-not-at-top
# We rely on th... | |
83,438 | 280,800 | 111 | keras/utils/tf_utils.py | 36 | 8 | def can_jit_compile(warn=False):
if platform.system() == "Darwin" and "arm" in pla | Set `jit_compile` only when TensorFlow XLA is available for the platform.
Fixes issue of using new optimizers on Mac M1 as TF on Mac M1 is not built with XLA.
PiperOrigin-RevId: 497158007 | can_jit_compile | 1b32391798a952176b733660c940b1589c2fc8a4 | keras | tf_utils.py | 13 | 9 | https://github.com/keras-team/keras.git | 4 | 43 | 0 | 34 | 82 | Python | {
"docstring": "Returns True if TensorFlow XLA is available for the platform.",
"language": "en",
"n_whitespaces": 9,
"n_words": 10,
"vocab_size": 10
} | def can_jit_compile(warn=False):
if platform.system() == "Darwin" and "arm" in platform.processor().lower():
if warn:
logging.warning(
"Tensorflow is not compiled with XLA on Mac M1 Arm processors, "
"so cannot set `jit_compile` to True."
)
... | |
28,767 | 128,633 | 637 | rllib/algorithms/algorithm.py | 165 | 23 | def __setstate__(self, state) -> None:
# TODO (sven): Validate that our config and the config in state are compatible.
# For example, the model architectures may differ.
# Also, what should the behavior be if e.g. some training parameter
# (e.g. lr) changed?
if hasat... | [RLlib] Algorithm/Policy checkpoint overhaul and Policy Model export (in native formats). (#28166) | __setstate__ | 23b3a599b9df8100558c477e94b0b19b1a38ac27 | ray | algorithm.py | 16 | 33 | https://github.com/ray-project/ray.git | 12 | 176 | 0 | 106 | 318 | Python | {
"docstring": "Sets the algorithm to the provided state.\n\n Args:\n state: The state dict to restore this Algorithm instance to. `state` may\n have been returned by a call to an Algorithm's `__getstate__()` method.\n ",
"language": "en",
"n_whitespaces": 71,
"n_words": ... | def __setstate__(self, state) -> None:
# TODO (sven): Validate that our config and the config in state are compatible.
# For example, the model architectures may differ.
# Also, what should the behavior be if e.g. some training parameter
# (e.g. lr) changed?
if hasat... | |
73,880 | 251,891 | 33 | test/mitmproxy/proxy/layers/http/test_http2.py | 18 | 15 | def test_http2_client_aborts(tctx, stream, when, how):
server = P | make it black! | test_http2_client_aborts | b3587b52b25077f68116b9852b041d33e7fc6601 | mitmproxy | test_http2.py | 8 | 86 | https://github.com/mitmproxy/mitmproxy.git | 12 | 494 | 0 | 15 | 62 | Python | {
"docstring": "\n Test handling of the case where a client aborts during request or response transmission.\n\n If the client aborts the request transmission, we must trigger an error hook,\n if the client disconnects during response transmission, no error hook is triggered.\n ",
"language": "en",
"n_... | def test_http2_client_aborts(tctx, stream, when, how):
server = Placeholder(Server)
flow = Placeholder(HTTPFlow)
playbook, cff = start_h2_client(tctx)
resp = Placeholder(bytes)
| |
50,366 | 203,421 | 277 | django/contrib/admin/options.py | 49 | 21 | def get_changelist_instance(self, request):
list_display = self.get_list_display(request)
list_display_links = self.get_list_display_links(request, list_display)
# Add the action checkboxes if any actions are available.
if self.get_actions(request):
list_display = ["... | Refs #33476 -- Reformatted code with Black. | get_changelist_instance | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | options.py | 10 | 23 | https://github.com/django/django.git | 2 | 117 | 0 | 43 | 174 | Python | {
"docstring": "\n Return a `ChangeList` instance based on `request`. May raise\n `IncorrectLookupParameters`.\n ",
"language": "en",
"n_whitespaces": 32,
"n_words": 10,
"vocab_size": 10
} | def get_changelist_instance(self, request):
list_display = self.get_list_display(request)
list_display_links = self.get_list_display_links(request, list_display)
# Add the action checkboxes if any actions are available.
if self.get_actions(request):
list_display = ["... | |
26,878 | 120,516 | 156 | jax/_src/lax/qdwh.py | 103 | 29 | def _use_cholesky(u, m, n, params):
a, b, c = params
_, N = u.shape
x = c * (u.T.conj() @ u) + jnp.eye(N, dtype=jnp.dtype(u))
# Pads the lower-right corner with the identity matrix to prevent the Cholesky
# decomposition from failing due to the matrix not being PSD if padded with
# zeros.
x = _mask(x, ... | Add support for padded arrays in QDWH algorithm.
This change is in preparation for adding a jit-table QDWH-eig implementation.
PiperOrigin-RevId: 448571523 | _use_cholesky | db73670ec3fc72f75e6f832351620ac79e9b0c6f | jax | qdwh.py | 13 | 13 | https://github.com/google/jax.git | 1 | 174 | 0 | 72 | 261 | Python | {
"docstring": "QDWH iteration using Cholesky decomposition.\n\n Args:\n u: a matrix, with static (padded) shape M x N\n m, n: the dynamic shape of the matrix, where m <= M and n <= N.\n params: the QDWH parameters.\n ",
"language": "en",
"n_whitespaces": 41,
"n_words": 36,
"vocab_size": 29
} | def _use_cholesky(u, m, n, params):
a, b, c = params
_, N = u.shape
x = c * (u.T.conj() @ u) + jnp.eye(N, dtype=jnp.dtype(u))
# Pads the lower-right corner with the identity matrix to prevent the Cholesky
# decomposition from failing due to the matrix not being PSD if padded with
# zeros.
x = _mask(x, ... | |
47,450 | 195,863 | 39 | sympy/functions/elementary/piecewise.py | 18 | 11 | def piecewise_integrate(self, x, **kwargs):
from sympy.integrals import integrate
return self.func(*[(integrate(e, x, **kwargs), c) for e, c in self.args])
| Improved documentation formatting | piecewise_integrate | cda8dfe6f45dc5ed394c2f5cda706cd6c729f713 | sympy | piecewise.py | 13 | 3 | https://github.com/sympy/sympy.git | 2 | 47 | 0 | 17 | 71 | Python | {
"docstring": "Return the Piecewise with each expression being\n replaced with its antiderivative. To obtain a continuous\n antiderivative, use the :func:`~.integrate` function or method.\n\n Examples\n ========\n\n >>> from sympy import Piecewise\n >>> from sympy.abc import... | def piecewise_integrate(self, x, **kwargs):
from sympy.integrals import integrate
return self.func(*[(integrate(e, x, **kwargs), c) for e, c in self.args])
| |
40,835 | 173,321 | 527 | cps/helper.py | 114 | 21 | def check_send_to_ereader(entry):
formats = list()
book_formats = list()
if len(entry.data):
for ele in iter(entry.data):
if ele.uncompressed_size < config.mail_size:
formats.append(ele.format)
if 'EPUB' in formats:
book_formats.append({'format': ... | Eenabled send epubs to E-Reader devices | check_send_to_ereader | fbac3e38ac116855b930ee60fb3c997337ae17b7 | calibre-web | helper.py | 17 | 29 | https://github.com/janeczku/calibre-web.git | 9 | 202 | 0 | 65 | 370 | Python | {
"docstring": "\n returns all available book formats for sending to E-Reader\n ",
"language": "en",
"n_whitespaces": 20,
"n_words": 9,
"vocab_size": 9
} | def check_send_to_ereader(entry):
formats = list()
book_formats = list()
if len(entry.data):
for ele in iter(entry.data):
if ele.uncompressed_size < config.mail_size:
formats.append(ele.format)
if 'EPUB' in formats:
book_formats.append({'format': ... | |
48,606 | 197,528 | 37 | sympy/stats/joint_rv_types.py | 23 | 8 | def Multinomial(syms, n, *p):
if not isinstance(p[0], list):
p = (list(p), )
return multivariate_rv(MultinomialDistribution, syms, n, p[0])
#-------------------------------------------------------------------------------
# Negative | Improved some documentation in the stats module | Multinomial | 7fe8e027ae1d7f683243c0229b961671a6cbb4c5 | sympy | joint_rv_types.py | 11 | 4 | https://github.com/sympy/sympy.git | 2 | 46 | 0 | 22 | 71 | Python | {
"docstring": "\n Creates a discrete random variable with Multinomial Distribution.\n\n The density of the said distribution can be found at [1].\n\n Parameters\n ==========\n\n n : Positive integer\n Represents number of trials\n p : List of event probabilites\n Must be in the range ... | def Multinomial(syms, n, *p):
if not isinstance(p[0], list):
p = (list(p), )
return multivariate_rv(MultinomialDistribution, syms, n, p[0])
#-------------------------------------------------------------------------------
# Negative Multinomial Distribution -----------------------------------------... | |
49,473 | 199,984 | 64 | sympy/physics/qho_1d.py | 43 | 16 | def psi_n(n, x, m, omega):
# sympify arguments
n, x, m, omega = map(S, [n, x, m, omega])
nu = m * omega / hbar
# | applied backtick correction to the remainder of the project | psi_n | a0989bcfd26470833cf03737941bfd80f511c745 | sympy | qho_1d.py | 14 | 5 | https://github.com/sympy/sympy.git | 1 | 97 | 0 | 31 | 146 | Python | {
"docstring": "\n Returns the wavefunction psi_{n} for the One-dimensional harmonic oscillator.\n\n Parameters\n ==========\n\n n :\n the \"nodal\" quantum number. Corresponds to the number of nodes in the\n wavefunction. ``n >= 0``\n x :\n x coordinate.\n m :\n Mass o... | def psi_n(n, x, m, omega):
# sympify arguments
n, x, m, omega = map(S, [n, x, m, omega])
nu = m * omega / hbar
# normalization coefficient
C = (nu/pi)**Rational(1, 4) * sqrt(1/(2**n*factorial(n)))
return C * exp(-nu* x**2 /2) * hermite(n, sqrt(nu)*x)
| |
55,321 | 218,459 | 201 | python3.10.4/Lib/inspect.py | 85 | 23 | def getsourcefile(object):
filename = getfile(object)
all_bytecode_suffixes = importlib.machinery.DEBUG_BYTECODE_SUFFIXES[:]
all_bytecode_suffixes += importlib.machinery.OPTIMIZED_BYTECODE_SUFFIXES[:]
if any(filename.endswith(s) for s in all_bytecode_suffixes):
filename = (os.path.splitext(... | add python 3.10.4 for windows | getsourcefile | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | inspect.py | 14 | 19 | https://github.com/XX-net/XX-Net.git | 9 | 158 | 0 | 49 | 250 | Python | {
"docstring": "Return the filename that can be used to locate an object's source.\n Return None if no way can be identified to get the source.\n ",
"language": "en",
"n_whitespaces": 30,
"n_words": 24,
"vocab_size": 18
} | def getsourcefile(object):
filename = getfile(object)
all_bytecode_suffixes = importlib.machinery.DEBUG_BYTECODE_SUFFIXES[:]
all_bytecode_suffixes += importlib.machinery.OPTIMIZED_BYTECODE_SUFFIXES[:]
if any(filename.endswith(s) for s in all_bytecode_suffixes):
filename = (os.path.splitext(... | |
55,086 | 218,024 | 76 | python3.10.4/Lib/importlib/_abc.py | 27 | 7 | def module_repr(self, module):
warnings.warn("importlib.abc.Loader.module_repr() is deprecated and "
"slated for removal in Python 3.12", DeprecationWarning | add python 3.10.4 for windows | module_repr | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | _abc.py | 9 | 4 | https://github.com/XX-net/XX-Net.git | 1 | 19 | 0 | 27 | 37 | Python | {
"docstring": "Return a module's repr.\n\n Used by the module type when the method does not raise\n NotImplementedError.\n\n This method is deprecated.\n\n ",
"language": "en",
"n_whitespaces": 48,
"n_words": 20,
"vocab_size": 18
} | def module_repr(self, module):
warnings.warn("importlib.abc.Loader.module_repr() is deprecated and "
"slated for removal in Python 3.12", DeprecationWarning)
# The exception will cause ModuleType.__repr__ to ignore this method.
raise NotImplementedError
| |
72,960 | 249,509 | 275 | tests/push/test_email.py | 24 | 25 | def test_need_validated_email(self):
with | Support enabling/disabling pushers (from MSC3881) (#13799)
Partial implementation of MSC3881 | test_need_validated_email | 8ae42ab8fa3c6b52d74c24daa7ca75a478fa4fbb | synapse | test_email.py | 14 | 17 | https://github.com/matrix-org/synapse.git | 1 | 99 | 0 | 23 | 161 | Python | {
"docstring": "Test that we can only add an email pusher if the user has validated\n their email.\n ",
"language": "en",
"n_whitespaces": 30,
"n_words": 16,
"vocab_size": 16
} | def test_need_validated_email(self):
with self.assertRaises(SynapseError) as cm:
self.get_success_or_raise(
self.hs.get_pusherpool().add_or_update_pusher(
user_id=self.user_id,
access_token=self.token_id,
kind="emai... | |
117,028 | 319,962 | 419 | src/documents/tests/test_api.py | 64 | 28 | def test_get_existing_comments(self):
doc = Document.objects.create(
title="test",
mime_type="application/pdf",
content="this is a document which will have comments!",
)
comment = Comment.objects.create(
comment="This is a comment.",
... | Starts on implementing tests for the new API | test_get_existing_comments | 6d5d308d6c7b7e359ba72964a300634e1065ace9 | paperless-ngx | test_api.py | 13 | 33 | https://github.com/paperless-ngx/paperless-ngx.git | 1 | 164 | 0 | 51 | 275 | Python | {
"docstring": "\n GIVEN:\n - A document with a single comment\n WHEN:\n - API reuqest for document comments is made\n THEN:\n - The associated comment is returned\n ",
"language": "en",
"n_whitespaces": 86,
"n_words": 24,
"vocab_size": 19
} | def test_get_existing_comments(self):
doc = Document.objects.create(
title="test",
mime_type="application/pdf",
content="this is a document which will have comments!",
)
comment = Comment.objects.create(
comment="This is a comment.",
... | |
77,276 | 262,565 | 155 | TTS/vocoder/datasets/wavegrad_dataset.py | 62 | 16 | def collate_full_clips(batch):
max_mel_length = max([b[0].sh | Fix tune wavegrad (#1844)
* fix imports in tune_wavegrad
* load_config returns Coqpit object instead None
* set action (store true) for flag "--use_cuda"; start to tune if module is running as the main program
* fix var order in the result of batch collating
* make style
* make style with black and isor... | collate_full_clips | 2c9f00a808e0aa76a82af2e8b325abb71f50d1df | TTS | wavegrad_dataset.py | 13 | 11 | https://github.com/coqui-ai/TTS.git | 6 | 185 | 0 | 38 | 272 | Python | {
"docstring": "This is used in tune_wavegrad.py.\n It pads sequences to the max length.",
"language": "en",
"n_whitespaces": 18,
"n_words": 12,
"vocab_size": 12
} | def collate_full_clips(batch):
max_mel_length = max([b[0].shape[1] for b in batch]) if len(batch) > 1 else batch[0][0].shape[1]
max_audio_length = max([b[1].shape[0] for b in batch]) if len(batch) > 1 else batch[0][1].shape[0]
mels = torch.zeros([len(batch), batch[0][0].shape[0], max_m... | |
38,550 | 160,178 | 249 | numpy/distutils/ccompiler_opt.py | 81 | 19 | def feature_test(self, name, force_flags=None, macros=[]):
if force_flags is None:
force_flags = self.feature_flags(name)
self.dist_log(
"testing feature '%s' with flags (%s)" % (
name, ' '.join(force_flags)
))
# Each CPU feature must have C ... | Add space after argument name | feature_test | f404e9e92e87a3990712d723d5c562a89300ac01 | numpy | ccompiler_opt.py | 12 | 18 | https://github.com/numpy/numpy.git | 4 | 123 | 0 | 65 | 204 | Python | {
"docstring": "\n Test a certain CPU feature against the compiler through its own\n check file.\n\n Parameters\n ----------\n name : str\n Supported CPU feature name.\n\n force_flags : list or None, optional\n If None(default), the returned flags from `... | def feature_test(self, name, force_flags=None, macros=[]):
if force_flags is None:
force_flags = self.feature_flags(name)
self.dist_log(
"testing feature '%s' with flags (%s)" % (
name, ' '.join(force_flags)
))
# Each CPU feature must have C ... | |
46,282 | 189,987 | 48 | manim/mobject/svg/svg_mobject.py | 16 | 6 | def get_file_path(self) -> str:
if self.file_name is None:
raise ValueError("Must specify file for SVGMobject")
return get_full_vector_image_path(self.f | Ported improved implementation of :class:`.SVGMobject` from 3b1b/manim (#2898)
* port SVGMobject from 3b1b/manim
* added svgelements as dependency
* revert change of default values
* [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
* set default stroke... | get_file_path | 309c9d41eb734ca85a7aea5533f88a6d4ee7c944 | manim | svg_mobject.py | 10 | 5 | https://github.com/ManimCommunity/manim.git | 2 | 27 | 0 | 16 | 48 | Python | {
"docstring": "Search for an existing file based on the specified file name.",
"language": "en",
"n_whitespaces": 10,
"n_words": 11,
"vocab_size": 10
} | def get_file_path(self) -> str:
if self.file_name is None:
raise ValueError("Must specify file for SVGMobject")
return get_full_vector_image_path(self.file_name)
| |
72,223 | 248,340 | 42 | tests/replication/_base.py | 14 | 7 | def default_config(self) -> Dict[str, Any]:
base = super().default_config()
base["redis"] = {"enabled": True}
| Lay some foundation work to allow workers to only subscribe to some kinds of messages, reducing replication traffic. (#12672) | default_config | 177b884ad7cc1ecdd92ff74188732734df203150 | synapse | _base.py | 10 | 10 | https://github.com/matrix-org/synapse.git | 1 | 34 | 0 | 12 | 60 | Python | {
"docstring": "\n Overrides the default config to enable Redis.\n Even if the test only uses make_worker_hs, the main process needs Redis\n enabled otherwise it won't create a Fake Redis server to listen on the\n Redis port and accept fake TCP connections.\n ",
"language": "en",
... | def default_config(self) -> Dict[str, Any]:
base = super().default_config()
base["redis"] = {"enabled": True}
return base
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.