ast_errors stringlengths 0 3.2k | d_id int64 44 121k | id int64 70 338k | n_whitespaces int64 3 14k | path stringlengths 8 134 | n_words int64 4 4.82k | n_identifiers int64 1 131 | random_cut stringlengths 16 15.8k | commit_message stringlengths 2 15.3k | fun_name stringlengths 1 84 | commit_id stringlengths 40 40 | repo stringlengths 3 28 | file_name stringlengths 5 79 | ast_levels int64 6 31 | nloc int64 1 548 | url stringlengths 31 59 | complexity int64 1 66 | token_counts int64 6 2.13k | n_ast_errors int64 0 28 | vocab_size int64 4 1.11k | n_ast_nodes int64 15 19.2k | language stringclasses 1
value | documentation dict | code stringlengths 101 62.2k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
94,918 | 295,910 | 35 | tests/components/siren/test_init.py | 16 | 10 | async def test_missing_tones_list(hass):
siren = MockSirenEntity(SirenEntityFeature.TONES, ["a", "b"])
siren.hass = hass
with pytest.raises(ValueError):
process_turn_on_params(siren, {"tone": "test"})
| Add EntityFeature enum to Siren (#69585)
Co-authored-by: Franck Nijhof <frenck@frenck.nl> | test_missing_tones_list | a61ac3ddc6d65522dfa1eb599adf73420a9267dc | core | test_init.py | 12 | 5 | https://github.com/home-assistant/core.git | 1 | 43 | 0 | 15 | 80 | Python | {
"docstring": "Test ValueError when setting a tone that is missing from available_tones list.",
"language": "en",
"n_whitespaces": 11,
"n_words": 12,
"vocab_size": 12
} | async def test_missing_tones_list(hass):
siren = MockSirenEntity(SirenEntityFeature.TONES, ["a", "b"])
siren.hass = hass
with pytest.raises(ValueError):
process_turn_on_params(siren, {"tone": "test"})
| |
@cli_utils.action_cli(check_db=False)
@suppress_logs_and_warning | 8,146 | 44,090 | 109 | airflow/cli/commands/task_command.py | 52 | 28 | def task_failed_deps(args):
dag = get_dag(args.subdir, args.dag_id)
task = dag.get_task(task_id=args.task_id)
ti = _get_ti(task, args.execution_date_or_run_id, args.map_index)
dep_context = DepContext(deps=SCHEDULER_QUEUED_DEPS)
failed_deps = list(ti.get_failed_dep_statuses(dep_context=dep_con... | Add `--map-index` parameter to task CLI commands (#20980) | task_failed_deps | 8dabce8887f02216c1037be35e80c214edcbadfe | airflow | task_command.py | 14 | 12 | https://github.com/apache/airflow.git | 3 | 88 | 1 | 44 | 180 | Python | {
"docstring": "\n Returns the unmet dependencies for a task instance from the perspective of the\n scheduler (i.e. why a task instance doesn't get scheduled and then queued by the\n scheduler, and then run by an executor).\n >>> airflow tasks failed-deps tutorial sleep 2015-01-01\n Task instance depen... | def task_failed_deps(args):
dag = get_dag(args.subdir, args.dag_id)
task = dag.get_task(task_id=args.task_id)
ti = _get_ti(task, args.execution_date_or_run_id, args.map_index)
dep_context = DepContext(deps=SCHEDULER_QUEUED_DEPS)
failed_deps = list(ti.get_failed_dep_statuses(dep_context=dep_con... |
54,959 | 217,834 | 100 | python3.10.4/Lib/http/cookies.py | 24 | 9 | def load(self, rawdata):
if isinstance(rawdata, str):
| add python 3.10.4 for windows | load | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | cookies.py | 12 | 7 | https://github.com/XX-net/XX-Net.git | 3 | 42 | 0 | 23 | 70 | Python | {
"docstring": "Load cookies from a string (presumably HTTP_COOKIE) or\n from a dictionary. Loading cookies from a dictionary 'd'\n is equivalent to calling:\n map(Cookie.__setitem__, d.keys(), d.values())\n ",
"language": "en",
"n_whitespaces": 57,
"n_words": 24,
"vocab_size"... | def load(self, rawdata):
if isinstance(rawdata, str):
self.__parse_string(rawdata)
else:
# self.update() wouldn't call our custom __setitem__
for key, value in rawdata.items():
self[key] = value
return
| |
22,909 | 107,773 | 125 | lib/matplotlib/axis.py | 32 | 12 | def _reset_major_tick_kw(self, keep_tick_and_label_visibility=False):
backup = {name: value for name, value in self._major_tick_kw.items()
if name in ['tick1On', 'tick2On', 'label1On', 'label2On']}
self. | Refactor handling of tick and ticklabel visiblity in Axis.clear()
This is a follow-up to #20826, which makes the exceptions from clearing
more explicit. | _reset_major_tick_kw | 2357c92d87d96d519c8470776e76180e71663d0b | matplotlib | axis.py | 11 | 9 | https://github.com/matplotlib/matplotlib.git | 5 | 87 | 0 | 27 | 150 | Python | {
"docstring": "\n Reset major tick params to defaults.\n\n Shared subplots pre-configure tick and label visibility. To keep this\n beyond an Axis.clear() operation, we may\n *keep_tick_and_label_visibility*.\n ",
"language": "en",
"n_whitespaces": 59,
"n_words": 23,
"vocab_... | def _reset_major_tick_kw(self, keep_tick_and_label_visibility=False):
backup = {name: value for name, value in self._major_tick_kw.items()
if name in ['tick1On', 'tick2On', 'label1On', 'label2On']}
self._major_tick_kw.clear()
if keep_tick_and_label_visibility:
... | |
28,990 | 129,623 | 55 | python/ray/tune/tests/test_integration_comet.py | 13 | 12 | def test_class_variable_to_instance(self):
logger = self.logger
self.assertEqual(logger._to_exclude, logger._exclude_results)
self.assertEqual(logger._to_system, lo | Comet Integration (#20766)
This PR adds a `CometLoggerCallback` to the Tune Integrations, allowing users to log runs from Ray to [Comet](https://www.comet.ml/site/).
Co-authored-by: Michael Cullan <mjcullan@gmail.com>
Co-authored-by: Antoni Baum <antoni.baum@protonmail.com> | test_class_variable_to_instance | 3d79815cd08c1be8e56c245e662f34366523847e | ray | test_integration_comet.py | 8 | 6 | https://github.com/ray-project/ray.git | 1 | 59 | 0 | 13 | 93 | Python | {
"docstring": "Test that class variables get properly assigned to instance\n variables.\n ",
"language": "en",
"n_whitespaces": 24,
"n_words": 10,
"vocab_size": 10
} | def test_class_variable_to_instance(self):
logger = self.logger
self.assertEqual(logger._to_exclude, logger._exclude_results)
self.assertEqual(logger._to_system, logger._system_results)
self.assertEqual(logger._to_other, logger._other_results)
self.assertEqual(logger._to... | |
4,085 | 21,881 | 574 | pipenv/patched/pip/_vendor/chardet/__init__.py | 120 | 37 | def detect_all(byte_str, ignore_threshold=False):
if not isinstance(byte_str, bytearray):
if not isinstance(byte_str, bytes):
raise TypeError(
f"Expected object of type bytes or bytearray, got: {type(byte_str)}"
)
b | Rename notpip to pip. Vendor in pip-22.2.1 and latest requirementslib and vistir. | detect_all | cd5a9683be69c86c8f3adcd13385a9bc5db198ec | pipenv | __init__.py | 17 | 36 | https://github.com/pypa/pipenv.git | 14 | 219 | 0 | 88 | 372 | Python | {
"docstring": "\n Detect all the possible encodings of the given byte string.\n\n :param byte_str: The byte sequence to examine.\n :type byte_str: ``bytes`` or ``bytearray``\n :param ignore_threshold: Include encodings that are below\n ``UniversalDetector.... | def detect_all(byte_str, ignore_threshold=False):
if not isinstance(byte_str, bytearray):
if not isinstance(byte_str, bytes):
raise TypeError(
f"Expected object of type bytes or bytearray, got: {type(byte_str)}"
)
byte_str = bytearray(byte_str)
detec... | |
54,715 | 217,317 | 184 | python3.10.4/Lib/enum.py | 39 | 10 | def __call__(cls, value, names=None, *, module=None, qualname=None, type=None, start=1):
if names is None: # simple value lookup
return cls.__new__(cls, value)
# otherwise, functional API: we're creating a new Enum type
return cls._create_(
value,
... | add python 3.10.4 for windows | __call__ | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | enum.py | 9 | 11 | https://github.com/XX-net/XX-Net.git | 2 | 70 | 0 | 36 | 100 | Python | {
"docstring": "\n Either returns an existing member, or creates a new enum class.\n\n This method is used both when an enum class is given a value to match\n to an enumeration member (i.e. Color(3)) and for the functional API\n (i.e. Color = Enum('Color', names='RED GREEN BLUE')).\n\n ... | def __call__(cls, value, names=None, *, module=None, qualname=None, type=None, start=1):
if names is None: # simple value lookup
return cls.__new__(cls, value)
# otherwise, functional API: we're creating a new Enum type
return cls._create_(
value,
... | |
7,567 | 42,482 | 166 | nltk/util.py | 62 | 13 | def edges2dot(edges, shapes=None, attr=None):
if not shapes:
| Fix some tests in Wordnet-related DocStrings | edges2dot | 692adaff901dd9daf29400fdf3385130aefbfb2a | nltk | util.py | 17 | 16 | https://github.com/nltk/nltk.git | 8 | 97 | 0 | 39 | 214 | Python | {
"docstring": "\n :param edges: the set (or list) of edges of a directed graph.\n\n :return dot_string: a representation of 'edges' as a string in the DOT\n graph language, which can be converted to an image by the 'dot' program\n from the Graphviz package, or nltk.parse.dependencygraph.dot2img(d... | def edges2dot(edges, shapes=None, attr=None):
if not shapes:
shapes = dict()
if not attr:
attr = dict()
dot_string = "digraph G {\n"
for pair in attr.items():
dot_string += f"{pair[0]} = {pair[1]};\n"
for edge in edges:
for shape in shapes.items():
... | |
3,341 | 20,356 | 92 | pipenv/patched/notpip/_vendor/pygments/formatters/img.py | 23 | 10 | def _draw_line_numbers(self):
if not self.line_numbers:
return
for p in range(self.maxlineno):
n = p + self.line_number_start
if (n % self.line_number_step) == 0:
| check point progress on only bringing in pip==22.0.4 (#4966)
* vendor in pip==22.0.4
* updating vendor packaging version
* update pipdeptree to fix pipenv graph with new version of pip.
* Vendoring of pip-shims 0.7.0
* Vendoring of requirementslib 1.6.3
* Update pip index safety restrictions patch for p... | _draw_line_numbers | f3166e673fe8d40277b804d35d77dcdb760fc3b3 | pipenv | img.py | 11 | 7 | https://github.com/pypa/pipenv.git | 4 | 49 | 0 | 21 | 80 | Python | {
"docstring": "\n Create drawables for the line numbers.\n ",
"language": "en",
"n_whitespaces": 21,
"n_words": 6,
"vocab_size": 6
} | def _draw_line_numbers(self):
if not self.line_numbers:
return
for p in range(self.maxlineno):
n = p + self.line_number_start
if (n % self.line_number_step) == 0:
self._draw_linenumber(p, n)
| |
46,249 | 189,904 | 197 | manim/cli/cfg/group.py | 58 | 21 | def export(ctx, directory):
directory_path = Path(directory)
if directory_path.absolute == Path.cwd().absolute:
console.print(
,
style="red bold",
end="",
)
proc | Migrate from os.path to pathlib in SVGMobject and other locations (#2687)
* fixed style
* fixed changes
* Update group.py
* Remove extra `Path` call
Co-authored-by: ad_chaos <90276965+Kiran-Raj-Dev@users.noreply.github.com>
* Remove unused context manager
Sorry, just committing here myself so that th... | export | a20f8aeb6ccd30d6b9d5c34285c3a718b0f5a59b | manim | group.py | 13 | 23 | https://github.com/ManimCommunity/manim.git | 4 | 126 | 0 | 43 | 234 | Python | {
"docstring": "You are reading the config from the same directory you are exporting to.\nThis means that the exported config will overwrite the config for this directory.\nAre you sure you want to continue? (y/n)",
"language": "en",
"n_whitespaces": 31,
"n_words": 34,
"vocab_size": 26
} | def export(ctx, directory):
directory_path = Path(directory)
if directory_path.absolute == Path.cwd().absolute:
console.print(
,
style="red bold",
end="",
)
proceed = input().lower() == "y"
else:
proceed = True
if proceed:
if no... | |
23,770 | 109,834 | 153 | lib/matplotlib/backend_managers.py | 40 | 13 | def update_keymap(self, name, key):
if name not in self._tools:
raise KeyError(f'{name!r} not in Tools')
| Add tests for ToolManager | update_keymap | 0d6ee255831adae452af355c025497c0f07aa296 | matplotlib | backend_managers.py | 15 | 11 | https://github.com/matplotlib/matplotlib.git | 5 | 70 | 0 | 31 | 135 | Python | {
"docstring": "\n Set the keymap to associate with the specified tool.\n\n Parameters\n ----------\n name : str\n Name of the Tool.\n key : str or list of str\n Keys to associate with the tool.\n ",
"language": "en",
"n_whitespaces": 96,
"n_words"... | def update_keymap(self, name, key):
if name not in self._tools:
raise KeyError(f'{name!r} not in Tools')
self._remove_keys(name)
if isinstance(key, str):
key = [key]
for k in key:
if k in self._keys:
_api.warn_external(
... | |
24,602 | 112,161 | 145 | nni/retiarii/oneshot/pytorch/supermodule/differentiable.py | 34 | 12 | def named_parameters(self, *args, **kwargs):
arch = kwargs.pop('arch', False)
for name, p in super().named_parameters(*args, **kwargs):
if any(name == par_name for par_name in self._arch_parameter_names):
if arch:
| Valuechoice oneshot lightning (#4602) | named_parameters | 14d2966b9e91ae16dcc39de8f41017a75cec8ff9 | nni | differentiable.py | 14 | 9 | https://github.com/microsoft/nni.git | 6 | 71 | 0 | 22 | 117 | Python | {
"docstring": "Named parameters excluding architecture parameters.",
"language": "en",
"n_whitespaces": 4,
"n_words": 5,
"vocab_size": 5
} | def named_parameters(self, *args, **kwargs):
arch = kwargs.pop('arch', False)
for name, p in super().named_parameters(*args, **kwargs):
if any(name == par_name for par_name in self._arch_parameter_names):
if arch:
yield name, p
else:
... | |
26,788 | 120,182 | 66 | jax/_src/util.py | 49 | 10 | def unzip3(xyzs):
# Note: we deliberately don't use zip(*xyzs) because it is lazily evaluated,
# is too permissive about inputs, and does not guarantee a length-3 output.
xs = []
ys = []
zs = []
for x, y, z in xyzs:
xs.append(x)
ys.append | Comment on implementation of unzip2 & unzip3 | unzip3 | 72470dee3a5181c8bfe0f0a4725564efbef80f92 | jax | util.py | 9 | 9 | https://github.com/google/jax.git | 2 | 60 | 0 | 43 | 101 | Python | {
"docstring": "Unzip sequence of length-3 tuples into three tuples.",
"language": "en",
"n_whitespaces": 7,
"n_words": 8,
"vocab_size": 8
} | def unzip3(xyzs):
# Note: we deliberately don't use zip(*xyzs) because it is lazily evaluated,
# is too permissive about inputs, and does not guarantee a length-3 output.
xs = []
ys = []
zs = []
for x, y, z in xyzs:
xs.append(x)
ys.append(y)
zs.append(z)
return tuple(xs), tuple(ys), tuple(z... | |
17,609 | 83,182 | 158 | zerver/tests/test_subs.py | 60 | 18 | def test_json_get_subscribers_for_guest_user(self) -> None:
guest_user = self.example_user("polonius")
never_subscribed = gather_subscriptions_helper(guest_user, True).never_subscribed
# A guest user can only see never subscribed streams that are web-public.
# For Polonius, the... | docs: Consistently hyphenate “web-public”.
In English, compound adjectives should essentially always be
hyphenated. This makes them easier to parse, especially for users who
might not recognize that the words “web public” go together as a
phrase.
Signed-off-by: Anders Kaseorg <anders@zulip.com> | test_json_get_subscribers_for_guest_user | 90e202cd38d00945c81da4730d39e3f5c5b1e8b1 | zulip | test_subs.py | 11 | 15 | https://github.com/zulip/zulip.git | 1 | 98 | 0 | 50 | 172 | Python | {
"docstring": "\n Guest users should have access to subscribers of web-public streams, even\n if they aren't subscribed or have never subscribed to that stream.\n ",
"language": "en",
"n_whitespaces": 44,
"n_words": 22,
"vocab_size": 19
} | def test_json_get_subscribers_for_guest_user(self) -> None:
guest_user = self.example_user("polonius")
never_subscribed = gather_subscriptions_helper(guest_user, True).never_subscribed
# A guest user can only see never subscribed streams that are web-public.
# For Polonius, the... | |
57,798 | 226,118 | 402 | packages/python/chart-studio/chart_studio/plotly/chunked_requests/chunked_request.py | 70 | 16 | def _reconnect(self):
if not self._isconnected():
try:
self._connec | switch to black .22 | _reconnect | 43e3a4011080911901176aab919c0ecf5046ddd3 | plotly.py | chunked_request.py | 17 | 17 | https://github.com/plotly/plotly.py.git | 6 | 95 | 0 | 53 | 166 | Python | {
"docstring": "Connect if disconnected.\n Retry self.maxtries times with delays\n ",
"language": "en",
"n_whitespaces": 22,
"n_words": 8,
"vocab_size": 8
} | def _reconnect(self):
if not self._isconnected():
try:
self._connect()
except http_client.socket.error as e:
# Attempt to reconnect if the connection was refused
if e.errno == 61 or e.errno == 10061:
# errno 61 ... | |
54,450 | 216,173 | 30 | salt/modules/cp.py | 14 | 6 | def list_master(saltenv=None, prefix=""):
if not saltenv:
| fixes saltstack/salt#61562 cp functions derive saltenv from config | list_master | 2bd6323ef5f87d871891a59917ee96f44ef55e75 | salt | cp.py | 11 | 4 | https://github.com/saltstack/salt.git | 3 | 35 | 0 | 14 | 63 | Python | {
"docstring": "\n .. versionchanged:: 3005\n ``saltenv`` will use value from config if not explicitly set\n\n List all of the files stored on the master\n\n CLI Example:\n\n .. code-block:: bash\n\n salt '*' cp.list_master\n ",
"language": "en",
"n_whitespaces": 60,
"n_words": 30,
... | def list_master(saltenv=None, prefix=""):
if not saltenv:
saltenv = __opts__["saltenv"] or "base"
return _client().file_list(saltenv, prefix)
| |
1,492 | 8,732 | 83 | tests/ludwig/utils/test_tokenizers.py | 47 | 22 | def test_bert_hf_tokenizer_parity(tmpdir, pretrained_model_name_or_path):
from ludwig.utils.tokenizers import get_hf_tokenizer, HFTokenizer
inputs = "Hello, ``I'm'' ónë of 1,205,000 sentences!"
hf_tokenizer = HFTokenizer(pretrained_model_name_or_path)
torchtext_tokenizer = get_hf_tokenizer(pretrai... | [TorchScript] Add user-defined HF Bert tokenizers (#2733)
* first working set
* wip todo: add never_split kwarg
* adds new never_split kwarg
* clean up
* get tests passing
* updated py38 tests
* pr revisions
* [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https:... | test_bert_hf_tokenizer_parity | f3fbfbbe7e4c772d60dbc4374811d3a959699f2b | ludwig | test_tokenizers.py | 10 | 11 | https://github.com/ludwig-ai/ludwig.git | 1 | 84 | 0 | 38 | 140 | Python | {
"docstring": "Tests the BERTTokenizer implementation.\n\n Asserts both tokens and token IDs are the same by initializing the BERTTokenizer as a standalone tokenizer and as a\n HF tokenizer.\n ",
"language": "en",
"n_whitespaces": 35,
"n_words": 26,
"vocab_size": 20
} | def test_bert_hf_tokenizer_parity(tmpdir, pretrained_model_name_or_path):
from ludwig.utils.tokenizers import get_hf_tokenizer, HFTokenizer
inputs = "Hello, ``I'm'' ónë of 1,205,000 sentences!"
hf_tokenizer = HFTokenizer(pretrained_model_name_or_path)
torchtext_tokenizer = get_hf_tokenizer(pretrai... | |
13,723 | 64,790 | 20 | erpnext/accounts/doctype/bank_reconciliation_tool/bank_reconciliation_tool.py | 27 | 6 | def get_pe_matching_query(amount_condition, account_from_to, transaction):
# get matching payment entries query
if transaction.deposit > 0:
currency_field = "paid_to_account_currency | style: format code with black | get_pe_matching_query | 494bd9ef78313436f0424b918f200dab8fc7c20b | erpnext | bank_reconciliation_tool.py | 10 | 28 | https://github.com/frappe/erpnext.git | 2 | 27 | 0 | 23 | 60 | Python | {
"docstring": "\n\tSELECT\n\t\t(CASE WHEN reference_no=%(reference_no)s THEN 1 ELSE 0 END\n\t\t+ CASE WHEN (party_type = %(party_type)s AND party = %(party)s ) THEN 1 ELSE 0 END\n\t\t+ 1 ) AS rank,\n\t\t'Payment Entry' as doctype,\n\t\tname,\n\t\tpaid_amount,\n\t\treference_no,\n\t\treference_date,\n\t\tparty,\n\t\... | def get_pe_matching_query(amount_condition, account_from_to, transaction):
# get matching payment entries query
if transaction.deposit > 0:
currency_field = "paid_to_account_currency as currency"
else:
currency_field = "paid_from_account_currency as currency"
return f
| |
50,325 | 203,351 | 334 | django/contrib/admin/checks.py | 60 | 23 | def _check_list_display_links(self, obj):
from django.contrib.admin.options import ModelAdmin
if obj.list_display_links is None:
return []
elif not isinstance(obj.list_display_links, (list, tuple)):
return must_be(
"a list, a tuple, or None",
... | Refs #33476 -- Reformatted code with Black. | _check_list_display_links | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | checks.py | 16 | 21 | https://github.com/django/django.git | 5 | 107 | 0 | 50 | 168 | Python | {
"docstring": "Check that list_display_links is a unique subset of list_display.",
"language": "en",
"n_whitespaces": 8,
"n_words": 9,
"vocab_size": 9
} | def _check_list_display_links(self, obj):
from django.contrib.admin.options import ModelAdmin
if obj.list_display_links is None:
return []
elif not isinstance(obj.list_display_links, (list, tuple)):
return must_be(
"a list, a tuple, or None",
... | |
89,351 | 290,233 | 131 | homeassistant/components/zha/core/channels/lighting.py | 40 | 7 | def min_mireds(self) -> int:
min_mireds = self.cluster.get("color_temp_physical_min", self.MIN_MIREDS)
if min_mireds == 0:
self.warning(
" | Fix invalid min and max color temp in bad ZHA light devices (#81604)
* Fix ZHA default color temps
* update test | min_mireds | 83c6a7e18b1b0e4d5a302e304f117dee11d3aa51 | core | lighting.py | 10 | 10 | https://github.com/home-assistant/core.git | 2 | 45 | 0 | 35 | 76 | Python | {
"docstring": "Return the coldest color_temp that this channel supports.",
"language": "en",
"n_whitespaces": 7,
"n_words": 8,
"vocab_size": 8
} | def min_mireds(self) -> int:
min_mireds = self.cluster.get("color_temp_physical_min", self.MIN_MIREDS)
if min_mireds == 0:
self.warning(
"[Min mireds is 0, setting to %s] Please open an issue on the quirks repo to have this device corrected",
self.MIN... | |
44,547 | 184,318 | 128 | src/textual/app.py | 35 | 15 | def pop_screen(self) -> Screen:
screen_stack = self._screen_stack
if len(screen_stack) <= 1:
raise ScreenStackError(
"Can't pop screen; there must be at least one screen on the stack"
)
screen = screen_stack.pop()
screen.post_me | prototype screens api | pop_screen | ff55dafb8638f6674f3662aa526a5fc35a007b24 | textual | app.py | 10 | 16 | https://github.com/Textualize/textual.git | 2 | 69 | 0 | 32 | 117 | Python | {
"docstring": "Pop the current screen from the stack, and switch to the previous screen.\n\n Returns:\n Screen: The screen that was replaced.\n ",
"language": "en",
"n_whitespaces": 45,
"n_words": 20,
"vocab_size": 17
} | def pop_screen(self) -> Screen:
screen_stack = self._screen_stack
if len(screen_stack) <= 1:
raise ScreenStackError(
"Can't pop screen; there must be at least one screen on the stack"
)
screen = screen_stack.pop()
screen.post_message_no_wa... | |
71,115 | 246,234 | 221 | tests/handlers/test_appservice.py | 39 | 25 | def test_notify_interested_services_ephemeral(self):
interested_service = self._mkservice(is_interested=True)
services = [interested_service]
self.mock_store.get_app_services.return_value = services
self.mock_store.get_type_stream_id_for_appservice.return_v | Send to-device messages to application services (#11215)
Co-authored-by: Richard van der Hoff <1389908+richvdh@users.noreply.github.com> | test_notify_interested_services_ephemeral | 64ec45fc1b0856dc7daacca7d3ab75d50bd89f84 | synapse | test_appservice.py | 11 | 22 | https://github.com/matrix-org/synapse.git | 1 | 119 | 0 | 27 | 192 | Python | {
"docstring": "\n Test sending ephemeral events to the appservice handler are scheduled\n to be pushed out to interested appservices, and that the stream ID is\n updated accordingly.\n ",
"language": "en",
"n_whitespaces": 54,
"n_words": 25,
"vocab_size": 22
} | def test_notify_interested_services_ephemeral(self):
interested_service = self._mkservice(is_interested=True)
services = [interested_service]
self.mock_store.get_app_services.return_value = services
self.mock_store.get_type_stream_id_for_appservice.return_value = make_awaitable(... | |
12,732 | 61,868 | 425 | .venv/lib/python3.8/site-packages/pip/_vendor/distlib/compat.py | 74 | 23 | def convert(self, value):
if not isinstance(value, ConvertingDict) and isinstance(value, dict):
value = ConvertingDict(value)
value.con | upd; format | convert | f638f5d0e6c8ebed0e69a6584bc7f003ec646580 | transferlearning | compat.py | 15 | 22 | https://github.com/jindongwang/transferlearning.git | 10 | 161 | 0 | 40 | 256 | Python | {
"docstring": "\n Convert values to an appropriate type. dicts, lists and tuples are\n replaced by their converting alternatives. Strings are checked to\n see if they have a conversion format and are converted if they do.\n ",
"language": "en",
"n_whitespaces": 78,
"... | def convert(self, value):
if not isinstance(value, ConvertingDict) and isinstance(value, dict):
value = ConvertingDict(value)
value.configurator = self
elif not isinstance(value, ConvertingList) and isinstance(value, list):
value = Con... | |
43,706 | 181,983 | 60 | tests/test_css_parse.py | 30 | 16 | def test_parse_transition(duration, parsed_duration):
css = f
stylesheet = Stylesheet()
stylesheet.parse(css)
rule = stylesheet.rules[0].styles
assert len(stylesheet.rules) == 1
assert len(stylesheet.rule | Stop parsing time as scalar | test_parse_transition | 644fdc7ed181a22773556236e83fb5343efe8fd5 | textual | test_css_parse.py | 12 | 13 | https://github.com/Textualize/textual.git | 1 | 80 | 0 | 24 | 130 | Python | {
"docstring": "#some-widget {{\n transition: offset {duration} in_out_cubic;\n }}\n ",
"language": "en",
"n_whitespaces": 20,
"n_words": 7,
"vocab_size": 7
} | def test_parse_transition(duration, parsed_duration):
css = f
stylesheet = Stylesheet()
stylesheet.parse(css)
rule = stylesheet.rules[0].styles
assert len(stylesheet.rules) == 1
assert len(stylesheet.rules[0].errors) == 0
assert rule.transitions == {
"offset": Transition(duration=p... | |
71,652 | 247,396 | 129 | tests/rest/media/v1/test_html_preview.py | 29 | 6 | def test_meta_charset(self) -> None:
encodings = _get_html_media_encodings(
b,
"text/html",
)
self.assertEqual(list(encodings), ["ascii", "utf-8", "cp1252"])
# A less well-form | Add type hints to `tests/rest` (#12146)
* Add type hints to `tests/rest`
* newsfile
* change import from `SigningKey` | test_meta_charset | 7e91107be1a4287873266e588a3c5b415279f4c8 | synapse | test_html_preview.py | 9 | 22 | https://github.com/matrix-org/synapse.git | 1 | 62 | 0 | 19 | 111 | Python | {
"docstring": "A character encoding is found via the meta tag.\n <html>\n <head><meta charset=\"ascii\">\n </head>\n </html>\n \n <html>\n <head>< meta charset = ascii>\n </head>\n </html>\n ",
"language": "en",
"n_whitespaces": 93,
"n_words... | def test_meta_charset(self) -> None:
encodings = _get_html_media_encodings(
b,
"text/html",
)
self.assertEqual(list(encodings), ["ascii", "utf-8", "cp1252"])
# A less well-formed version.
encodings = _get_html_media_encodings(
b,
... | |
15,407 | 70,182 | 285 | glances/amps_list.py | 69 | 17 | def _build_amps_list(self, amp_value, processlist):
ret = []
try:
# Search in both cmdline and name (for kernel thread, see #1261)
for p in processlist:
if (re.search(amp_value.regex(), p['name']) is not None) or (
p['cmdline'] is not ... | AMP: regex with special chars #2152 | _build_amps_list | 1aa5596cc25fbd74cac65c5e4d6b16bd90091138 | glances | amps_list.py | 19 | 15 | https://github.com/nicolargo/glances.git | 7 | 134 | 0 | 57 | 228 | Python | {
"docstring": "Return the AMPS process list according to the amp_value\n\n Search application monitored processes by a regular expression\n ",
"language": "en",
"n_whitespaces": 31,
"n_words": 17,
"vocab_size": 16
} | def _build_amps_list(self, amp_value, processlist):
ret = []
try:
# Search in both cmdline and name (for kernel thread, see #1261)
for p in processlist:
if (re.search(amp_value.regex(), p['name']) is not None) or (
p['cmdline'] is not ... | |
80,526 | 270,689 | 25 | keras/engine/base_layer.py | 10 | 5 | def call(self, inputs, *args, **kwargs): # pylint: disable=unused-argument
return inputs
| Reformatting the codebase with black.
PiperOrigin-RevId: 450093126 | call | 84afc5193d38057e2e2badf9c889ea87d80d8fbf | keras | base_layer.py | 6 | 2 | https://github.com/keras-team/keras.git | 1 | 16 | 0 | 10 | 27 | Python | {
"docstring": "This is where the layer's logic lives.\n\n The `call()` method may not create state (except in its first invocation,\n wrapping the creation of variables or other resources in `tf.init_scope()`).\n It is recommended to create state in `__init__()`, or the `build()` method\n ... | def call(self, inputs, *args, **kwargs): # pylint: disable=unused-argument
return inputs
| |
25,676 | 116,155 | 494 | tests/unit/test_executor.py | 144 | 31 | def test_use_predictor_with_view(self, mock_handler):
# set integration data
df = pd.DataFrame([
{'a': 1, 'b': 'one'},
{'a': 2, 'b': 'two'},
{'a': 1, 'b': 'three'},
])
self.set_handler(mock_handler, name='pg', tables={'tasks': df})
view_name ... | executor tests | test_use_predictor_with_view | 02a831997cdffafca7cb160eb1938e72020ee049 | mindsdb | test_executor.py | 13 | 39 | https://github.com/mindsdb/mindsdb.git | 1 | 254 | 0 | 90 | 442 | Python | {
"docstring": "\n select task_model.p \n from views.{view_name}\n join mindsdb.task_model\n where {view_name}.a = 2\n ",
"language": "en",
"n_whitespaces": 59,
"n_words": 10,
"vocab_size": 10
} | def test_use_predictor_with_view(self, mock_handler):
# set integration data
df = pd.DataFrame([
{'a': 1, 'b': 'one'},
{'a': 2, 'b': 'two'},
{'a': 1, 'b': 'three'},
])
self.set_handler(mock_handler, name='pg', tables={'tasks': df})
view_name ... | |
28,017 | 125,895 | 421 | rllib/connectors/tests/test_agent.py | 114 | 31 | def test_vr_connector_shift_by_one(self):
view_rq_dict = {
"state": ViewRequirement("obs"),
"next_state": ViewRequirement(
"obs", shift=1, used_for_compute_actions=False
),
"prev_state": ViewRequirement("obs", shift=-1),
}
... | [RLlib] Implemented ViewRequirementConnector (#26998) | test_vr_connector_shift_by_one | 8ddcf89096e5631c6b6e0d04dc094b458a15c9f9 | ray | test_agent.py | 15 | 26 | https://github.com/ray-project/ray.git | 3 | 187 | 0 | 87 | 308 | Python | {
"docstring": "Test that the ViewRequirementConnector can handle shift by one correctly and\n can ignore future referencing view_requirements to respect causality",
"language": "en",
"n_whitespaces": 25,
"n_words": 19,
"vocab_size": 18
} | def test_vr_connector_shift_by_one(self):
view_rq_dict = {
"state": ViewRequirement("obs"),
"next_state": ViewRequirement(
"obs", shift=1, used_for_compute_actions=False
),
"prev_state": ViewRequirement("obs", shift=-1),
}
... | |
18,743 | 91,232 | 127 | tools/flake8_plugin.py | 85 | 19 | def adapt_error(cls, e):
return e._replace(message=e.message.format(*e.vars))[:4]
error = namedtuple("error", "lineno col message type vars")
Error = partial(partial, error, message="", type=SentryCheck, vars=())
S001 = Error(
message="S001: Avoid us | Revert "ref: simplify and type flake8 plugin (#35645)" (#35651) | adapt_error | 8b9bcdc92d8ff23ec9f44d90d14348d9464d476b | sentry | flake8_plugin.py | 13 | 2 | https://github.com/getsentry/sentry.git | 1 | 31 | 0 | 75 | 227 | Python | {
"docstring": "Adapts the extended error namedtuple to be compatible with Flake8.",
"language": "en",
"n_whitespaces": 9,
"n_words": 10,
"vocab_size": 10
} | def adapt_error(cls, e):
return e._replace(message=e.message.format(*e.vars))[:4]
error = namedtuple("error", "lineno col message type vars")
Error = partial(partial, error, message="", type=SentryCheck, vars=())
S001 = Error(
message="S001: Avoid using the {} mock call as it is "
"confusing... | |
80,335 | 269,925 | 43 | keras/callbacks.py | 11 | 6 | def on_train_begin(self, logs=None):
logs = self._process_logs(logs)
for callback in self.callbacks:
callback.on_train_begin(logs)
| Reformatting the codebase with black.
PiperOrigin-RevId: 450093126 | on_train_begin | 84afc5193d38057e2e2badf9c889ea87d80d8fbf | keras | callbacks.py | 9 | 4 | https://github.com/keras-team/keras.git | 2 | 31 | 0 | 11 | 51 | Python | {
"docstring": "Calls the `on_train_begin` methods of its callbacks.\n\n Args:\n logs: Dict. Currently, no data is passed via this argument\n for this method, but that may change in the future.\n ",
"language": "en",
"n_whitespaces": 66,
"n_words": 28,
"vocab_size": 26
} | def on_train_begin(self, logs=None):
logs = self._process_logs(logs)
for callback in self.callbacks:
callback.on_train_begin(logs)
| |
76,068 | 260,093 | 80 | sklearn/utils/tests/test_param_validation.py | 40 | 11 | def test_decorate_validated_function():
decorated_function = deprecated()(_func)
with pytest.warns(FutureWarning, match="Function _func is deprecated"):
decorated_function(1, 2, c=3)
# outer decorator does not inte | MNT Param validation: do not expose internal values in error msg (#23459)
* allow to not expose internal valid params in error msg
* ensure deprecated and internal do not overlap
* deprecated and internal must be subsets of options
* black | test_decorate_validated_function | 122876e9ab1ab494b4bf0ca3360d5a1527caf2e7 | scikit-learn | test_param_validation.py | 13 | 7 | https://github.com/scikit-learn/scikit-learn.git | 1 | 70 | 0 | 29 | 123 | Python | {
"docstring": "Check that validate_params functions can be decorated",
"language": "en",
"n_whitespaces": 6,
"n_words": 7,
"vocab_size": 7
} | def test_decorate_validated_function():
decorated_function = deprecated()(_func)
with pytest.warns(FutureWarning, match="Function _func is deprecated"):
decorated_function(1, 2, c=3)
# outer decorator does not interfer with validation
with pytest.warns(FutureWarning, match="Function _func... | |
74,253 | 253,816 | 48 | d2l/jax.py | 19 | 16 | def set_axes(axes, xlabel, ylabel, xlim, ylim, xscale, yscale, legend):
axes.set_xlabel(xlabel), axes.set_ylabel(ylabel)
axes.set_xscale(xscale), axes.set_yscale(yscale)
axes.set_xlim(xlim), axes.set_ylim(ylim)
if legend:
axes.legend(legend)
axes.grid()
| [Jax] Add calculus | set_axes | 7487da3edb1a68af60104e0290216f0849a8765c | d2l-en | jax.py | 9 | 7 | https://github.com/d2l-ai/d2l-en.git | 2 | 73 | 0 | 19 | 111 | Python | {
"docstring": "Set the axes for matplotlib.\n\n Defined in :numref:`sec_calculus`",
"language": "en",
"n_whitespaces": 10,
"n_words": 8,
"vocab_size": 8
} | def set_axes(axes, xlabel, ylabel, xlim, ylim, xscale, yscale, legend):
axes.set_xlabel(xlabel), axes.set_ylabel(ylabel)
axes.set_xscale(xscale), axes.set_yscale(yscale)
axes.set_xlim(xlim), axes.set_ylim(ylim)
if legend:
axes.legend(legend)
axes.grid()
| |
73,504 | 250,551 | 397 | mitmproxy/addonmanager.py | 91 | 27 | def register(self, addon):
api_changes = {
# mitmproxy 6 -> mitmproxy 7
"clientconnect": "client_connected",
"clientdisconnect": "client_disconnected",
"serverconnect": "server_connect and server_connected",
"serverdisconnect": "server_disconn... | Rename new async helper functions.
async_trigger -> trigger_event
invoke_addon -> invoke_addon_sync (API breakage)
async_invoke_addon -> invoke_addon | register | ee4999e8e4380f7b67faef92f04c361deffba412 | mitmproxy | addonmanager.py | 16 | 26 | https://github.com/mitmproxy/mitmproxy.git | 7 | 164 | 0 | 68 | 283 | Python | {
"docstring": "\n Register an addon, call its load event, and then register all its\n sub-addons. This should be used by addons that dynamically manage\n addons.\n\n If the calling addon is already running, it should follow with\n running and configure events. M... | def register(self, addon):
api_changes = {
# mitmproxy 6 -> mitmproxy 7
"clientconnect": "client_connected",
"clientdisconnect": "client_disconnected",
"serverconnect": "server_connect and server_connected",
"serverdisconnect": "server_disconn... | |
@method_decorator(never_cache, name='dispatch') | 45,927 | 188,689 | 292 | apps/authentication/views/login.py | 74 | 41 | def get_context_data(self, **kwargs):
from tickets.models import Ticket
from tickets.const import TICKET_DETAIL_URL
ticket_id = self.request.session.get("auth_ticket_id")
if not ticket_id:
| fix: login confirm bug (#7914)
Co-authored-by: feng626 <1304903146@qq.com> | get_context_data | 08ff8fa285575b8ca5ee187d297d807bf197a161 | jumpserver | login.py | 15 | 26 | https://github.com/jumpserver/jumpserver.git | 4 | 180 | 1 | 52 | 320 | Python | {
"docstring": "Wait for <b>{}</b> confirm, You also can copy link to her/him <br/>\n Don't close this page",
"language": "en",
"n_whitespaces": 32,
"n_words": 16,
"vocab_size": 16
} | def get_context_data(self, **kwargs):
from tickets.models import Ticket
from tickets.const import TICKET_DETAIL_URL
ticket_id = self.request.session.get("auth_ticket_id")
if not ticket_id:
ticket = None
else:
ticket = Ticket.all().filter(pk=ticket_id).firs... |
5,748 | 31,457 | 319 | src/transformers/modeling_tf_utils.py | 137 | 29 | def tf_shard_checkpoint(weights, max_shard_size="10GB"):
max_shard_size = convert_file_size_to_int(max_shard_size)
sharded_state_dicts = []
current_block = []
current_block_size = 0
total_size = 0
for item in weights:
weight_size = item.numpy().size * dtype_byte_size(item.dtype)
... | TF Sharded (#17713)
* initial commit
* update modeeling tf utils
* quality
* clean and update args
* update
* remove potential bug
* code quality
* update
* update max shard
* update tests for sharding from pretrained
* fix remaining test
* make style
* h5py if tf available
* upd... | tf_shard_checkpoint | 7cced021fa8ddc59f0f77384300760d34545394e | transformers | modeling_tf_utils.py | 14 | 29 | https://github.com/huggingface/transformers.git | 6 | 181 | 0 | 83 | 324 | Python | {
"docstring": "\n Splits a model state dictionary in sub-checkpoints so that the final size of each sub-checkpoint does not exceed a\n given size.\n\n The sub-checkpoints are determined by iterating through the `state_dict` in the order of its keys, so there is no\n optimization made to make each sub-che... | def tf_shard_checkpoint(weights, max_shard_size="10GB"):
max_shard_size = convert_file_size_to_int(max_shard_size)
sharded_state_dicts = []
current_block = []
current_block_size = 0
total_size = 0
for item in weights:
weight_size = item.numpy().size * dtype_byte_size(item.dtype)
... | |
47,761 | 196,261 | 72 | sympy/geometry/curve.py | 22 | 14 | def scale(self, x=1, y=1, pt=None):
if pt:
pt = Point(pt, dim=2)
return self.translate(*(-pt).args).scale(x, y).translate(*pt.args)
fx, | Updated import locations | scale | 498015021131af4dbb07eb110e5badaba8250c7b | sympy | curve.py | 17 | 6 | https://github.com/sympy/sympy.git | 2 | 85 | 0 | 20 | 130 | Python | {
"docstring": "Override GeometryEntity.scale since Curve is not made up of Points.\n\n Returns\n =======\n\n Curve :\n returns scaled curve.\n\n Examples\n ========\n\n >>> from sympy import Curve\n >>> from sympy.abc import x\n >>> Curve((x, x), (x,... | def scale(self, x=1, y=1, pt=None):
if pt:
pt = Point(pt, dim=2)
return self.translate(*(-pt).args).scale(x, y).translate(*pt.args)
fx, fy = self.functions
return self.func((fx*x, fy*y), self.limits)
| |
48,536 | 197,428 | 740 | sympy/physics/vector/frame.py | 217 | 58 | def orient_body_fixed(self, parent, angles, rotation_order):
_check_frame(parent)
amounts = list(angles)
for i, v in enumerate(amounts):
if not isinstance(v, Vector):
amounts[i] = sympify(v)
approved_orders = ('123', '231', '312', '132', '213', '32... | Restores pre-1.7 simplify behavior for orient_body_fixed() | orient_body_fixed | 5afe37b3dee65554d7e4592c20f922cb551fde31 | sympy | frame.py | 12 | 40 | https://github.com/sympy/sympy.git | 7 | 394 | 0 | 155 | 626 | Python | {
"docstring": "Rotates this reference frame relative to the parent reference frame\n by right hand rotating through three successive body fixed simple axis\n rotations. Each subsequent axis of rotation is about the \"body fixed\"\n unit vectors of a new intermediate reference frame. This type of... | def orient_body_fixed(self, parent, angles, rotation_order):
_check_frame(parent)
amounts = list(angles)
for i, v in enumerate(amounts):
if not isinstance(v, Vector):
amounts[i] = sympify(v)
approved_orders = ('123', '231', '312', '132', '213', '32... | |
14,124 | 66,175 | 42 | erpnext/hr/doctype/leave_block_list/leave_block_list.py | 66 | 18 | def get_applicable_block_lists(employee=None, company=None, all_lists=False):
block_lists = []
if not employee:
employee = frappe.db.get_value("Employee", {"user_id": frappe.session.user})
if not employee:
return []
if not company:
company = frappe.db.get_value("Employee", employee, "company")
def add_b... | style: format code with black | get_applicable_block_lists | 494bd9ef78313436f0424b918f200dab8fc7c20b | erpnext | leave_block_list.py | 14 | 20 | https://github.com/frappe/erpnext.git | 6 | 132 | 0 | 43 | 257 | Python | {
"docstring": "select name from `tabLeave Block List`\n\t\twhere applies_to_all_departments=1 and company=%s",
"language": "en",
"n_whitespaces": 8,
"n_words": 10,
"vocab_size": 10
} | def get_applicable_block_lists(employee=None, company=None, all_lists=False):
block_lists = []
if not employee:
employee = frappe.db.get_value("Employee", {"user_id": frappe.session.user})
if not employee:
return []
if not company:
company = frappe.db.get_value("Employee", employee, "company")
def add_b... | |
25,164 | 114,363 | 79 | mindsdb/integrations/libs/storage_handler.py | 20 | 9 | def _setup_connection(self):
# noqa
cur = self.connection.cursor()
if ('store',) not in lis | feat: add docs, improve base class signatures | _setup_connection | 27a34a6a706a06e1241671d29c8cab93d77a19c1 | mindsdb | storage_handler.py | 11 | 6 | https://github.com/mindsdb/mindsdb.git | 2 | 45 | 0 | 20 | 83 | Python | {
"docstring": " Checks that a key-value table exists, otherwise creates it. create table store (key text, value text)",
"language": "en",
"n_whitespaces": 16,
"n_words": 16,
"vocab_size": 15
} | def _setup_connection(self):
# noqa
cur = self.connection.cursor()
if ('store',) not in list(cur.execute("SELECT name FROM sqlite_master WHERE type='table';")):
cur.execute(
)
self.internal_registry.commit()
| |
36,574 | 156,129 | 214 | dask/optimization.py | 64 | 20 | def cull(dsk, keys):
if not isinstance(keys, (lis | absolufy-imports - No relative - PEP8 (#8796)
Conversation in https://github.com/dask/distributed/issues/5889 | cull | cccb9d8d8e33a891396b1275c2448c352ef40c27 | dask | optimization.py | 15 | 19 | https://github.com/dask/dask.git | 6 | 121 | 0 | 44 | 193 | Python | {
"docstring": "Return new dask with only the tasks required to calculate keys.\n\n In other words, remove unnecessary tasks from dask.\n ``keys`` may be a single key or list of keys.\n\n Examples\n --------\n >>> def inc(x):\n ... return x + 1\n\n >>> def add(x, y):\n ... return x + y... | def cull(dsk, keys):
if not isinstance(keys, (list, set)):
keys = [keys]
seen = set()
dependencies = dict()
out = {}
work = list(set(flatten(keys)))
while work:
new_work = []
for k in work:
dependencies_k = get_dependencies(dsk, k, as_list=True) # fuse... | |
34,339 | 148,815 | 92 | freqtrade/exchange/exchange.py | 23 | 11 | def fill_leverage_tiers(self) -> None:
leverage_tiers = self.load_leverage_tiers()
for pair, tiers in leverage_tiers.items():
tiers = []
| freqtrade.exchange edited load_leverage_tiers | fill_leverage_tiers | 41d8330fbc95224020a046bd46eea6252374ee15 | freqtrade | exchange.py | 13 | 11 | https://github.com/freqtrade/freqtrade.git | 3 | 54 | 0 | 17 | 89 | Python | {
"docstring": "\n Assigns property _leverage_tiers to a dictionary of information about the leverage\n allowed on each pair\n ",
"language": "en",
"n_whitespaces": 37,
"n_words": 15,
"vocab_size": 15
} | def fill_leverage_tiers(self) -> None:
leverage_tiers = self.load_leverage_tiers()
for pair, tiers in leverage_tiers.items():
tiers = []
for tier in tiers:
tiers.append(self.parse_leverage_tier(tier))
self._leverage_tiers[pair] = tiers
| |
47,889 | 196,389 | 57 | sympy/matrices/expressions/kronecker.py | 21 | 7 | def kronecker_product(*matrices):
if not matrices: | Moved imports to higher level | kronecker_product | 59d22b6bb7287613d598611027f640d068ca5748 | sympy | kronecker.py | 13 | 8 | https://github.com/sympy/sympy.git | 3 | 46 | 0 | 19 | 82 | Python | {
"docstring": "\n The Kronecker product of two or more arguments.\n\n This computes the explicit Kronecker product for subclasses of\n ``MatrixBase`` i.e. explicit matrices. Otherwise, a symbolic\n ``KroneckerProduct`` object is returned.\n\n\n Examples\n ========\n\n For ``MatrixSymbol`` argume... | def kronecker_product(*matrices):
if not matrices:
raise TypeError("Empty Kronecker product is undefined")
validate(*matrices)
if len(matrices) == 1:
return matrices[0]
else:
return KroneckerProduct(*matrices).doit()
| |
50,364 | 203,419 | 97 | django/contrib/admin/options.py | 33 | 18 | def _get_obj_does_not_exist_redirect(self, request, opts, object_id):
msg = _("%(name)s with ID “%(key)s” doesn’t exist. Perhaps it was deleted?") % {
"name": opts.verbose_name,
"key": unquote(object_id),
}
self.message_user(request, | Refs #33476 -- Reformatted code with Black. | _get_obj_does_not_exist_redirect | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | options.py | 11 | 8 | https://github.com/django/django.git | 1 | 65 | 0 | 32 | 106 | Python | {
"docstring": "\n Create a message informing the user that the object doesn't exist\n and return a redirect to the admin index page.\n ",
"language": "en",
"n_whitespaces": 42,
"n_words": 20,
"vocab_size": 17
} | def _get_obj_does_not_exist_redirect(self, request, opts, object_id):
msg = _("%(name)s with ID “%(key)s” doesn’t exist. Perhaps it was deleted?") % {
"name": opts.verbose_name,
"key": unquote(object_id),
}
self.message_user(request, msg, messages.WARNING)
... | |
55,829 | 219,816 | 999 | python3.10.4/Lib/_pydecimal.py | 183 | 19 | def compare_total(self, other, context=None):
other = _convert_other(other, raiseit=True)
# if one is negative and the other is positive, it's easy
if self._sign and not other._sign:
return _NegativeOne
if not self._sign and other._sign:
return _One
... | add python 3.10.4 for windows | compare_total | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | _pydecimal.py | 14 | 57 | https://github.com/XX-net/XX-Net.git | 27 | 242 | 0 | 61 | 391 | Python | {
"docstring": "Compares self to other using the abstract representations.\n\n This is not like the standard compare, which use their numerical\n value. Note that a total ordering is defined for all possible abstract\n representations.\n ",
"language": "en",
"n_whitespaces": 60,
"n_w... | def compare_total(self, other, context=None):
other = _convert_other(other, raiseit=True)
# if one is negative and the other is positive, it's easy
if self._sign and not other._sign:
return _NegativeOne
if not self._sign and other._sign:
return _One
... | |
69,581 | 241,553 | 161 | pytorch_lightning/utilities/enums.py | 61 | 11 | def detect_current_mode(cls) -> _FaultTolerantMode:
env_value = os.getenv("PL_FAULT_TOLERANT_TRAINING", "0").lower()
# the int values are kept for backwards compatibility, but long-term we want to keep only the strings
if env_value in ("0", "disabled"):
return _FaultT | Add typing for utilities/enums.py (#11298) | detect_current_mode | a610e043d797ca0bae1ce186829fece79077407a | lightning | enums.py | 11 | 12 | https://github.com/Lightning-AI/lightning.git | 4 | 66 | 0 | 52 | 122 | Python | {
"docstring": "This classmethod detects if `Fault Tolerant` is activated and maps its value to `_FaultTolerantMode`.",
"language": "en",
"n_whitespaces": 13,
"n_words": 14,
"vocab_size": 14
} | def detect_current_mode(cls) -> _FaultTolerantMode:
env_value = os.getenv("PL_FAULT_TOLERANT_TRAINING", "0").lower()
# the int values are kept for backwards compatibility, but long-term we want to keep only the strings
if env_value in ("0", "disabled"):
return _FaultTolerant... | |
@keras_export("keras.preprocessing.image.random_channel_shift") | 81,449 | 275,711 | 71 | keras/preprocessing/image.py | 41 | 16 | def apply_channel_shift(x, intensity, channel_axis=0):
x = np.rollaxis(x, channel_axis, 0)
min_x, max_x = np.min(x), np.max(x)
channel_images = [
np.clip(x_channel + intensity, min_x, max_x) for x_channel in x
]
x = np.stack(channel_imag | Reformatting the codebase with black.
PiperOrigin-RevId: 450093126 | apply_channel_shift | 84afc5193d38057e2e2badf9c889ea87d80d8fbf | keras | image.py | 10 | 9 | https://github.com/keras-team/keras.git | 2 | 89 | 1 | 29 | 144 | Python | {
"docstring": "Performs a channel shift.\n\n Args:\n x: Input tensor. Must be 3D.\n intensity: Transformation intensity.\n channel_axis: Index of axis for channels in the input tensor.\n\n Returns:\n Numpy image tensor.\n ",
"language": "en",
"n_whitespaces": 65,
"n_words":... | def apply_channel_shift(x, intensity, channel_axis=0):
x = np.rollaxis(x, channel_axis, 0)
min_x, max_x = np.min(x), np.max(x)
channel_images = [
np.clip(x_channel + intensity, min_x, max_x) for x_channel in x
]
x = np.stack(channel_images, axis=0)
x = np.rollaxis(x, 0, channel_axis... |
118,377 | 323,132 | 79 | paddlenlp/trainer/trainer_base.py | 22 | 7 | def _nested_gather(self, tensors, name=None):
if tensors is None:
return
if self.ar | [Trainer] Add init version of paddlenlp trainer and apply finetune for ernie-1.0 pretraining. (#1761)
* add some datasets for finetune.
* support fine tune for all tastks.
* add trainer prototype.
* init verison for paddlenlp trainer.
* refine trainer.
* update for some details.
* support multi-card... | _nested_gather | 44a290e94d1becd1f09fddc3d873f9e19c9d6919 | PaddleNLP | trainer_base.py | 10 | 6 | https://github.com/PaddlePaddle/PaddleNLP.git | 3 | 36 | 0 | 18 | 60 | Python | {
"docstring": "\n Gather value of `tensors` (tensor or list/tuple of nested tensors) and convert them to numpy before\n concatenating them to `gathered`\n ",
"language": "en",
"n_whitespaces": 42,
"n_words": 20,
"vocab_size": 17
} | def _nested_gather(self, tensors, name=None):
if tensors is None:
return
if self.args.local_rank != -1:
tensors = distributed_concat(tensors)
return tensors
# Copied from Accelerate. | |
@pytest.mark.parametrize("solver", SOLVERS)
@pytest.mark.parametrize("fit_intercept", [True, False]) | 76,234 | 260,411 | 595 | sklearn/linear_model/_glm/tests/test_glm.py | 269 | 37 | def test_glm_regression_unpenalized(solver, fit_intercept, glm_dataset):
model, X, y, coef, _, _, _ = glm_dataset
n_samples, n_features = X.shape
alpha = 0 # unpenalized
params = dict(
alpha=alpha,
fit_intercept=fit_intercept,
# solver=solver, # only lbfgs available
... | TST tight tests for GLMs (#23619)
Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org> | test_glm_regression_unpenalized | 9d863aba2b6dab9c9cbbcf2f7c3b7a99b6ad168f | scikit-learn | test_glm.py | 15 | 31 | https://github.com/scikit-learn/scikit-learn.git | 4 | 241 | 1 | 170 | 413 | Python | {
"docstring": "Test that unpenalized GLM converges for all solvers to correct solution.\n\n We work with a simple constructed data set with known solution.\n Note: This checks the minimum norm solution for wide X, i.e.\n n_samples < n_features:\n min ||w||_2 subject to w = argmin deviance(X, y, w)\n ... | def test_glm_regression_unpenalized(solver, fit_intercept, glm_dataset):
model, X, y, coef, _, _, _ = glm_dataset
n_samples, n_features = X.shape
alpha = 0 # unpenalized
params = dict(
alpha=alpha,
fit_intercept=fit_intercept,
# solver=solver, # only lbfgs available
... |
42,392 | 177,486 | 55 | networkx/algorithms/bipartite/redundancy.py | 33 | 11 | def _node_redundancy(G, v):
n = len(G[v])
overlap = sum(
1 for (u, w) in combinations(G[v], 2) if (G[u].keys() & G[w].keys()) - {v} | Minor Python 2 cleanup (#6219)
Python3 cleanup
Use dict.keys() for set operations rather than explicitly
creating sets. | _node_redundancy | 1f033118f2e0cca12c6e2375708dc92197b62da6 | networkx | redundancy.py | 15 | 6 | https://github.com/networkx/networkx.git | 3 | 79 | 0 | 29 | 121 | Python | {
"docstring": "Returns the redundancy of the node `v` in the bipartite graph `G`.\n\n If `G` is a graph with `n` nodes, the redundancy of a node is the ratio\n of the \"overlap\" of `v` to the maximum possible overlap of `v`\n according to its degree. The overlap of `v` is the number of pairs of\n neighb... | def _node_redundancy(G, v):
n = len(G[v])
overlap = sum(
1 for (u, w) in combinations(G[v], 2) if (G[u].keys() & G[w].keys()) - {v}
)
return (2 * overlap) / (n * (n - 1))
| |
49,063 | 198,893 | 958 | sympy/physics/continuum_mechanics/truss.py | 199 | 45 | def solve(self):
count_reaction_loads = 0
for node in self._nodes:
if node in list(self._supports):
if self._supports[node[0]]=='pinned':
count_reaction_loads += 2
elif self._supports[node[0]]=='roller':
count_r... | solve method added for the truss class | solve | af847114b9138a321933574fc3c3ec73af8b3459 | sympy | truss.py | 20 | 61 | https://github.com/sympy/sympy.git | 22 | 749 | 0 | 80 | 1,153 | Python | {
"docstring": "\n This method solves for all reaction forces of all supports and all internal forces\n of all the members in the truss, provided the Truss is solvable.\n\n A Truss is solvable if the following condition is met,\n\n 2n >= r + m\n\n Where n is the number of nodes, r i... | def solve(self):
count_reaction_loads = 0
for node in self._nodes:
if node in list(self._supports):
if self._supports[node[0]]=='pinned':
count_reaction_loads += 2
elif self._supports[node[0]]=='roller':
count_r... | |
48,565 | 197,463 | 319 | sympy/polys/galoistools.py | 127 | 28 | def gf_edf_zassenhaus(f, n, p, K):
factors = [f]
if gf_degree(f) <= n:
return factors
N = gf_degree(f) // n
if p != 2:
b = gf_frobenius_monomial_base(f, p, K)
t = [K.one, K.zero]
while len(factors) < N:
if p == 2:
h = r = t
for i in range(... | Improve `gf_edf_zassenhaus()`.
For the case p == 2, we use Algorithm 3.4.8 of [Cohen93], instead of the current procedure.
The current algorithm was failing to terminate on at least one known case (factoring cyclotomic_poly(17) mod 2).
A simple bugfix would have been to change the iteration to `for i in range(n - 1)... | gf_edf_zassenhaus | d8bc197a19c0f4ea76c088da6f1655f1586cd700 | sympy | galoistools.py | 16 | 24 | https://github.com/sympy/sympy.git | 8 | 247 | 0 | 67 | 352 | Python | {
"docstring": "\n Cantor-Zassenhaus: Probabilistic Equal Degree Factorization\n\n Given a monic square-free polynomial ``f`` in ``GF(p)[x]`` and\n an integer ``n``, such that ``n`` divides ``deg(f)``, returns all\n irreducible factors ``f_1,...,f_d`` of ``f``, each of degree ``n``.\n EDF procedure giv... | def gf_edf_zassenhaus(f, n, p, K):
factors = [f]
if gf_degree(f) <= n:
return factors
N = gf_degree(f) // n
if p != 2:
b = gf_frobenius_monomial_base(f, p, K)
t = [K.one, K.zero]
while len(factors) < N:
if p == 2:
h = r = t
for i in range(... | |
41,753 | 176,187 | 168 | networkx/linalg/algebraicconnectivity.py | 89 | 23 | def _tracemin_fiedler(L, X, normalized, tol, method):
import n | Use scipy.sparse array datastructure (#5139)
* Step 1: use sparse arrays in nx.to_scipy_sparse_matrix.
Seems like a reasonable place to start.
nx.to_scipy_sparse_matrix is one of the primary interfaces to
scipy.sparse from within NetworkX.
* 1: Use np.outer instead of mult col/row vectors
Fix two instances ... | _tracemin_fiedler | 5dfd57af2a141a013ae3753e160180b82bec9469 | networkx | algebraicconnectivity.py | 14 | 42 | https://github.com/networkx/networkx.git | 7 | 412 | 0 | 61 | 178 | Python | {
"docstring": "Compute the Fiedler vector of L using the TraceMIN-Fiedler algorithm.\n\n The Fiedler vector of a connected undirected graph is the eigenvector\n corresponding to the second smallest eigenvalue of the Laplacian matrix\n of the graph. This function starts with the Laplacian L, not the Graph.\n... | def _tracemin_fiedler(L, X, normalized, tol, method):
import numpy as np
import scipy as sp
import scipy.linalg # call as sp.linalg
import scipy.linalg.blas # call as sp.linalg.blas
import scipy.sparse # call as sp.sparse
n = X.shape[0]
if normalized:
# Form the normalized ... | |
@pytest.mark.skipif(
not os.environ.get("OPENAI_API_KEY", None),
reason="Please export an env var called OPENAI_API_KEY containing the OpenAI API key to run this test.",
) | 75,229 | 258,374 | 98 | test/nodes/test_prompt_node.py | 55 | 22 | def test_complex_pipeline_with_shared_prompt_model_and_prompt_template_yaml(tmp_path):
with open(tmp_path / "tmp_config_with_prompt_template.yml", "w") as tmp_file:
tmp_file.write(
f
)
pipeline = Pipeline.load_from_yam | feat: Expand LLM support with PromptModel, PromptNode, and PromptTemplate (#3667)
Co-authored-by: ZanSara <sarazanzo94@gmail.com> | test_complex_pipeline_with_shared_prompt_model_and_prompt_template_yaml | 9ebf164cfdfb320503b7161493420c1b0ec577a3 | haystack | test_prompt_node.py | 13 | 43 | https://github.com/deepset-ai/haystack.git | 1 | 78 | 1 | 50 | 181 | Python | {
"docstring": "\n version: ignore\n components:\n - name: pmodel\n type: PromptModel\n params:\n model_name_or_path: google/flan-t5-small\n model_kwargs:\n torch_dtype: torch.bfloat16\n - name: questi... | def test_complex_pipeline_with_shared_prompt_model_and_prompt_template_yaml(tmp_path):
with open(tmp_path / "tmp_config_with_prompt_template.yml", "w") as tmp_file:
tmp_file.write(
f
)
pipeline = Pipeline.load_from_yaml(path=tmp_path / "tmp_config_with_prompt_template.yml")
resul... |
19,851 | 100,362 | 595 | lib/utils.py | 109 | 33 | def _download_model(self):
self.logger.info("Downloading model: '%s' from: %s", self._model_name, self._url_download)
for attempt in range(self._retries):
try:
downloaded_size = | Update code to support Tensorflow versions up to 2.8 (#1213)
* Update maximum tf version in setup + requirements
* - bump max version of tf version in launcher
- standardise tf version check
* update keras get_custom_objects for tf>2.6
* bugfix: force black text in GUI file dialogs (linux)
* dssim loss -... | _download_model | c1512fd41d86ef47a5d1ce618d6d755ef7cbacdf | faceswap | utils.py | 17 | 26 | https://github.com/deepfakes/faceswap.git | 5 | 220 | 0 | 89 | 366 | Python | {
"docstring": " Download the model zip from github to the cache folder. ",
"language": "en",
"n_whitespaces": 11,
"n_words": 10,
"vocab_size": 9
} | def _download_model(self):
self.logger.info("Downloading model: '%s' from: %s", self._model_name, self._url_download)
for attempt in range(self._retries):
try:
downloaded_size = self._url_partial_size
req = urllib.request.Request(self._url_download)
... | |
76,333 | 260,546 | 36 | sklearn/manifold/_locally_linear.py | 8 | 7 | def fit_transform(self, X, y=None):
self._validate_params()
self._fit_transform(X)
| MAINT Use _validate_params in LocallyLinearEmbedding (#23938)
Co-authored-by: jeremiedbb <jeremiedbb@yahoo.fr> | fit_transform | ceeda362402bfc978bcc93d02481fe28e21a07ad | scikit-learn | _locally_linear.py | 7 | 4 | https://github.com/scikit-learn/scikit-learn.git | 1 | 27 | 0 | 8 | 45 | Python | {
"docstring": "Compute the embedding vectors for data X and transform X.\n\n Parameters\n ----------\n X : array-like of shape (n_samples, n_features)\n Training set.\n\n y : Ignored\n Not used, present here for API consistency by convention.\n\n Returns\n ... | def fit_transform(self, X, y=None):
self._validate_params()
self._fit_transform(X)
return self.embedding_
| |
5,345 | 30,144 | 169 | tests/types/test_song.py | 84 | 20 | def test_song_from_data_dump():
# Loads from str
song = Song.from_data_dump(
)
assert song.name == "Ropes"
assert song.artists == ["Dirty Palm", "Chandler Jewels"]
assert song.album_name == "Ropes"
assert song.album_artist == "Dirty Palm"
assert song.genres == ["gaming ed... | v4 init | test_song_from_data_dump | fa2ad657482aca9dc628e6d7062b8badf2706bb6 | spotify-downloader | test_song.py | 9 | 47 | https://github.com/spotDL/spotify-downloader.git | 1 | 119 | 0 | 50 | 207 | Python | {
"docstring": "\n Tests if Song.from_data_dump() works correctly.\n \n {\n \"name\": \"Ropes\",\n \"artists\": [\"Dirty Palm\", \"Chandler Jewels\"],\n \"album_name\": \"Ropes\",\n \"album_artist\": \"Dirty Palm\",\n \"genres\": [\"gaming edm\", \"m... | def test_song_from_data_dump():
# Loads from str
song = Song.from_data_dump(
)
assert song.name == "Ropes"
assert song.artists == ["Dirty Palm", "Chandler Jewels"]
assert song.album_name == "Ropes"
assert song.album_artist == "Dirty Palm"
assert song.genres == ["gaming ed... | |
77,125 | 262,100 | 482 | TTS/tts/models/vits.py | 73 | 36 | def test_run(self) -> Tuple[Dict, Dict]:
print(" | > Synthesizing test sentences.")
test_audios = {}
test_figures = {}
test_sentences = self.config.test_sentences
for idx, s_info in enumerate(test_sentences):
try:
aux_inputs = self.get_aux_inp... | Update VITS for the new API | test_run | ea965a5683c56a39570b4cc91e86cd2bb9799308 | TTS | vits.py | 22 | 35 | https://github.com/coqui-ai/TTS.git | 3 | 190 | 0 | 63 | 304 | Python | {
"docstring": "Generic test run for `tts` models used by `Trainer`.\n\n You can override this for a different behaviour.\n\n Returns:\n Tuple[Dict, Dict]: Test figures and audios to be projected to Tensorboard.\n ",
"language": "en",
"n_whitespaces": 61,
"n_words": 29,
"vocab_... | def test_run(self) -> Tuple[Dict, Dict]:
print(" | > Synthesizing test sentences.")
test_audios = {}
test_figures = {}
test_sentences = self.config.test_sentences
for idx, s_info in enumerate(test_sentences):
try:
aux_inputs = self.get_aux_inp... | |
56,469 | 221,674 | 298 | python3.10.4/Lib/configparser.py | 60 | 19 | def read_dict(self, dictionary, source='<dict>'):
elements_added = set()
for section, keys in dictionary.items():
section = str(section)
try:
self.add_section(section)
except (DuplicateSectionError, ValueError):
| add python 3.10.4 for windows | read_dict | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | configparser.py | 14 | 18 | https://github.com/XX-net/XX-Net.git | 9 | 141 | 0 | 42 | 222 | Python | {
"docstring": "Read configuration from a dictionary.\n\n Keys are section names, values are dictionaries with keys and values\n that should be present in the section. If the used dictionary type\n preserves order, sections and their keys will be added in order.\n\n All types held in the d... | def read_dict(self, dictionary, source='<dict>'):
elements_added = set()
for section, keys in dictionary.items():
section = str(section)
try:
self.add_section(section)
except (DuplicateSectionError, ValueError):
if self._strict... | |
54,501 | 216,287 | 382 | salt/channel/client.py | 67 | 22 | def send(self, load, tries=3, timeout=60, raw=False):
_try = 1
while True:
try:
if self.crypt == "clear":
log.trace("ReqChannel send clear load=%r", load)
ret = yield self._uncrypted_transfer(load, timeout=timeout)
... | Move retries to channel | send | 25e7a51c729cca539778c53f0858d6070e7d76cb | salt | client.py | 17 | 21 | https://github.com/saltstack/salt.git | 5 | 125 | 0 | 49 | 206 | Python | {
"docstring": "\n Send a request, return a future which will complete when we send the message\n\n :param dict load: A load to send across the wire\n :param int tries: The number of times to make before failure\n :param int timeout: The number of seconds on a response before failing\n ... | def send(self, load, tries=3, timeout=60, raw=False):
_try = 1
while True:
try:
if self.crypt == "clear":
log.trace("ReqChannel send clear load=%r", load)
ret = yield self._uncrypted_transfer(load, timeout=timeout)
... | |
50,438 | 203,542 | 83 | django/contrib/admin/utils.py | 25 | 13 | def get_fields_from_path(model, path):
pieces = path.split(LOOKUP_SEP)
fields = []
for piece in pieces:
if fields:
parent = get_model_from_relation(fields[-1])
else:
parent = model
fields.app | Refs #33476 -- Reformatted code with Black. | get_fields_from_path | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | utils.py | 14 | 10 | https://github.com/django/django.git | 3 | 58 | 0 | 20 | 96 | Python | {
"docstring": "Return list of Fields given path relative to model.\n\n e.g. (ModelX, \"user__groups__name\") -> [\n <django.db.models.fields.related.ForeignKey object at 0x...>,\n <django.db.models.fields.related.ManyToManyField object at 0x...>,\n <django.db.models.fields.CharField object at... | def get_fields_from_path(model, path):
pieces = path.split(LOOKUP_SEP)
fields = []
for piece in pieces:
if fields:
parent = get_model_from_relation(fields[-1])
else:
parent = model
fields.append(parent._meta.get_field(piece))
return fields
| |
56,788 | 222,870 | 310 | python3.10.4/Lib/distutils/dist.py | 92 | 23 | def find_config_files(self):
files = []
check_environ()
# Where to look for the system-wide Distutils config file
sys_dir = os.path.dirname(sys.modules['distutils'].__file__)
# Look for the system config file
sys_file = os.path.join(sys_dir, "distutils.cfg")
... | add python 3.10.4 for windows | find_config_files | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | dist.py | 13 | 21 | https://github.com/XX-net/XX-Net.git | 7 | 150 | 0 | 61 | 267 | Python | {
"docstring": "Find as many configuration files as should be processed for this\n platform, and return a list of filenames in the order in which they\n should be parsed. The filenames returned are guaranteed to exist\n (modulo nasty race conditions).\n\n There are three possible config f... | def find_config_files(self):
files = []
check_environ()
# Where to look for the system-wide Distutils config file
sys_dir = os.path.dirname(sys.modules['distutils'].__file__)
# Look for the system config file
sys_file = os.path.join(sys_dir, "distutils.cfg")
... | |
10,339 | 51,516 | 87 | modules/image/Image_gan/gan/stgan_bald/processor.py | 42 | 15 | def get_save_image_name(org_im_path, output_dir, num):
# name prefix of orginal image
org_im_name = os.path.split(org_im_path)[-1]
im_prefix = os.path.splitext(org_im_name)[0]
ext = '.png'
# save image path
save_im_path = os.path.join(output_dir, im_prefix + ext)
if os.path.exists(sav | update stgan_bald (#2022) | get_save_image_name | 02d7e5514b0da9a7ebabb004533b274056c954e2 | PaddleHub | processor.py | 14 | 9 | https://github.com/PaddlePaddle/PaddleHub.git | 2 | 85 | 0 | 28 | 137 | Python | {
"docstring": "\n Get save image name from source image path.\n ",
"language": "en",
"n_whitespaces": 15,
"n_words": 8,
"vocab_size": 7
} | def get_save_image_name(org_im_path, output_dir, num):
# name prefix of orginal image
org_im_name = os.path.split(org_im_path)[-1]
im_prefix = os.path.splitext(org_im_name)[0]
ext = '.png'
# save image path
save_im_path = os.path.join(output_dir, im_prefix + ext)
if os.path.exists(save_... | |
24,007 | 110,265 | 310 | lib/matplotlib/colors.py | 175 | 24 | def rgb_to_hsv(arr):
arr = np.asarray(arr)
# check length of the last dimension, should be _some_ sort of rgb
if arr.shape[-1] != 3:
raise ValueError("Last dimension of input array must be 3; "
"shape {} was found.".format(arr.shape))
in_shape = arr.shape
arr ... | DOC: improve grammar and consistency | rgb_to_hsv | 9b6abd0b4933811e0a45c2535ab8fd107db65dd9 | matplotlib | colors.py | 13 | 28 | https://github.com/matplotlib/matplotlib.git | 2 | 308 | 0 | 95 | 452 | Python | {
"docstring": "\n Convert float RGB values (in the range [0, 1]), in a numpy array to HSV\n values.\n\n Parameters\n ----------\n arr : (..., 3) array-like\n All values must be in the range [0, 1]\n\n Returns\n -------\n (..., 3) ndarray\n Colors converted to HSV values in range [... | def rgb_to_hsv(arr):
arr = np.asarray(arr)
# check length of the last dimension, should be _some_ sort of rgb
if arr.shape[-1] != 3:
raise ValueError("Last dimension of input array must be 3; "
"shape {} was found.".format(arr.shape))
in_shape = arr.shape
arr ... | |
8,969 | 46,726 | 382 | tests/jobs/test_scheduler_job.py | 64 | 38 | def test_scheduler_verify_pool_full(self, dag_maker, configs):
with conf_vars(configs):
with dag_maker(dag_id='test_scheduler_verify_pool_full'):
BashOperator(
task_id='dummy',
pool='test_scheduler_verify_pool_full',
... | Add dag-processor cli command (#22305) | test_scheduler_verify_pool_full | f5f11aefea775448105098b59c4657fa1809fb94 | airflow | test_scheduler_job.py | 13 | 25 | https://github.com/apache/airflow.git | 1 | 173 | 0 | 49 | 285 | Python | {
"docstring": "\n Test task instances not queued when pool is full\n ",
"language": "en",
"n_whitespaces": 24,
"n_words": 9,
"vocab_size": 9
} | def test_scheduler_verify_pool_full(self, dag_maker, configs):
with conf_vars(configs):
with dag_maker(dag_id='test_scheduler_verify_pool_full'):
BashOperator(
task_id='dummy',
pool='test_scheduler_verify_pool_full',
... | |
4,481 | 22,868 | 215 | VoiceAssistant/Project_Basic_struct/textRead.py | 74 | 21 | def ms_word():
# TODO : Take location input from the user
try:
speak("Enter the document's location - ")
| VoiceAssistant
This is Voice Assistant coded using Python which can do the following: -
1. Speak Text entered by User.
2. Search anything on Google.
3. Search anything on Wikipedia.
4. Read an MS Word(docx) document.
5. Read a book(PDF).
6. Can be used as a Dictator. | ms_word | 39c49e07066b2a53e176d555af6a7bf8aabb8a9c | Python | textRead.py | 12 | 16 | https://github.com/geekcomputers/Python.git | 3 | 86 | 0 | 57 | 166 | Python | {
"docstring": "[Print and speak out a ms_word docx file as specified in the path]\r\n ",
"language": "en",
"n_whitespaces": 16,
"n_words": 13,
"vocab_size": 13
} | def ms_word():
# TODO : Take location input from the user
try:
speak("Enter the document's location - ")
location = input("Enter the document's location - ")
file_loc = doubleslash(location)
doc = docx.Document(file_loc)
fullText = []
... | |
41,792 | 176,252 | 369 | networkx/readwrite/json_graph/tree.py | 151 | 18 | def tree_data(G, root, attrs=None, ident="id", children="children"):
if G.number_of_nodes() != G.number_of_edges() + 1:
raise TypeError("G is not a tree.")
if not G.is_directed():
raise TypeError("G is not directed.")
| Add exception for unconnected graph (#5287) | tree_data | cceb43d15e1d01476c8c15ff273399dee0e3b1aa | networkx | tree.py | 11 | 31 | https://github.com/networkx/networkx.git | 6 | 167 | 0 | 104 | 247 | Python | {
"docstring": "Returns data in tree format that is suitable for JSON serialization\n and use in Javascript documents.\n\n Parameters\n ----------\n G : NetworkX graph\n G must be an oriented tree\n\n root : node\n The root of the tree\n\n attrs : dict\n A dictionary that contains... | def tree_data(G, root, attrs=None, ident="id", children="children"):
if G.number_of_nodes() != G.number_of_edges() + 1:
raise TypeError("G is not a tree.")
if not G.is_directed():
raise TypeError("G is not directed.")
if not nx.is_weakly_connected(G):
raise TypeError("G is not w... | |
24,166 | 110,450 | 150 | lib/mpl_toolkits/mplot3d/tests/test_axes3d.py | 87 | 14 | def test_mutating_input_arrays_y_and_z(fig_test, fig_ref):
ax1 = fig_test.add_subplot(111, projection='3d')
x = [1, 2, 3]
y = [0.0, 0.0, 0.0]
z = [0.0, 0.0, 0.0]
ax1.plot(x, y, z, 'o-')
ax1.set_ylim([0, 4])
ax1.set_zlim([0, 4])
fig_test.draw_without_rendering()
# mutate y,z to... | Test that plot results aren't affected by mutating input arrays | test_mutating_input_arrays_y_and_z | 7a1df7830f7685a99291d90c5e79bfc5e7876f31 | matplotlib | test_axes3d.py | 10 | 19 | https://github.com/matplotlib/matplotlib.git | 1 | 208 | 0 | 46 | 277 | Python | {
"docstring": "\n Test to see if the `z` axis does not get mutated\n after a call to `Axes3D.plot`\n\n test cases came from GH#8990\n ",
"language": "en",
"n_whitespaces": 34,
"n_words": 21,
"vocab_size": 20
} | def test_mutating_input_arrays_y_and_z(fig_test, fig_ref):
ax1 = fig_test.add_subplot(111, projection='3d')
x = [1, 2, 3]
y = [0.0, 0.0, 0.0]
z = [0.0, 0.0, 0.0]
ax1.plot(x, y, z, 'o-')
ax1.set_ylim([0, 4])
ax1.set_zlim([0, 4])
fig_test.draw_without_rendering()
# mutate y,z to... | |
16,431 | 75,623 | 184 | wagtail/search/tests/elasticsearch_common_tests.py | 40 | 20 | def test_search_with_hyphen(self):
book = models.Book.objects.create(
title="Harry Potter and the Half-Blood Prince",
publication_date=date(2009, 7, 15),
number_of_pages=607,
)
index = self.backend.get_index_for_model(models.Book)
index.add_i... | Reformat with black | test_search_with_hyphen | d10f15e55806c6944827d801cd9c2d53f5da4186 | wagtail | elasticsearch_common_tests.py | 11 | 16 | https://github.com/wagtail/wagtail.git | 2 | 93 | 0 | 32 | 148 | Python | {
"docstring": "\n This tests that punctuation characters are treated the same\n way in both indexing and querying.\n\n See: https://github.com/wagtail/wagtail/issues/937\n ",
"language": "en",
"n_whitespaces": 46,
"n_words": 17,
"vocab_size": 17
} | def test_search_with_hyphen(self):
book = models.Book.objects.create(
title="Harry Potter and the Half-Blood Prince",
publication_date=date(2009, 7, 15),
number_of_pages=607,
)
index = self.backend.get_index_for_model(models.Book)
index.add_i... | |
12,738 | 61,879 | 344 | .venv/lib/python3.8/site-packages/pip/_vendor/distlib/compat.py | 58 | 21 | def resolve(self, s):
name = s.split('.')
used = name.pop(0)
try:
found = self.importer(used)
for frag in name:
used += '.' + frag
try:
found = getattr(found, frag)
... | upd; format | resolve | f638f5d0e6c8ebed0e69a6584bc7f003ec646580 | transferlearning | compat.py | 15 | 18 | https://github.com/jindongwang/transferlearning.git | 4 | 114 | 0 | 38 | 189 | Python | {
"docstring": "\n Resolve strings to objects using standard import and attribute\n syntax.\n ",
"language": "en",
"n_whitespaces": 44,
"n_words": 10,
"vocab_size": 10
} | def resolve(self, s):
name = s.split('.')
used = name.pop(0)
try:
found = self.importer(used)
for frag in name:
used += '.' + frag
try:
found = getattr(found, frag)
... | |
76,867 | 261,569 | 38 | examples/ensemble/plot_gradient_boosting_oob.py | 19 | 16 | def heldout_score(clf, X_test, y_test):
score = np.zeros((n_estimators,), dtype=np.float64)
for i, y_pred in enumerate(clf.staged_decision_function(X_test)):
| DOC Fix FutureWarning in ensemble/plot_gradient_boosting_oob.py (#24948) | heldout_score | 2c1581c32e641e535305647eb57a1787bcf803f0 | scikit-learn | plot_gradient_boosting_oob.py | 12 | 5 | https://github.com/scikit-learn/scikit-learn.git | 2 | 59 | 0 | 17 | 91 | Python | {
"docstring": "compute deviance scores on ``X_test`` and ``y_test``.",
"language": "en",
"n_whitespaces": 6,
"n_words": 7,
"vocab_size": 7
} | def heldout_score(clf, X_test, y_test):
score = np.zeros((n_estimators,), dtype=np.float64)
for i, y_pred in enumerate(clf.staged_decision_function(X_test)):
score[i] = binomial_deviance(y_test, y_pred.ravel())
return score
| |
@pytest.mark.parametrize(
"search, expected_names",
(
("", ["The best juices", "The best beers", "The worst beers"]),
("best", ["The best juices", "The best beers"]),
("worst", ["The worst beers"]),
("average", []),
),
) | 5,214 | 29,299 | 130 | saleor/graphql/product/tests/queries/test_product_types_query.py | 72 | 17 | def test_product_types_query_ids_not_exists(user_api_client, category):
query = NOT_EXISTS_IDS_COLLECTIONS_QUERY
variables = {"filter": {"ids": ["fTEJRuFHU6fd2RU=", "2XwnQNNhwCdEjhP="]}}
response = user_api_client.post_graphql(query, variables)
content = get_graphql_content(response, ignore_errors=True)... | Split test_product.py and test_variant.py into multiple files (#11173)
* Split test_product.py into multiple files
* Split test_variant.py into multiple files | test_product_types_query_ids_not_exists | d90be220d6b687d08153934a51354011a3cb5ca1 | saleor | test_product_types_query.py | 12 | 9 | https://github.com/saleor/saleor.git | 1 | 81 | 1 | 52 | 234 | Python | {
"docstring": "\n query($filters: ProductTypeFilterInput) {\n productTypes(first: 10, filter: $filters) {\n edges {\n node {\n name\n }\n }\n }\n }\n",
"language": "en",
"n_whitespaces": 76,
"n_words": 17,
"vocab_size": 11
} | def test_product_types_query_ids_not_exists(user_api_client, category):
query = NOT_EXISTS_IDS_COLLECTIONS_QUERY
variables = {"filter": {"ids": ["fTEJRuFHU6fd2RU=", "2XwnQNNhwCdEjhP="]}}
response = user_api_client.post_graphql(query, variables)
content = get_graphql_content(response, ignore_errors=True)... |
21,945 | 104,721 | 384 | datasets/hans/hans.py | 90 | 11 | def _generate_examples(self, filepath):
for idx, line in enumerate(open(filepath, "r", encoding="utf-8")):
if idx == 0:
continue # skip header
line = line.strip()
split_line = line.split("\t")
# Examples not marked with a three out of fiv... | Make HANS dataset streamable (#4155)
* Make HANS dataset streamable
* Fix tags | _generate_examples | 0060f4c7d3f8e4fb7a3694a925ca3b7f44e1f2ea | datasets | hans.py | 12 | 20 | https://github.com/huggingface/datasets.git | 4 | 132 | 0 | 76 | 223 | Python | {
"docstring": "Generate hans examples.\n\n Args:\n filepath: a string\n\n Yields:\n dictionaries containing \"premise\", \"hypothesis\" and \"label\" strings\n ",
"language": "en",
"n_whitespaces": 54,
"n_words": 15,
"vocab_size": 15
} | def _generate_examples(self, filepath):
for idx, line in enumerate(open(filepath, "r", encoding="utf-8")):
if idx == 0:
continue # skip header
line = line.strip()
split_line = line.split("\t")
# Examples not marked with a three out of fiv... | |
87,754 | 288,598 | 241 | homeassistant/components/light/__init__.py | 72 | 15 | def _light_internal_color_mode(self) -> str:
if (color_mode := self.color_mode) is None:
# Backwards compatibility for color_mode added in 2021.4
# Add warning in 2021.6, remove in 2021.10
supported = self._light_internal_supported_color_modes
if ColorMo... | Use Kelvin as the preferred color temperature unit (#79591)
* Use Kelvin as the preferred white temperature unit
* Update homekit
* Adjust tests | _light_internal_color_mode | 47d0598e75487f63901931875f69f802a477df13 | core | __init__.py | 10 | 14 | https://github.com/home-assistant/core.git | 9 | 95 | 0 | 38 | 150 | Python | {
"docstring": "Return the color mode of the light with backwards compatibility.",
"language": "en",
"n_whitespaces": 9,
"n_words": 10,
"vocab_size": 9
} | def _light_internal_color_mode(self) -> str:
if (color_mode := self.color_mode) is None:
# Backwards compatibility for color_mode added in 2021.4
# Add warning in 2021.6, remove in 2021.10
supported = self._light_internal_supported_color_modes
if ColorMo... | |
39,609 | 164,815 | 30 | pandas/plotting/_core.py | 13 | 7 | def kde(self, bw_method=None, ind=None, **kwargs):
| DOC: fix URLs, formatting and typos (#45920) | kde | 1b5338e95917a8b94a9f7b2e1881442dd663c02d | pandas | _core.py | 9 | 2 | https://github.com/pandas-dev/pandas.git | 1 | 35 | 0 | 13 | 59 | Python | {
"docstring": "\n Generate Kernel Density Estimate plot using Gaussian kernels.\n\n In statistics, `kernel density estimation`_ (KDE) is a non-parametric\n way to estimate the probability density function (PDF) of a random\n variable. This function uses Gaussian kernels and includes autom... | def kde(self, bw_method=None, ind=None, **kwargs):
return self(kind="kde", bw_method=bw_method, ind=ind, **kwargs)
density = kde
| |
@pytest.fixture | 5,005 | 26,447 | 21 | saleor/plugins/webhook/tests/subscription_webhooks/fixtures.py | 10 | 8 | def subscription_invoice_requested_webhook(subscription_webhook):
return subscription_webhook(
INVOICE_REQUESTED_SUBSCRIPTION_QUERY, WebhookEventAsyncType.INVOICE_REQUESTED
)
INVOICE_DELETED_SUBSCRIPTION_QUERY =
@pytest | Add Webhook payload via graphql subscriptions (#9394)
* Add PoC of webhook subscriptions
* add async webhooks subscription payloads feature
* remove unneeded file
* add translations subscription handling, fixes after review
* remove todo
* add descriptions
* add descriptions, move subsrciption_payloa... | subscription_invoice_requested_webhook | aca6418d6c36956bc1ab530e6ef7e146ec9df90c | saleor | fixtures.py | 8 | 4 | https://github.com/saleor/saleor.git | 1 | 14 | 1 | 10 | 36 | Python | {
"docstring": "\n subscription{\n event{\n ...on InvoiceDeleted{\n invoice{\n id\n }\n }\n }\n }\n",
"language": "en",
"n_whitespaces": 69,
"n_words": 10,
"vocab_size": 7
} | def subscription_invoice_requested_webhook(subscription_webhook):
return subscription_webhook(
INVOICE_REQUESTED_SUBSCRIPTION_QUERY, WebhookEventAsyncType.INVOICE_REQUESTED
)
INVOICE_DELETED_SUBSCRIPTION_QUERY =
@pytest.fixture |
46,647 | 191,522 | 146 | tests/unit_tests/prompts/test_prompt.py | 45 | 11 | def test_prompt_from_examples_valid() -> None:
template =
input_variables = ["question"]
example_separator = "\n\n"
prefix =
suffix =
examples = [
,
,
]
prompt_from_examples = PromptTemplate.from_examples(
examples,
suffix,
input_variables,... | add few shot example (#148) | test_prompt_from_examples_valid | c02eb199b6587aeeb50fbb083693572bd2f030cc | langchain | test_prompt.py | 9 | 32 | https://github.com/hwchase17/langchain.git | 1 | 81 | 0 | 34 | 143 | Python | {
"docstring": "Test prompt can be successfully constructed from examples.Test Prompt:\n\nQuestion: who are you?\nAnswer: foo\n\nQuestion: what are you?\nAnswer: bar\n\nQuestion: {question}\nAnswer:Test Prompt:Question: {question}\\nAnswer:Question: who are you?\\nAnswer: fooQuestion: what are you?\\nAnswer: bar",
... | def test_prompt_from_examples_valid() -> None:
template =
input_variables = ["question"]
example_separator = "\n\n"
prefix =
suffix =
examples = [
,
,
]
prompt_from_examples = PromptTemplate.from_examples(
examples,
suffix,
input_variables,... | |
13,149 | 63,105 | 455 | .venv/lib/python3.8/site-packages/pip/_vendor/pkg_resources/__init__.py | 163 | 13 | def compatible_platforms(provided, required):
if provided is None or required is None or provided == required:
# easy case
return True
# Mac OS X special cases
reqMac = macosVersionString.match(required)
if reqMac:
provMac = macosVersionString.match(provided)
# is ... | upd; format | compatible_platforms | f638f5d0e6c8ebed0e69a6584bc7f003ec646580 | transferlearning | __init__.py | 16 | 22 | https://github.com/jindongwang/transferlearning.git | 14 | 168 | 0 | 95 | 281 | Python | {
"docstring": "Can code for the `provided` platform run on the `required` platform?\n\n Returns true if either platform is ``None``, or the platforms are equal.\n\n XXX Needs compatibility checks for Linux and other unixy OSes.\n ",
"language": "en",
"n_whitespaces": 42,
"n_words": 33,
"vocab_size":... | def compatible_platforms(provided, required):
if provided is None or required is None or provided == required:
# easy case
return True
# Mac OS X special cases
reqMac = macosVersionString.match(required)
if reqMac:
provMac = macosVersionString.match(provided)
# is ... | |
41,989 | 176,590 | 422 | networkx/algorithms/shortest_paths/weighted.py | 125 | 24 | def find_negative_cycle(G, source, weight="weight"):
weight = _weight_function(G, weight)
pred = {source: []}
v = _inner_bellman_ford(G, [source], weight, pred=pred)
if v is None:
raise nx.NetworkXError("No negative cycles detected.")
# negative cycle detected... find it
neg_cycle... | Corrected the documentation of find_negative_cycle() solving issue #5610 (#5613)
* issue
* Update branchings.py
* Update weakly_connected.py | find_negative_cycle | ec2e239764c92adf3b1abcf12817198a878d8772 | networkx | weighted.py | 17 | 31 | https://github.com/networkx/networkx.git | 9 | 221 | 0 | 83 | 358 | Python | {
"docstring": "Returns a cycle with negative total weight if it exists.\n\n Bellman-Ford is used to find shortest_paths. That algorithm\n stops if there exists a negative cycle. This algorithm\n picks up from there and returns the found negative cycle.\n\n The cycle consists of a list of nodes in the cyc... | def find_negative_cycle(G, source, weight="weight"):
weight = _weight_function(G, weight)
pred = {source: []}
v = _inner_bellman_ford(G, [source], weight, pred=pred)
if v is None:
raise nx.NetworkXError("No negative cycles detected.")
# negative cycle detected... find it
neg_cycle... | |
70,030 | 243,427 | 133 | src/PIL/ImageOps.py | 61 | 26 | def expand(image, border=0, fill=0):
left, top, right, bottom = _border(border)
width = left + image.size[0] + right
height = top + image.size[1] + bottom
color = _color(fill, | Use getpalette() in ImageOps | expand | 279ddf4ce6c76498ac29df2552a3023b9aaa76c1 | Pillow | ImageOps.py | 13 | 16 | https://github.com/python-pillow/Pillow.git | 5 | 149 | 0 | 45 | 230 | Python | {
"docstring": "\n Add border to the image\n\n :param image: The image to expand.\n :param border: Border width, in pixels.\n :param fill: Pixel fill value (a color value). Default is 0 (black).\n :return: An image.\n ",
"language": "en",
"n_whitespaces": 52,
"n_words": 32,
"vocab_size": 28... | def expand(image, border=0, fill=0):
left, top, right, bottom = _border(border)
width = left + image.size[0] + right
height = top + image.size[1] + bottom
color = _color(fill, image.mode)
if image.mode == "P" and image.palette:
palette = ImagePalette.ImagePalette(palette=image.getpalett... | |
18,085 | 86,210 | 174 | tests/sentry/integrations/slack/notifications/test_issue_alert.py | 54 | 32 | def test_digest_enabled(self, digests, mock_func):
backend = RedisBackend()
digests.digest = backend.digest
digests.enabled.return_value = True
rule = Rule.objects.create(project=self.project, label="my rule")
ProjectOwnership.objects.create(project_id=self.project.id, ... | feat(workflow): Set project ownership fallthrough default false (#39305) | test_digest_enabled | 210295c5ed1d7286ae808b15d14f6e83356af16e | sentry | test_issue_alert.py | 12 | 16 | https://github.com/getsentry/sentry.git | 1 | 150 | 0 | 45 | 260 | Python | {
"docstring": "\n Test that with digests enabled, but Slack notification settings\n (and not email settings), we send a Slack notification\n ",
"language": "en",
"n_whitespaces": 40,
"n_words": 18,
"vocab_size": 16
} | def test_digest_enabled(self, digests, mock_func):
backend = RedisBackend()
digests.digest = backend.digest
digests.enabled.return_value = True
rule = Rule.objects.create(project=self.project, label="my rule")
ProjectOwnership.objects.create(project_id=self.project.id, ... | |
47,393 | 195,738 | 68 | sympy/physics/control/control_plots.py | 37 | 18 | def pole_zero_numerical_data(system):
_check_system(system)
system = system.doit() # Get the equivalent TransferFunction object.
num_poly = Poly(system.num, system.var).all_coeffs()
den_poly = Poly(system.den, system.var).all_coeffs()
num_poly = np.array(num_poly, dtype=np.complex128)
de... | Allow complex transfer functions in pole-zero plot | pole_zero_numerical_data | bf1cb469061d7ad07bfbf687f5635d9f4ec569dd | sympy | control_plots.py | 11 | 10 | https://github.com/sympy/sympy.git | 1 | 97 | 0 | 26 | 157 | Python | {
"docstring": "\n Returns the numerical data of poles and zeros of the system.\n It is internally used by ``pole_zero_plot`` to get the data\n for plotting poles and zeros. Users can use this data to further\n analyse the dynamics of the system or plot using a different\n backend/plotting-module.\n\n ... | def pole_zero_numerical_data(system):
_check_system(system)
system = system.doit() # Get the equivalent TransferFunction object.
num_poly = Poly(system.num, system.var).all_coeffs()
den_poly = Poly(system.den, system.var).all_coeffs()
num_poly = np.array(num_poly, dtype=np.complex128)
de... | |
56,767 | 222,834 | 25 | python3.10.4/Lib/distutils/cygwinccompiler.py | 16 | 5 | def get_versions():
commands = ['gcc -dumpversion', 'ld -v', 'dllwrap --version']
return t | add python 3.10.4 for windows | get_versions | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | cygwinccompiler.py | 10 | 3 | https://github.com/XX-net/XX-Net.git | 2 | 28 | 0 | 16 | 51 | Python | {
"docstring": " Try to find out the versions of gcc, ld and dllwrap.\n\n If not possible it returns None for it.\n ",
"language": "en",
"n_whitespaces": 26,
"n_words": 19,
"vocab_size": 19
} | def get_versions():
commands = ['gcc -dumpversion', 'ld -v', 'dllwrap --version']
return tuple([_find_exe_version(cmd) for cmd in commands])
| |
53,868 | 215,170 | 63 | salt/beacons/napalm_beacon.py | 26 | 11 | def __virtual__():
if salt.utils.napalm.virtual(__opts__, __virtualname__, __file__):
return __virtualname__
else:
err_msg = "NAPALM is not installed."
log.error("Unable to load %s beacon: %s", __virtualname__, err_msg) | Align enhanced logging accross beacons | __virtual__ | 4e3632254fb73210ce3e1954ec507473433018b8 | salt | napalm_beacon.py | 11 | 7 | https://github.com/saltstack/salt.git | 2 | 42 | 0 | 23 | 71 | Python | {
"docstring": "\n This beacon can only work when running under a regular or a proxy minion, managed through napalm.\n ",
"language": "en",
"n_whitespaces": 24,
"n_words": 17,
"vocab_size": 16
} | def __virtual__():
if salt.utils.napalm.virtual(__opts__, __virtualname__, __file__):
return __virtualname__
else:
err_msg = "NAPALM is not installed."
log.error("Unable to load %s beacon: %s", __virtualname__, err_msg)
return False, err_msg
| |
120,971 | 337,188 | 37 | examples/community/lpw_stable_diffusion.py | 19 | 7 | def parse_prompt_attention(text):
res = []
round_brackets = []
square_brackets = []
round_bracket_multiplier = 1.1
square_bracket_multiplier = 1 / 1.1
| [Community Pipelines] Long Prompt Weighting Stable Diffusion Pipelines (#907)
* [Community Pipelines] Long Prompt Weighting
* Update README.md
* fix
* style
* fix style
* Update examples/community/README.md
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com> | parse_prompt_attention | 2a0c823527694058d410ed6f91b52e7dd9f94ebe | diffusers | lpw_stable_diffusion.py | 7 | 38 | https://github.com/huggingface/diffusers.git | 16 | 299 | 0 | 12 | 49 | Python | {
"docstring": "\n Parses a string with attention tokens and returns a list of pairs: text and its assoicated weight.\n Accepted tokens are:\n (abc) - increases attention to abc by a multiplier of 1.1\n (abc:3.12) - increases attention to abc by a multiplier of 3.12\n [abc] - decreases attention ... | def parse_prompt_attention(text):
res = []
round_brackets = []
square_brackets = []
round_bracket_multiplier = 1.1
square_bracket_multiplier = 1 / 1.1
| |
78,638 | 266,879 | 364 | lib/ansible/galaxy/dependency_resolution/providers.py | 178 | 12 | def get_dependencies(self, candidate):
# type: (Candidate) -> list[Candidate]
r
# FIXME: If there's several galaxy servers set, there may be a
# FIXME: situation when the metadata of the same collection
# FIXME: differs. So how do we resolve this case? Priority?
# FIXME: ... | galaxy - Clean up type hints and imports. | get_dependencies | 8b2e6285650ec42ec4a19075a8567047e8304ba2 | ansible | providers.py | 11 | 13 | https://github.com/ansible/ansible.git | 4 | 60 | 0 | 125 | 115 | Python | {
"docstring": "Get direct dependencies of a candidate.\n\n :returns: A collection of requirements that `candidate` \\\n specifies as its dependencies.\n ",
"language": "en",
"n_whitespaces": 49,
"n_words": 18,
"vocab_size": 17
} | def get_dependencies(self, candidate):
# type: (Candidate) -> list[Candidate]
r
# FIXME: If there's several galaxy servers set, there may be a
# FIXME: situation when the metadata of the same collection
# FIXME: differs. So how do we resolve this case? Priority?
# FIXME: ... | |
71,047 | 246,153 | 208 | tests/rest/admin/test_user.py | 40 | 14 | def test_set_displayname(self) -> None:
# Modify user
channel = self.make_request(
"PUT",
self.url_other_user,
access_token=self.admin_user_tok,
content={"displayname": "foobar"},
)
self.assertEqual(HTTPStatus.OK, channel.code, m... | Add type hints to `tests/rest/admin` (#11851) | test_set_displayname | 901b264c0c88f39cbfb8b2229e0dc57968882658 | synapse | test_user.py | 12 | 21 | https://github.com/matrix-org/synapse.git | 1 | 142 | 0 | 25 | 235 | Python | {
"docstring": "\n Test setting the displayname of another user.\n ",
"language": "en",
"n_whitespaces": 22,
"n_words": 7,
"vocab_size": 7
} | def test_set_displayname(self) -> None:
# Modify user
channel = self.make_request(
"PUT",
self.url_other_user,
access_token=self.admin_user_tok,
content={"displayname": "foobar"},
)
self.assertEqual(HTTPStatus.OK, channel.code, m... | |
117,847 | 321,664 | 437 | qutebrowser/browser/webkit/network/networkmanager.py | 94 | 37 | def on_ssl_errors(self, reply, qt_errors):
| Refactor certificate error handling
- Make CertificateErrorWrapper responsible for accepting/rejecting certs
- Try to avoid dealing with unclear booleans
- Implement support for deferred errors (#4616) - disabled due to PyQt bug
- Implement support for Qt 6 API (#7086) | on_ssl_errors | e5340c449f23608803c286da0563b62f58ba25b0 | qutebrowser | networkmanager.py | 13 | 35 | https://github.com/qutebrowser/qutebrowser.git | 8 | 220 | 0 | 62 | 353 | Python | {
"docstring": "Decide if SSL errors should be ignored or not.\n\n This slot is called on SSL/TLS errors by the self.sslErrors signal.\n\n Args:\n reply: The QNetworkReply that is encountering the errors.\n qt_errors: A list of errors.\n ",
"language": "en",
"n_whitespac... | def on_ssl_errors(self, reply, qt_errors):
errors = certificateerror.CertificateErrorWrapper(reply, qt_errors)
log.network.debug("Certificate errors: {!r}".format(errors))
try:
host_tpl: Optional[urlutils.HostTupleType] = urlutils.host_tuple(
reply.url())
... | |
40,694 | 171,627 | 142 | pandas/_version.py | 36 | 4 | def render_pep440(pieces):
i | BLD: use nonvendor versioneer (#49924)
* BLD: remove vendored versioneer
* run vis
* move config to pyproject.toml
* add versioneer to deps
* run pyupgrade
* fix isort and pylint
* fix ci
* fix env | render_pep440 | e2df99823758210fb2b7c4aba39e23f3445f7cd3 | pandas | _version.py | 14 | 13 | https://github.com/pandas-dev/pandas.git | 6 | 65 | 0 | 20 | 163 | Python | {
"docstring": "Build up version string, with post-release \"local version identifier\".\n\n Our goal: TAG[+DISTANCE.gHEX[.dirty]] . Note that if you\n get a tagged build and then dirty it, you'll get TAG+0.gHEX.dirty\n\n Exceptions:\n 1: no tags. git_describe was just HEX. 0+untagged.DISTANCE.gHEX[.dirty... | def render_pep440(pieces):
if pieces["closest-tag"]:
rendered = pieces["closest-tag"]
if pieces["distance"] or pieces["dirty"]:
rendered += plus_or_dot(pieces)
rendered += f"{pieces['distance']}.g{pieces['short']}"
if pieces["dirty"]:
rendered... | |
50,052 | 202,099 | 69 | tests/cache/tests_async.py | 19 | 7 | async def test_aset_many(self):
self.assertEqual(await cache.aset_many({"a": 1, "b": 2}), [])
self.assert | Refs #33476 -- Reformatted code with Black. | test_aset_many | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | tests_async.py | 13 | 6 | https://github.com/django/django.git | 1 | 61 | 0 | 16 | 105 | Python | {
"docstring": "aset_many() does nothing for the dummy cache backend.",
"language": "en",
"n_whitespaces": 7,
"n_words": 8,
"vocab_size": 8
} | async def test_aset_many(self):
self.assertEqual(await cache.aset_many({"a": 1, "b": 2}), [])
self.assertEqual(
await cache.aset_many({"a": 1, "b": 2}, timeout=2, version="1"),
[],
)
| |
50,013 | 201,852 | 77 | tests/bash_completion/tests.py | 34 | 13 | def _user_input(self, input_str):
os.environ["COMP_WORDS"] = input_str
idx = | Refs #33476 -- Reformatted code with Black. | _user_input | 9c19aff7c7561e3a82978a272ecdaad40dda5c00 | django | tests.py | 12 | 6 | https://github.com/django/django.git | 2 | 63 | 0 | 27 | 110 | Python | {
"docstring": "\n Set the environment and the list of command line arguments.\n\n This sets the bash variables $COMP_WORDS and $COMP_CWORD. The former is\n an array consisting of the individual words in the current command\n line, the latter is the index of the current cursor position, so... | def _user_input(self, input_str):
os.environ["COMP_WORDS"] = input_str
idx = len(input_str.split(" ")) - 1 # Index of the last word
comp_cword = idx + 1 if input_str.endswith(" ") else idx
os.environ["COMP_CWORD"] = str(comp_cword)
sys.argv = input_str.split()
| |
26,352 | 118,671 | 229 | lib/streamlit/config.py | 102 | 9 | def _check_conflicts() -> None:
# Node-related conflicts
# When using the Node server, we must always connect to 8501 (this is
# hard-coded in JS). Otherwise, the browser would decide what port to
# connect to based on window.location.port, which in dev is going to
# be (3000)
# Import logger ... | Report sharing removal (#4260)
The report sharing feature is a substantial but completely unused portion of the code in Streamlit's underlying machinery. The feature was created early on, used by just a few groups, and has not been used by anyone for a while, as indicated by no activity in the associated S3 buckets. T... | _check_conflicts | dd9084523e365e637443ea351eaaaa25f52d8412 | streamlit | config.py | 12 | 25 | https://github.com/streamlit/streamlit.git | 5 | 65 | 0 | 74 | 132 | Python | {
"docstring": "\nWarning: the config option 'server.enableCORS=false' is not compatible with 'server.enableXsrfProtection=true'.\nAs a result, 'server.enableCORS' is being overridden to 'true'.\n\nMore information:\nIn order to protect against CSRF attacks, we send a cookie with each request.\nTo do so, we must spec... | def _check_conflicts() -> None:
# Node-related conflicts
# When using the Node server, we must always connect to 8501 (this is
# hard-coded in JS). Otherwise, the browser would decide what port to
# connect to based on window.location.port, which in dev is going to
# be (3000)
# Import logger ... | |
21,967 | 104,785 | 35 | src/datasets/dataset_dict.py | 14 | 9 | def num_columns(self) -> Dict[str, int]:
self._check_values_type()
return {k: dataset.num_columns for k, datase | Add code examples for DatasetDict (#4245)
* 📝 add code examples for DatasetDict
* 🖍 apply quentin review | num_columns | 1904d0c0a3a96330d9b870cdca3e9a3a137f2977 | datasets | dataset_dict.py | 9 | 14 | https://github.com/huggingface/datasets.git | 2 | 36 | 0 | 14 | 58 | Python | {
"docstring": "Number of columns in each split of the dataset.\n\n Example:\n\n ```py\n >>> from datasets import load_dataset\n >>> ds = load_dataset(\"rotten_tomatoes\")\n >>> ds.num_columns\n {'test': 2, 'train': 2, 'validation': 2}\n ```\n ",
"language": "... | def num_columns(self) -> Dict[str, int]:
self._check_values_type()
return {k: dataset.num_columns for k, dataset in self.items()}
| |
36,586 | 156,162 | 207 | dask/bag/random.py | 72 | 23 | def _sample_with_replacement_map_partitions(population, k):
stream = iter(population)
e = next(stream)
reservoir, stream_length = [e for _ in range(k)], 1
w = [rnd.random() for _ in range(k)]
nxt = [_geometric(wi) for wi in w]
min_nxt = min(nxt)
for i, e in enumerate(stream, 1):
... | Bag: add implementation for reservoir sampling (#7068) (#7636)
- Implement the [L algorithm](https://en.wikipedia.org/wiki/Reservoir_sampling#An_optimal_algorithm) for reservoir sampling without replacement.
- Use the **k** reservoir of size 1 strategy for sampling with replacement (see [reference](http://uto... | _sample_with_replacement_map_partitions | 4e5dfe7463028a39a90e026c7fb9220969093ab3 | dask | random.py | 17 | 17 | https://github.com/dask/dask.git | 8 | 144 | 0 | 43 | 224 | Python | {
"docstring": "\n Reservoir sampling with replacement, the main idea is to use k reservoirs of size 1\n See Section Applications in http://utopia.duth.gr/~pefraimi/research/data/2007EncOfAlg.pdf\n ",
"language": "en",
"n_whitespaces": 30,
"n_words": 20,
"vocab_size": 20
} | def _sample_with_replacement_map_partitions(population, k):
stream = iter(population)
e = next(stream)
reservoir, stream_length = [e for _ in range(k)], 1
w = [rnd.random() for _ in range(k)]
nxt = [_geometric(wi) for wi in w]
min_nxt = min(nxt)
for i, e in enumerate(stream, 1):
... | |
41,589 | 175,300 | 61 | Lib/enum.py | 22 | 9 | def __setattr__(cls, name, value):
member_map = cls.__dict__.get('_member_map_', {})
if name in member_map:
raise AttributeError('cannot reassign member %r' % (name, ))
super().__s | bpo-40066: [Enum] update str() and format() output (GH-30582)
Undo rejected PEP-663 changes:
- restore `repr()` to its 3.10 status
- restore `str()` to its 3.10 status
New changes:
- `IntEnum` and `IntFlag` now leave `__str__` as the original `int.__str__` so that str() and format() return the same result
... | __setattr__ | acf7403f9baea3ae1119fc6b4a3298522188bf96 | cpython | enum.py | 11 | 5 | https://github.com/python/cpython.git | 2 | 48 | 0 | 22 | 80 | Python | {
"docstring": "\n Block attempts to reassign Enum members.\n\n A simple assignment to the class namespace only changes one of the\n several possible ways to get an Enum member from the Enum class,\n resulting in an inconsistent Enumeration.\n ",
"language": "en",
"n_whitespaces... | def __setattr__(cls, name, value):
member_map = cls.__dict__.get('_member_map_', {})
if name in member_map:
raise AttributeError('cannot reassign member %r' % (name, ))
super().__setattr__(name, value)
| |
56,453 | 221,633 | 92 | python3.10.4/Lib/configparser.py | 19 | 7 | def read_file(self, f, source=None):
if source is None:
try:
source = f.name
except AttributeError:
| add python 3.10.4 for windows | read_file | 8198943edd73a363c266633e1aa5b2a9e9c9f526 | XX-Net | configparser.py | 12 | 7 | https://github.com/XX-net/XX-Net.git | 3 | 38 | 0 | 16 | 64 | Python | {
"docstring": "Like read() but the argument must be a file-like object.\n\n The `f' argument must be iterable, returning one line at a time.\n Optional second argument is the `source' specifying the name of the\n file being read. If not given, it is taken from f.name. If `f' has no\n `nam... | def read_file(self, f, source=None):
if source is None:
try:
source = f.name
except AttributeError:
source = '<???>'
self._read(f, source)
| |
83,766 | 281,449 | 286 | gamestonk_terminal/cryptocurrency/defi/substack_model.py | 90 | 42 | def get_newsletters() -> pd.DataFrame:
urls = [
"https://defiweekly.substack.com/archive",
"https://newsletter.thedefiant.io/archive",
"https://thedailygwei.substack.com/archive",
"https://todayindefi.substack.com/archive",
"https://newsletter.banklesshq.com/archive",
... | Terminal Wide Rich (#1161)
* My idea for how we handle Rich moving forward
* remove independent consoles
* FIxed pylint issues
* add a few vars
* Switched print to console
* More transitions
* Changed more prints
* Replaced all prints
* Fixing tabulate
* Finished replace tabulate
* Finish... | get_newsletters | 82747072c511beb1b2672846ae2ee4aec53eb562 | OpenBBTerminal | substack_model.py | 15 | 40 | https://github.com/OpenBB-finance/OpenBBTerminal.git | 6 | 242 | 0 | 72 | 419 | Python | {
"docstring": "Scrape all substack newsletters from url list.\n [Source: substack.com]\n\n Returns\n -------\n pd.DataFrame\n DataFrame with recent news from most popular DeFi related newsletters.\n ",
"language": "en",
"n_whitespaces": 44,
"n_words": 22,
"vocab_size": 21
} | def get_newsletters() -> pd.DataFrame:
urls = [
"https://defiweekly.substack.com/archive",
"https://newsletter.thedefiant.io/archive",
"https://thedailygwei.substack.com/archive",
"https://todayindefi.substack.com/archive",
"https://newsletter.banklesshq.com/archive",
... | |
343 | 2,712 | 76 | packages/syft/src/syft/core/node/common/action/get_enum_attribute_action.py | 11 | 9 | def _object2proto(self) -> GetEnumAttributeAction_PB:
return GetEnumAttri | [syft.core.node.common.action] Change syft import absolute -> relative | _object2proto | e272ed2fa4c58e0a89e273a3e85da7d13a85e04c | PySyft | get_enum_attribute_action.py | 11 | 18 | https://github.com/OpenMined/PySyft.git | 1 | 45 | 0 | 11 | 70 | Python | {
"docstring": "Returns a protobuf serialization of self.\n As a requirement of all objects which inherit from Serializable,\n this method transforms the current object into the corresponding\n Protobuf object so that it can be further serialized.\n :return: returns a protobuf object\n ... | def _object2proto(self) -> GetEnumAttributeAction_PB:
return GetEnumAttributeAction_PB(
path=self.path,
id_at_location=serialize(self.id_at_location),
address=serialize(self.address),
msg_id=serialize(self.id),
)
| |
107,924 | 309,217 | 253 | tests/components/seventeentrack/test_sensor.py | 70 | 27 | async def test_becomes_delivered_not_shown_notification(hass):
package = Package(
tracking_number="456",
destination_country=206,
friendly_name="friendly name 1",
info_text="info text 1",
location="location 1",
timestamp="2020-08-10 10:32",
origin_country... | Import persistent notification (part 4) (#63901) | test_becomes_delivered_not_shown_notification | a672dc3437b95734e44cb3f61b3f3c299627bb1a | core | test_sensor.py | 11 | 33 | https://github.com/home-assistant/core.git | 1 | 159 | 0 | 43 | 265 | Python | {
"docstring": "Ensure notification is triggered when package becomes delivered.",
"language": "en",
"n_whitespaces": 7,
"n_words": 8,
"vocab_size": 8
} | async def test_becomes_delivered_not_shown_notification(hass):
package = Package(
tracking_number="456",
destination_country=206,
friendly_name="friendly name 1",
info_text="info text 1",
location="location 1",
timestamp="2020-08-10 10:32",
origin_country... | |
75,913 | 259,775 | 939 | sklearn/ensemble/_iforest.py | 230 | 38 | def fit(self, X, y=None, sample_weight=None):
X = self._validate_data(X, accept_sparse=["csc"])
if issparse(X):
# Pre-sort indices to avoid that each individual tree of the
# ensemble sorts the indices.
X.sort_indices()
rnd = check_random_state(self.... | ENH Optimize runtime for IsolationForest (#23149) | fit | 767e9ae7e4fec8bea36c0433ab42f500aacfde64 | scikit-learn | _iforest.py | 15 | 54 | https://github.com/scikit-learn/scikit-learn.git | 10 | 318 | 0 | 139 | 503 | Python | {
"docstring": "\n Fit estimator.\n\n Parameters\n ----------\n X : {array-like, sparse matrix} of shape (n_samples, n_features)\n The input samples. Use ``dtype=np.float32`` for maximum\n efficiency. Sparse matrices are also supported, use sparse\n ``csc_m... | def fit(self, X, y=None, sample_weight=None):
X = self._validate_data(X, accept_sparse=["csc"])
if issparse(X):
# Pre-sort indices to avoid that each individual tree of the
# ensemble sorts the indices.
X.sort_indices()
rnd = check_random_state(self.... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.