Skip to content

Commit 7eac12e

Browse files
Fix typos
1 parent a13b06c commit 7eac12e

8 files changed

Lines changed: 22 additions & 22 deletions

File tree

distributed/comm/tests/test_comms.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1321,7 +1321,7 @@ async def check_addresses(a, b):
13211321

13221322

13231323
@gen_test()
1324-
async def test_tcp_adresses(tcp):
1324+
async def test_tcp_addresses(tcp):
13251325
a, b = await get_tcp_comm_pair()
13261326
await check_addresses(a, b)
13271327

@@ -1336,13 +1336,13 @@ async def test_get_stream_address_raises_if_closed():
13361336

13371337

13381338
@gen_test()
1339-
async def test_tls_adresses(tcp):
1339+
async def test_tls_addresses(tcp):
13401340
a, b = await get_tls_comm_pair()
13411341
await check_addresses(a, b)
13421342

13431343

13441344
@gen_test()
1345-
async def test_inproc_adresses():
1345+
async def test_inproc_addresses():
13461346
a, b = await get_inproc_comm_pair()
13471347
await check_addresses(a, b)
13481348

distributed/dashboard/components/scheduler.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -2331,7 +2331,7 @@ def __init__(self, scheduler, **kwargs):
23312331
@log_errors
23322332
def update(self):
23332333
# If there are too many tasks in the scheduler we'll disable this
2334-
# compoonents to not overload scheduler or client. Once we drop
2334+
# components to not overload scheduler or client. Once we drop
23352335
# below the threshold, the data is filled up again as usual
23362336
if len(self.scheduler.tasks) > self.max_items:
23372337
self.subtitle.text = "Scheduler has too many tasks to display."
@@ -2863,7 +2863,7 @@ def update(self):
28632863
# Add some status to hover
28642864
tasks_processing = tg.states["processing"]
28652865
tasks_memory = tg.states["memory"]
2866-
tasks_relased = tg.states["released"]
2866+
tasks_released = tg.states["released"]
28672867
tasks_erred = tg.states["erred"]
28682868

28692869
nodes_data["comp_tasks"].append(
@@ -2876,7 +2876,7 @@ def update(self):
28762876
f"{tasks_memory} ({tasks_memory / tot_tasks * 100:.0f} %)"
28772877
)
28782878
nodes_data["in_released"].append(
2879-
f"{tasks_relased} ({tasks_relased / tot_tasks * 100:.0f} %)"
2879+
f"{tasks_released} ({tasks_released / tot_tasks * 100:.0f} %)"
28802880
)
28812881
nodes_data["in_erred"].append(
28822882
f"{tasks_erred} ({tasks_erred / tot_tasks * 100:.0f} %)"
@@ -2963,7 +2963,7 @@ def _should_add_new_renderers(self) -> bool:
29632963
def _should_update(self) -> bool:
29642964
"""
29652965
Whether to update the ColumnDataSource. This is cheaper than redrawing,
2966-
but still not free, so we check whether we need it and whether the scheudler
2966+
but still not free, so we check whether we need it and whether the scheduler
29672967
is busy.
29682968
"""
29692969
return (

distributed/deploy/subprocess.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -193,7 +193,7 @@ def SubprocessCluster(
193193
host:
194194
Host address on which the scheduler will listen, defaults to localhost
195195
scheduler_port:
196-
Port fo the scheduler, defaults to 0 to choose a random port
196+
Port for the scheduler, defaults to 0 to choose a random port
197197
scheduler_kwargs:
198198
Keywords to pass on to scheduler
199199
dashboard_address:

distributed/protocol/tests/test_protocol_utils.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -36,17 +36,17 @@ def test_unpack_frames_partial():
3636
frames = [b"123", b"asdf"]
3737
frames.insert(0, pack_frames_prelude(frames))
3838

39-
frames2, missing_lenghts = unpack_frames(b"".join(frames), partial=True)
39+
frames2, missing_lengths = unpack_frames(b"".join(frames), partial=True)
4040
assert frames2 == frames[1:]
41-
assert missing_lenghts == []
41+
assert missing_lengths == []
4242

43-
frames2, missing_lenghts = unpack_frames(b"".join(frames[:-1]), partial=True)
43+
frames2, missing_lengths = unpack_frames(b"".join(frames[:-1]), partial=True)
4444
assert frames2 == frames[1:-1]
45-
assert missing_lenghts == [4]
45+
assert missing_lengths == [4]
4646

47-
frames2, missing_lenghts = unpack_frames(frames[0], partial=True)
47+
frames2, missing_lengths = unpack_frames(frames[0], partial=True)
4848
assert frames2 == []
49-
assert missing_lenghts == [3, 4]
49+
assert missing_lengths == [3, 4]
5050

5151
with pytest.raises(AssertionError):
5252
unpack_frames(b"".join(frames[:-1]))

distributed/security.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@
2020
# https://docs.python.org/3.10/library/ssl.html#ssl.SSLContext.minimum_version
2121
# https://docs.python.org/3.7/library/ssl.html#ssl.SSLContext.minimum_version
2222

23-
# these _set_mimimun_version and _set_maximum_version depend on the validation
23+
# these _set_minimum_version and _set_maximum_version depend on the validation
2424
# already performed in `Security._set_tls_version_field`,
2525
# and that they only apply to freshly created ssl.SSLContext instances in
2626
# _get_tls_context

distributed/shuffle/tests/test_shuffle.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -233,7 +233,7 @@ async def test_shuffle_with_array_conversion(c, s, a, b, npartitions):
233233
out = df.shuffle("x", npartitions=npartitions, force=True).values
234234

235235
# See distributed#7816. TaskSpec is currently blocking linear fusion. If
236-
# that was implemented, this may raise a P2PConsistencyErrro
236+
# that was implemented, this may raise a P2PConsistencyError
237237

238238
await c.compute(out)
239239

distributed/tests/test_client.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2520,7 +2520,7 @@ async def test_traceback_clean(c, s, a, b):
25202520

25212521

25222522
@gen_cluster(client=True)
2523-
async def test_map_differnet_lengths(c, s, a, b):
2523+
async def test_map_different_lengths(c, s, a, b):
25242524
assert len(c.map(add, [1, 2], [1, 2, 3])) == 2
25252525

25262526

distributed/tests/test_jupyter.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -77,11 +77,11 @@ async def test_jupyter_idle_timeout():
7777
# time, which is used to determine idleness. Instead of a full e2e test
7878
# launching a kernel and running commands on it, we'll just hook into this:
7979
# https://github.com/jupyter-server/jupyter_server/blob/e582e555/jupyter_server/serverapp.py#L385-L387
80-
extension_last_activty = web_app.settings["last_activity_times"]
80+
extension_last_activity = web_app.settings["last_activity_times"]
8181

8282
for _ in range(10):
8383
last = perf_counter()
84-
extension_last_activty["test"] = datetime.now(timezone.utc)
84+
extension_last_activity["test"] = datetime.now(timezone.utc)
8585

8686
await asyncio.sleep(s.idle_timeout / 2)
8787
if (d := perf_counter() - last) >= s.idle_timeout:
@@ -99,12 +99,12 @@ async def test_jupyter_idle_timeout_returned():
9999
"`check_idle` should return the last Jupyter idle time. Used in dask-kubernetes."
100100
async with Scheduler(jupyter=True, dashboard_address=":0") as s:
101101
web_app = s._jupyter_server_application.web_app
102-
extension_last_activty = web_app.settings["last_activity_times"]
102+
extension_last_activity = web_app.settings["last_activity_times"]
103103

104-
extension_last_activty["test"] = datetime.now(timezone.utc)
104+
extension_last_activity["test"] = datetime.now(timezone.utc)
105105
last_idle = s.check_idle()
106106
assert last_idle is not None
107-
extension_last_activty["test"] = datetime.now(timezone.utc) + timedelta(
107+
extension_last_activity["test"] = datetime.now(timezone.utc) + timedelta(
108108
seconds=1
109109
)
110110
next_idle = s.check_idle()

0 commit comments

Comments
 (0)