Skip to content

Commit ff6f673

Browse files
committed
Merge branch 'main' into wence/fix/tornado-dep
2 parents f613809 + 803c624 commit ff6f673

91 files changed

Lines changed: 3629 additions & 2077 deletions

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

.github/workflows/conda.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ jobs:
3030
with:
3131
fetch-depth: 0
3232
- name: Set up Python
33-
uses: conda-incubator/setup-miniconda@v2
33+
uses: conda-incubator/setup-miniconda@v2.2.0
3434
with:
3535
miniforge-variant: Mambaforge
3636
use-mamba: true

.github/workflows/test-report.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ jobs:
2222
- uses: actions/checkout@v3.1.0
2323

2424
- name: Setup Conda Environment
25-
uses: conda-incubator/setup-miniconda@v2
25+
uses: conda-incubator/setup-miniconda@v2.2.0
2626
with:
2727
miniforge-variant: Mambaforge
2828
miniforge-version: latest
@@ -50,8 +50,8 @@ jobs:
5050

5151
- name: Generate report
5252
run: |
53-
python continuous_integration/scripts/test_report.py --max-days 90 --max-runs 50 --nfails 1 -o test_report.html
54-
python continuous_integration/scripts/test_report.py --max-days 7 --max-runs 50 --nfails 2 -o test_short_report.html --title "Test Short Report"
53+
python continuous_integration/scripts/test_report.py --max-days 90 --max-runs 30 --nfails 1 -o test_report.html
54+
python continuous_integration/scripts/test_report.py --max-days 7 --max-runs 30 --nfails 2 -o test_short_report.html --title "Test Short Report"
5555
mkdir deploy
5656
mv test_report.html test_short_report.html deploy/
5757

.github/workflows/tests.yaml

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ jobs:
2424
matrix:
2525
os: [ubuntu-latest, windows-latest, macos-latest]
2626
python-version: ["3.8", "3.9", "3.10"]
27-
queuing: [no_queue]
27+
queuing: [queue]
2828
# Cherry-pick test modules to split the overall runtime roughly in half
2929
partition: [ci1, not ci1]
3030
exclude:
@@ -33,11 +33,11 @@ jobs:
3333
include:
3434
- os: ubuntu-latest
3535
python-version: 3.9
36-
queuing: queue
36+
queuing: no_queue
3737
partition: "ci1"
3838
- os: ubuntu-latest
3939
python-version: 3.9
40-
queuing: queue
40+
queuing: no_queue
4141
partition: "not ci1"
4242

4343
# Uncomment to stress-test the test suite for random failures.
@@ -68,7 +68,7 @@ jobs:
6868
fetch-depth: 0
6969

7070
- name: Setup Conda Environment
71-
uses: conda-incubator/setup-miniconda@v2
71+
uses: conda-incubator/setup-miniconda@v2.2.0
7272
with:
7373
miniforge-variant: Mambaforge
7474
miniforge-version: latest
@@ -144,8 +144,8 @@ jobs:
144144

145145
- name: Set up dask env for job queuing
146146
shell: bash -l {0}
147-
if: ${{ matrix.queuing == 'queue' }}
148-
run: echo "DASK_DISTRIBUTED__SCHEDULER__WORKER_SATURATION=1.0" >> $GITHUB_ENV
147+
if: ${{ matrix.queuing == 'no_queue' }}
148+
run: echo "DASK_DISTRIBUTED__SCHEDULER__WORKER_SATURATION=inf" >> $GITHUB_ENV
149149

150150
- name: Print host info
151151
shell: bash -l {0}
@@ -200,6 +200,8 @@ jobs:
200200
uses: codecov/codecov-action@v3
201201
with:
202202
name: ${{ env.TEST_ID }}
203+
# See https://community.codecov.com/t/upload-issues-unable-to-locate-build-via-github-actions-api/3954
204+
token: ${{ secrets.CODECOV_TOKEN }}
203205

204206
- name: Upload test results
205207
# ensure this runs even if pytest fails

.github/workflows/update-gpuci.yaml

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,6 +14,7 @@ jobs:
1414
- uses: actions/checkout@v3.1.0
1515

1616
- name: Parse current axis YAML
17+
id: rapids_current
1718
uses: the-coding-turtle/ga-yaml-parser@v0.1.2
1819
with:
1920
file: continuous_integration/gpuci/axis.yaml
@@ -39,8 +40,8 @@ jobs:
3940
FULL_RAPIDS_VER: ${{ steps.cudf_latest.outputs.version }}
4041
FULL_UCX_PY_VER: ${{ steps.ucx_py_latest.outputs.version }}
4142
run: |
42-
echo RAPIDS_VER=$RAPIDS_VER_0 >> $GITHUB_ENV
43-
echo UCX_PY_VER=$(curl -sL https://version.gpuci.io/rapids/$RAPIDS_VER_0) >> $GITHUB_ENV
43+
echo RAPIDS_VER=${{ steps.rapids_current.outputs.RAPIDS_VER_0 }} >> $GITHUB_ENV
44+
echo UCX_PY_VER=$(curl -sL https://version.gpuci.io/rapids/${{ steps.rapids_current.outputs.RAPIDS_VER_0 }}) >> $GITHUB_ENV
4445
echo NEW_RAPIDS_VER=${FULL_RAPIDS_VER::-10} >> $GITHUB_ENV
4546
echo NEW_UCX_PY_VER=${FULL_UCX_PY_VER::-10} >> $GITHUB_ENV
4647

.pre-commit-config.yaml

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,12 @@ repos:
3434
# NOTE: autoupdate does not pick up flake8-bugbear since it is a transitive
3535
# dependency. Make sure to update flake8-bugbear manually on a regular basis.
3636
- flake8-bugbear==22.9.23
37+
- repo: https://github.com/codespell-project/codespell
38+
rev: v2.1.0
39+
hooks:
40+
- id: codespell
41+
types_or: [rst, markdown]
42+
files: docs
3743
- repo: https://github.com/pre-commit/mirrors-mypy
3844
rev: v0.982
3945
hooks:

codecov.yml

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,11 @@
11
codecov:
22
require_ci_to_pass: yes
33

4+
ignore:
5+
# Files that exercise GPU-only functionality or are only tested by gpuCI
6+
# but don't interact with codecov are ignored.
7+
- "distributed/comm/ucx.py"
8+
49
coverage:
510
precision: 2
611
round: down

continuous_integration/gpuci/axis.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,6 @@ LINUX_VER:
88
- ubuntu18.04
99

1010
RAPIDS_VER:
11-
- "22.12"
11+
- "23.02"
1212

1313
excludes:

continuous_integration/recipes/distributed/meta.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ requirements:
3131
- python >=3.8
3232
- click >=7.0
3333
- cloudpickle >=1.5.0
34-
- cytoolz >=0.8.2
34+
- cytoolz >=0.10.0
3535
- {{ pin_compatible('dask-core', max_pin='x.x.x.x') }}
3636
- jinja2
3737
- locket >=1.0.0
@@ -41,8 +41,8 @@ requirements:
4141
- pyyaml
4242
- sortedcontainers !=2.0.0,!=2.0.1
4343
- tblib >=1.6.0
44-
- toolz >=0.8.2
45-
- tornado >=6.2
44+
- toolz >=0.10.0
45+
- tornado >=6.0.3
4646
- urllib3
4747
- zict >=0.1.3
4848
run_constrained:

distributed/cli/dask_worker.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -462,7 +462,7 @@ def _apportion_ports(
462462
463463
Returns
464464
=======
465-
List of kwargs to pass to the Worker or Nanny construtors
465+
List of kwargs to pass to the Worker or Nanny constructors
466466
"""
467467
seen = set()
468468

distributed/client.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -478,7 +478,7 @@ def __del__(self):
478478
try:
479479
self.release()
480480
except AttributeError:
481-
# Ocassionally we see this error when shutting down the client
481+
# Occasionally we see this error when shutting down the client
482482
# https://github.com/dask/distributed/issues/4305
483483
if not sys.is_finalizing():
484484
raise
@@ -2851,7 +2851,7 @@ def run(
28512851
wait : boolean (optional)
28522852
If the function is asynchronous whether or not to wait until that
28532853
function finishes.
2854-
nanny : bool, defualt False
2854+
nanny : bool, default False
28552855
Whether to run ``function`` on the nanny. By default, the function
28562856
is run on the worker process. If specified, the addresses in
28572857
``workers`` should still be the worker addresses, not the nanny addresses.
@@ -2885,7 +2885,7 @@ def run(
28852885
>>> def get_status(dask_worker):
28862886
... return dask_worker.status
28872887
2888-
>>> c.run(get_hostname) # doctest: +SKIP
2888+
>>> c.run(get_status) # doctest: +SKIP
28892889
{'192.168.0.100:9000': 'running',
28902890
'192.168.0.101:9000': 'running}
28912891
@@ -4206,7 +4206,7 @@ def get_scheduler_logs(self, n=None):
42064206
Parameters
42074207
----------
42084208
n : int
4209-
Number of logs to retrive. Maxes out at 10000 by default,
4209+
Number of logs to retrieve. Maxes out at 10000 by default,
42104210
configurable via the ``distributed.admin.log-length``
42114211
configuration value.
42124212
@@ -4222,7 +4222,7 @@ def get_worker_logs(self, n=None, workers=None, nanny=False):
42224222
Parameters
42234223
----------
42244224
n : int
4225-
Number of logs to retrive. Maxes out at 10000 by default,
4225+
Number of logs to retrieve. Maxes out at 10000 by default,
42264226
configurable via the ``distributed.admin.log-length``
42274227
configuration value.
42284228
workers : iterable

0 commit comments

Comments
 (0)