Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Dask tests failing with v2024.12.0 #638

Open
tomwhite opened this issue Dec 4, 2024 · 0 comments
Open

Dask tests failing with v2024.12.0 #638

tomwhite opened this issue Dec 4, 2024 · 0 comments
Labels
bug Something isn't working runtime

Comments

@tomwhite
Copy link
Member

tomwhite commented Dec 4, 2024

See https://github.com/cubed-dev/cubed/actions/runs/12164228442/job/33925369770

For example (running locally):

pytest -vs 'cubed/tests/test_array_api.py::test_astype[dask]'
========================================================== test session starts ===========================================================
platform darwin -- Python 3.10.15, pytest-8.3.4, pluggy-1.5.0 -- /Users/tom/miniforge3/envs/cubed-dask-m1/bin/python3.10
cachedir: .pytest_cache
rootdir: /Users/tom/workspace/cubed
configfile: pyproject.toml
plugins: cov-6.0.0, mock-3.14.0
collected 1 item                                                                                                                         

cubed/tests/test_array_api.py::test_astype[dask] FAILED

================================================================ FAILURES ================================================================
___________________________________________________________ test_astype[dask] ____________________________________________________________

spec = cubed.Spec(work_dir=/private/var/folders/9j/h1v35g4166z6zt816fq7wymc0000gn/T/pytest-of-tom/pytest-273/test_astype_dask_0, allowed_mem=100000, reserved_mem=0, executor=None, storage_options=None, zarr_compressor=default)
executor = <cubed.runtime.executors.dask.DaskExecutor object at 0x108115ba0>

    def test_astype(spec, executor):
        a = xp.asarray([[1, 2, 3], [4, 5, 6], [7, 8, 9]], chunks=(2, 2), spec=spec)
        b = xp.astype(a, xp.int32)
>       assert_array_equal(
            b.compute(executor=executor),
            np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]]),
        )

cubed/tests/test_array_api.py:153: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
../../miniforge3/envs/cubed-dask-m1/lib/python3.10/site-packages/numpy/_utils/__init__.py:85: in wrapper
    return fun(*args, **kwargs)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

args = (<built-in function eq>, array([[1, 2, 0],
       [4, 5, 0],
       [7, 8, 9]], dtype=int32), array([[1, 2, 3],
       [4, 5, 6],
       [7, 8, 9]]))
kwds = {'err_msg': '', 'header': 'Arrays are not equal', 'strict': False, 'verbose': True}

    @wraps(func)
    def inner(*args, **kwds):
        with self._recreate_cm():
>           return func(*args, **kwds)
E           AssertionError: 
E           Arrays are not equal
E           
E           Mismatched elements: 2 / 9 (22.2%)
E           Max absolute difference among violations: 6
E           Max relative difference among violations: 1.
E            ACTUAL: array([[1, 2, 0],
E                  [4, 5, 0],
E                  [7, 8, 9]], dtype=int32)
E            DESIRED: array([[1, 2, 3],
E                  [4, 5, 6],
E                  [7, 8, 9]])

../../miniforge3/envs/cubed-dask-m1/lib/python3.10/contextlib.py:79: AssertionError
----------------------------------------------------------- Captured log call ------------------------------------------------------------
INFO     distributed.http.proxy:proxy.py:85 To route to workers diagnostics web server please install jupyter-server-proxy: python -m pip install jupyter-server-proxy
INFO     distributed.scheduler:scheduler.py:1750 State start
INFO     distributed.scheduler:scheduler.py:4220   Scheduler at:     tcp://127.0.0.1:64249
INFO     distributed.scheduler:scheduler.py:4235   dashboard at:  http://127.0.0.1:8787/status
INFO     distributed.scheduler:scheduler.py:8114 Registering Worker plugin shuffle
INFO     distributed.nanny:nanny.py:368         Start Nanny at: 'tcp://127.0.0.1:64252'
INFO     distributed.nanny:nanny.py:368         Start Nanny at: 'tcp://127.0.0.1:64253'
INFO     distributed.nanny:nanny.py:368         Start Nanny at: 'tcp://127.0.0.1:64254'
INFO     distributed.nanny:nanny.py:368         Start Nanny at: 'tcp://127.0.0.1:64255'
INFO     distributed.scheduler:scheduler.py:4574 Register worker addr: tcp://127.0.0.1:64263 name: 1
INFO     distributed.scheduler:scheduler.py:6168 Starting worker compute stream, tcp://127.0.0.1:64263
INFO     distributed.core:core.py:883 Starting established connection to tcp://127.0.0.1:64265
INFO     distributed.scheduler:scheduler.py:4574 Register worker addr: tcp://127.0.0.1:64267 name: 3
INFO     distributed.scheduler:scheduler.py:6168 Starting worker compute stream, tcp://127.0.0.1:64267
INFO     distributed.core:core.py:883 Starting established connection to tcp://127.0.0.1:64271
INFO     distributed.scheduler:scheduler.py:4574 Register worker addr: tcp://127.0.0.1:64260 name: 0
INFO     distributed.scheduler:scheduler.py:6168 Starting worker compute stream, tcp://127.0.0.1:64260
INFO     distributed.core:core.py:883 Starting established connection to tcp://127.0.0.1:64262
INFO     distributed.scheduler:scheduler.py:4574 Register worker addr: tcp://127.0.0.1:64266 name: 2
INFO     distributed.scheduler:scheduler.py:6168 Starting worker compute stream, tcp://127.0.0.1:64266
INFO     distributed.core:core.py:883 Starting established connection to tcp://127.0.0.1:64269
INFO     distributed.scheduler:scheduler.py:5922 Receive client connection: Client-158a4e76-b266-11ef-8c05-66948709f435
INFO     distributed.core:core.py:883 Starting established connection to tcp://127.0.0.1:64272
INFO     distributed.scheduler:scheduler.py:5967 Remove client Client-158a4e76-b266-11ef-8c05-66948709f435
INFO     distributed.core:core.py:908 Received 'close-stream' from tcp://127.0.0.1:64272; closing.
INFO     distributed.scheduler:scheduler.py:5967 Remove client Client-158a4e76-b266-11ef-8c05-66948709f435
INFO     distributed.scheduler:scheduler.py:5959 Close client connection: Client-158a4e76-b266-11ef-8c05-66948709f435
INFO     distributed.scheduler:scheduler.py:7553 Retire worker addresses (stimulus_id='retire-workers-1733333700.0394292') (0, 1, 2, 3)
INFO     distributed.nanny:nanny.py:611 Closing Nanny at 'tcp://127.0.0.1:64252'. Reason: nanny-close
INFO     distributed.nanny:nanny.py:858 Nanny asking worker to close. Reason: nanny-close
INFO     distributed.nanny:nanny.py:611 Closing Nanny at 'tcp://127.0.0.1:64253'. Reason: nanny-close
INFO     distributed.nanny:nanny.py:858 Nanny asking worker to close. Reason: nanny-close
INFO     distributed.nanny:nanny.py:611 Closing Nanny at 'tcp://127.0.0.1:64254'. Reason: nanny-close
INFO     distributed.nanny:nanny.py:858 Nanny asking worker to close. Reason: nanny-close
INFO     distributed.nanny:nanny.py:611 Closing Nanny at 'tcp://127.0.0.1:64255'. Reason: nanny-close
INFO     distributed.nanny:nanny.py:858 Nanny asking worker to close. Reason: nanny-close
INFO     distributed.core:core.py:908 Received 'close-stream' from tcp://127.0.0.1:64265; closing.
INFO     distributed.scheduler:scheduler.py:5431 Remove worker addr: tcp://127.0.0.1:64263 name: 1 (stimulus_id='handle-worker-cleanup-1733333700.041016')
INFO     distributed.core:core.py:908 Received 'close-stream' from tcp://127.0.0.1:64262; closing.
INFO     distributed.scheduler:scheduler.py:5431 Remove worker addr: tcp://127.0.0.1:64260 name: 0 (stimulus_id='handle-worker-cleanup-1733333700.041572')
INFO     distributed.core:core.py:908 Received 'close-stream' from tcp://127.0.0.1:64269; closing.
INFO     distributed.scheduler:scheduler.py:5431 Remove worker addr: tcp://127.0.0.1:64266 name: 2 (stimulus_id='handle-worker-cleanup-1733333700.041902')
INFO     distributed.core:core.py:908 Received 'close-stream' from tcp://127.0.0.1:64271; closing.
INFO     distributed.scheduler:scheduler.py:5431 Remove worker addr: tcp://127.0.0.1:64267 name: 3 (stimulus_id='handle-worker-cleanup-1733333700.04224')
INFO     distributed.scheduler:scheduler.py:5566 Lost all workers
INFO     distributed.nanny:nanny.py:626 Nanny at 'tcp://127.0.0.1:64255' closed.
INFO     distributed.nanny:nanny.py:626 Nanny at 'tcp://127.0.0.1:64253' closed.
INFO     distributed.nanny:nanny.py:626 Nanny at 'tcp://127.0.0.1:64254' closed.
INFO     distributed.nanny:nanny.py:626 Nanny at 'tcp://127.0.0.1:64252' closed.
INFO     distributed.scheduler:scheduler.py:4284 Closing scheduler. Reason: unknown
INFO     distributed.scheduler:scheduler.py:4312 Scheduler closing all comms
======================================================== short test summary info =========================================================
FAILED cubed/tests/test_array_api.py::test_astype[dask] - AssertionError: 
=========================================================== 1 failed in 1.83s ============================================================
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working runtime
Projects
None yet
Development

No branches or pull requests

1 participant