dask-ml icon indicating copy to clipboard operation
dask-ml copied to clipboard

Implementation for make_s_curve

Open CC-Yeh opened this issue 2 years ago • 2 comments

Hi guys,

I implemented the make_s_curve function and corresponding tests for it in https://github.com/dask/dask-ml/pull/967.

However, one of the tests is failing consistently, even though I have verified the values in the outputs and the computation graph. Appreciate any feedback or suggestions on how to resolve this issue.

The test that failed.

@pytest.mark.parametrize(
    "generator",
    [
        dask_ml.datasets.make_blobs,
        dask_ml.datasets.make_classification,
        dask_ml.datasets.make_counts,
        dask_ml.datasets.make_regression,
        dask_ml.datasets.make_s_curve,
    ],
)
def test_deterministic(generator, scheduler):
    a, t = generator(chunks=100, random_state=10)
    b, u = generator(chunks=100, random_state=10)
    assert_eq(a, b)
    assert_eq(t, u)

Environment:

Dask version: 2023.3.2 Dask_ml versuib: 2022.5.27 Python version: 3.10.9 Operating System: OSX Install method (conda, pip, source): pip

Reproducible example

import numpy as np
import dask.array as da
from dask.array.utils import assert_eq
import dask_ml

def make_s_curve(
    n_samples=100,
    noise=0.0,
    random_state=None,
    chunks=None,
):
    rng = dask_ml.utils.check_random_state(random_state)

    t_scale = 3 * np.pi * 0.5
    t = rng.uniform(low=-t_scale, high=t_scale, size=(n_samples), chunks=(chunks,))
    X = da.empty(shape=(n_samples, 3), chunks=(chunks, 3), dtype="f8")
    X[:, 0] = da.sin(t)
    X[:, 1] = rng.uniform(low=0, high=2, size=n_samples, chunks=(chunks,))
    X[:, 2] = da.sign(t) * (da.cos(t) - 1)

    if noise > 0:
        X += rng.normal(scale=noise, size=X.shape, chunks=X.chunks)

    return X, t


if __name__ == '__main__':
    a, t = make_s_curve(chunks=100, random_state=10)
    b, u = make_s_curve(chunks=100, random_state=10)
    assert_eq(a, b)
    assert_eq(t, u)

Traceback

Traceback (most recent call last):
  File "/Users/dask_mltest.py", line 30, in <module>
    assert_eq(a, b)
  File "/Users/venv/lib/python3.9/site-packages/dask/array/utils.py", line 304, in assert_eq
    a, adt, a_meta, a_computed = _get_dt_meta_computed(
  File "/Users/venv/lib/python3.9/site-packages/dask/array/utils.py", line 259, in _get_dt_meta_computed
    _check_dsk(x.dask)
  File "/Users/venv/lib/python3.9/site-packages/dask/array/utils.py", line 210, in _check_dsk
    assert not non_one, non_one
AssertionError: {('uniform-c022afab3445a4ac294ad46da60634e2', 0): 2}

CC-Yeh avatar Apr 03 '23 09:04 CC-Yeh

Thanks. It might be a while before I can look at this.

_check_dsk is verifying things about your task graph. It looks like this is saying you have multiple layers that use the same key. I'm not sure if you're making your own layers / keys, but if you are you'll want to use dask.base.tokenize on all the arguments going into whatever that function is.

TomAugspurger avatar Apr 16 '23 13:04 TomAugspurger

Thanks for your reply.

Yes, _check_dsk has a side effect of checking an overlapping key in the task graph. But it's the nature of this computation.

Task Graph and Details of Layers

taskgraph

The items of layers from the output

Layer: empty_like-848136d20f353a621c3fb6120e93a97c
('empty_like-848136d20f353a621c3fb6120e93a97c', 0, 0)

Layer: setitem-af567baa0d133fb2414b22009feedc71
('uniform-92c494621cf5b5df81f5fc85445fdfbf', 0)
('sin-b72ce9449b1942fdf9eade38a01fb89e', 0)
('setitem-af567baa0d133fb2414b22009feedc71', 0, 0)

Layer: setitem-c722836162c6db55933255d42e1117d9
('uniform-cd1ff3741c1350d3149e974c0a71b242', 0)
('setitem-c722836162c6db55933255d42e1117d9', 0, 0)

Layer: setitem-63195b1fdf99be7d4bf37a116ed79ef1
('mul-d7bb22edc92ff9263d91f04ac910d2d8', 0)
('uniform-92c494621cf5b5df81f5fc85445fdfbf', 0)
('sign-6d3aa1923db2461ffc8668e5d496fcbc', 0)
('cos-514e590b2a8b5ed03c3f5dc4551ff230', 0)
('sub-5397dd1b68f17458187bee3512c849ca', 0)
('setitem-63195b1fdf99be7d4bf37a116ed79ef1', 0, 0)

We could observe that ('uniform-92c494621cf5b5df81f5fc85445fdfbf', 0) was used in layer 2 and 4, which is the expected behavior.

Also check sklearn make_s_curve

If we only want to check if outcomes are deterministic, maybe it's ok to skip checking the graph overlapping here?

CC-Yeh avatar Apr 21 '23 21:04 CC-Yeh