Joel
Joel
What about adding preview logic to the `AbstractDataSet` class? And then also implementing it for the pandas and spark datasets today? pandas -> `.head(5)` spark -> `.limit(5).toPandas().head()`
[It does](https://plotly.com/python/table/), but there are [better ones](https://hackernoon.com/10-data-table-libraries-for-javascript-5g263vdm) out there which support things I think our users would like filtering, excel export, text search etc. ...Question for you all if it's...
I'm going to reopen this - since users on Dicsord have been asking for the same thing
I would like to push that we also get into the habit of doing smaller, more frequent patch releases for this sort of thing.
Interesting so if I'm reading this right, you would like to have user defined pipeline templates? I think this is something we'd be interested in building but will take a...
So I think this would be really neat for some of the custom things people do in 3 main areas: - Custom DataSets - Hooks - CLI.py Slightly less useful...
Hi @Cortysus - Kedro supports dask in it's [own dataset](https://kedro.readthedocs.io/en/stable/kedro.extras.datasets.dask.ParquetDataSet.html) could you try saving to parquet via Pandas and transcoding to dask using this dataset. That way it's not the...
Hi this isn't maintained by the Kedro team - but this repo I've found looks like it implements things in the right way https://github.com/blaizeberry4/lastfm-recs-sandbox/blob/3f997ae2301d6108c72df8dda325c101398b77c9/src/etl/pipelines/eda.py
Ah have you tried using the `kedro run --runner=ThreadRunner` I think for remote execution environments like Spark and Dask you can't use ParallelRunner/Multiprocessing for that reason.
If you're using ThreadRunner we shouldn't be triggering any multi-processing from the Kedro side, so I'm not sure what going on her? Can you post the screenshot of the terminal...