zero323
zero323
https://github.com/zero323/pyspark-asyncactions > A proof of concept asynchronous actions for PySpark using concurent.futures Originally developed as proof-of-concept solution for SPARK-20347
https://github.com/arangodb/arangodb-spark-connector Right now it is a tad useless (doesn't support data source API), maybe it'll change in the future.
[spark-timeseries](https://github.com/sryza/spark-timeseries) is officially no longer developed: https://github.com/sryza/spark-timeseries/commit/17b78f7d62f4262e0d8b5d9ee7aec2bf221f1f07. Should we keep it in the awesome?
It seems like at this point retype doesn't support `overloaded` annotations. Tested with example annotation (`foo.pyi`): ```python from typing import overload @overload def f(x: str) -> str: ... @overload def...
In particular - It is not enforced. - It is used for execution optimizations. - Reflection and `Product` types (`Option` as a way to indicate `nullable` fields`). This can work...
It would be great to add a short description of the repository and individual examples. Something like --- ## Spark Examples (adjust to reflect new project name) _These are awesome...
I am trying to investigate some issues related to caching behavior. When testing project with complex dependencies, I see serious performance degradation (roughly 20 fold on just 36 tests) compared...