cubed icon indicating copy to clipboard operation
cubed copied to clipboard

Executor for Apache Spark

Open rbavery opened this issue 1 year ago • 7 comments

Could Spark be added as a supported executor?

Maybe RDD.map or RDD.mapPartitions would be the correct way to map a function similar to map_unordered in the Lithops executor.

https://spark.apache.org/docs/latest/api/python/reference/api/pyspark.RDD.mapPartitions.html#pyspark.RDD.mapPartitions

To support this a guess would need to be made up front on the reserved memory available for python UDFs. It sounds like currently this would be done globally but maybe later could be done on a per-operator basis?

rbavery avatar Jul 14 '24 22:07 rbavery