Daft
Daft copied to clipboard
Implementing hive-style read
When pyspark saves parquets to a folder on a partition, it creates folders of the partition=some_value. When I use daft to read_parquet the parent folder, I would like to get back the columns of the table which were used as partitions. It would be helpful if we could parse the key=value pairs from the hive partition paths back into columns.