ClickHouse
ClickHouse copied to clipboard
ClickHouse® is a free analytics DBMS for big data
Use metadata cache (min/max, bloom filter) to apply file filtering (depends on #480 ). This is generic for external table engines
See https://www.dremio.com/blog/puffins-and-icebergs-additional-stats-for-apache-iceberg-tables/
```INSERT INTO s3(‘s3:///myfiles*.parquet’)``` That will automatically split data into multiple files, use existing min_insert_block_size_rows/bytes. Should close https://github.com/ClickHouse/ClickHouse/issues/41537 For example, other systems implement it as follows: BigQuery The path must contain...
``` CREATE TABLE Engine = DeltaLake(...) CREATE TABLE Engine = Iceberg(...) ``` Current implementation assumes 3 versions of Iceberg. There should be only one. Also it should support named collections...