Eduard Tudenhoefner
Eduard Tudenhoefner
@koombal did you enable `s3.remote-signing-enabled` by any chance? Against which REST catalog implementation is this running? It's also possible that the server set the `s3.remote-signing-enabled` flag by sending it back...
@koombal remote signing exists only for S3. For azure there are vended credentials. For S3 you might want to use vended creds as well (https://docs.lakekeeper.io/docs/0.10.x/storage/?h=remote#s3)
@c-thiel is this something you could please take a look at?
@c-thiel the Table is broadcasted in https://github.com/apache/iceberg/blob/24ca356fb4ecd48d593949fd25c852c21bc87d53/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/source/SparkMicroBatchStream.java#L105 and the Table's IO instance is set in https://github.com/apache/iceberg/blob/e667f64f5bddbacb1a641ac8ea67fc21a76e434d/core/src/main/java/org/apache/iceberg/SerializableTable.java#L86 That IO is then used in https://github.com/apache/iceberg/blob/24ca356fb4ecd48d593949fd25c852c21bc87d53/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/source/SparkMicroBatchStream.java#L558-L569
Yes I think it's correct that the checkpoint uses the Table's FileIO, otherwise I don't see what other options we have and what other FileIO it would use. It sounds...
@koombal can you please share your entire catalog configuration and also what does `ckptDir` point to in `.option("checkpointLocation", ckptDir)`? > @nastra any idea why scala uses the tables FileIO while...
@koombal does your checkpoint location use `s3://` or `s3a://`?
@talatuyarer can you review this one please?
> will it be part of iceberg 1.10.1 ? ( since spark 4.1.0 was release today ) thanks No this won't be part of 1.10.1 since 1.10.1 is a patch...
It's not entirely clear whether this issue is caused by Iceberg itself. So it would be good to have a reproducible test in Iceberg for this in `CatalogTests`/`TestRESTUtil` but also...