paimon icon indicating copy to clipboard operation
paimon copied to clipboard

[Bug] Sink multi-parallelism error hive lock

Open A-little-bit-of-data opened this issue 3 months ago • 0 comments

Search before asking

  • [x] I searched in the issues and found nothing similar.

Paimon version

1.1.1

Compute Engine

flink 1.20.1

Minimal reproduce step

CREATE TABLE t_changelog_input ( age BIGINT, money BIGINT, name STRING, PRIMARY KEY (name) NOT ENFORCED )WITH ( "bucket" = "4", 'file.compression' = 'snappy', 'merge-engine' = 'deduplicate', 'changelog-producer' = 'input', "sink.parallelism" = "4" );

The table is created as above. When a large amount of data is written or a large amount of data is changed, the following error will appear. However, when sink.parallelism is 1, it will not appear. My data is stored on S3 and uses hivemetastore 3.1.2.Is there something wrong with my usage? Is there any way to increase the write parallelism?

Image

What doesn't meet your expectations?

I hope that even when the data is stored on S3, which does not have atomic storage, it can still support multiple concurrent writes.

Anything else?

No response

Are you willing to submit a PR?

  • [ ] I'm willing to submit a PR!

A-little-bit-of-data avatar Sep 13 '25 10:09 A-little-bit-of-data