[Bug] Sink multi-parallelism error hive lock
Search before asking
- [x] I searched in the issues and found nothing similar.
Paimon version
1.1.1
Compute Engine
flink 1.20.1
Minimal reproduce step
CREATE TABLE t_changelog_input ( age BIGINT, money BIGINT, name STRING, PRIMARY KEY (name) NOT ENFORCED )WITH ( "bucket" = "4", 'file.compression' = 'snappy', 'merge-engine' = 'deduplicate', 'changelog-producer' = 'input', "sink.parallelism" = "4" );
The table is created as above. When a large amount of data is written or a large amount of data is changed, the following error will appear. However, when sink.parallelism is 1, it will not appear. My data is stored on S3 and uses hivemetastore 3.1.2.Is there something wrong with my usage? Is there any way to increase the write parallelism?
What doesn't meet your expectations?
I hope that even when the data is stored on S3, which does not have atomic storage, it can still support multiple concurrent writes.
Anything else?
No response
Are you willing to submit a PR?
- [ ] I'm willing to submit a PR!