KnightChess

Results 70 comments of KnightChess

@nsivabalan @danny0405 @yihua hi, ci all sucess, can you help revie it.

@smileyboy2019 you need point profiles `spark3.4` pom.xml in idea

@zhangjw123321 look like `hoodie.bulkinsert.shuffle.parallelism` can not work on non-partitioned table in the code. In the spark ui, may be you not set `spark.default.parallelism` so `reduceBykey` will use the parent rdd...

@zhangjw123321 I test in my local, `spark.default.parallelism` look like can not effect in sql set, can you set when submit spark job, like --conf. before try it, how much cores...

@zhangjw123321 you can try set it in spark submit, --conf, or by code sparkconf.set('xxx','yyy'), will match other branch, not use parent rdd partition size ![image](https://github.com/apache/hudi/assets/20125927/4b21cb55-3bd6-471e-92d4-e3dade5eafaf)

@zhangjw123321 I create a issue to track it, https://issues.apache.org/jira/browse/HUDI-7277