KnightChess
KnightChess
@hudi-bot run azure
@hudi-bot run azure
@nsivabalan @danny0405 @yihua hi, ci all sucess, can you help revie it.
@danny0405 yes, is ready for review
@hudi-bot run azure
@smileyboy2019 you need point profiles `spark3.4` pom.xml in idea
@zhangjw123321 look like `hoodie.bulkinsert.shuffle.parallelism` can not work on non-partitioned table in the code. In the spark ui, may be you not set `spark.default.parallelism` so `reduceBykey` will use the parent rdd...
@zhangjw123321 I test in my local, `spark.default.parallelism` look like can not effect in sql set, can you set when submit spark job, like --conf. before try it, how much cores...
@zhangjw123321 you can try set it in spark submit, --conf, or by code sparkconf.set('xxx','yyy'), will match other branch, not use parent rdd partition size 
@zhangjw123321 I create a issue to track it, https://issues.apache.org/jira/browse/HUDI-7277