[Minor] Extend the Parquet writer's dictionary encoding benchmark.
Description
This PR extends the data cardinality and run length range for the existing parquet writer's encoding benchmark.
Checklist
- [x] I am familiar with the Contributing Guidelines.
- [x] New or existing tests cover these changes.
- [x] The documentation is up to date with these changes.
This pull request requires additional validation before any workflows can run on NVIDIA's runners.
Pull request vetters can view their responsibilities here.
Contributors can view more details about this message here.
/ok to test
what's the reason for this change?
what's the reason for this change?
First of all welcome back. Greg wanted me to push any updates I did to the benchmark for #16541. Though I think that all my local changes (even wider extended ranges) need not to be pushed upstream if not needed.
/ok to test
/ok to test
/ok to test
/merge
Would be nice to know how much time this increases in benchmark runs. If it is not available now, follow up with Randy on benchmark runs.
Results in #16541 (here) for which we are extending this. Each new benchmark in the matrix takes roughly 0.5s to run on my workstation (AMD Threadripper + RTX Ada 5880) so it should be roughly an 4s increase in total time (8x new benchmarks).