paasta icon indicating copy to clipboard operation
paasta copied to clipboard

Pass aws credentials to spark executors as environment variables

Open 88manpreet opened this issue 4 years ago • 2 comments

88manpreet avatar Apr 22 '20 23:04 88manpreet

@huadongliu can you review this one? Doesn't "--aws-credentials-yaml" do this for users already?

solarkennedy avatar Apr 23 '20 00:04 solarkennedy

@huadongliu can you review this one? Doesn't "--aws-credentials-yaml" do this for users already?

We only do it for spark driver, the change is for spark executor. Normally spark executor doesn't need this env. This env is only helpful when people trying to access AWS resources inside the executor. For native spark script, we don't recommend people access AWS resource inside executor, because it is higher chance to hit the AWS api limit. However, it looks like this is the quick fix for mrjob harness script, so maybe go with this fix first, and if people do see the s3 limit error, we can discuss with the team for the workaround.

lydian avatar Apr 23 '20 00:04 lydian

Cleaning up and closing some very old PRs. Please re-open or nudge me if you’re still planning to work on this.

mattmb avatar Feb 21 '24 11:02 mattmb