vector
vector copied to clipboard
chore(deps): Update aws-config crate
Also adds the aws-runtime crate as some types were moved there.
Closes: https://github.com/vectordotdev/vector/issues/20662
Datadog Report
Branch report: jszwedko/update-aws-config
Commit report: cc7a433
Test service: vector
:white_check_mark: 0 Failed, 7 Passed, 0 Skipped, 25.44s Total Time
@jszwedko could you please check the failing checks so this can get merged?
@jszwedko could you please check the failing checks so this can get merged?
Apologies for the delay. I had looked a bit before, the integration tests for AWS are failing, but I wasn't able to sort it out 😓 I'll try to take another pass soon.
@jszwedko maybe the issue is with the integration test itself? I am just guessing here, because I don't know what mockwatchlogs image does, but it seems like it is used to emulate AWS Cloudwatch, because back in the days (that container is 5 years old) localstack didn't support Cloudwatch Logs?
Using localstack for all Cloudwatch tests (metrics are using localstack) could do the trick.
Datadog Report
Branch report: jszwedko/update-aws-config
Commit report: fbd978c
Test service: vector
:white_check_mark: 0 Failed, 7 Passed, 0 Skipped, 25.47s Total Time
@jszwedko maybe the issue is with the integration test itself? I am just guessing here, because I don't know what
mockwatchlogsimage does, but it seems like it is used to emulate AWS Cloudwatch, because back in the days (that container is 5 years old)localstackdidn't support Cloudwatch Logs? Usinglocalstackfor all Cloudwatch tests (metrics are usinglocalstack) could do the trick.
Good observation! I opened a PR to switch from mockwatchlogs to localstack here: https://github.com/vectordotdev/vector/pull/21114. The integration tests are currently failing, but if we can fix that that might unblock this. I'll try to look at it again soon, but also happy to see someone else take a look if they have time.
@jszwedko now after we fixed the integration tests in another PR, could you please rebase this branch and see if it works now?
@jszwedko now after we fixed the integration tests in another PR, could you please rebase this branch and see if it works now?
👍 thanks for the bump on this. I merged in master 🤞
Looks like we are still seeing:
2024-09-06T20:47:02.214518Z WARN request{request_id=1}: vector::sinks::util::retries: Retrying after error. error=CloudwatchError::DescribeLogStreams: dispatch failure internal_log_rate_limit=true
I'll try to dig into this later unless someone else gets to it first. It may just be a matter of a missing feature flag.
Looks like we are still seeing:
2024-09-06T20:47:02.214518Z WARN request{request_id=1}: vector::sinks::util::retries: Retrying after error. error=CloudwatchError::DescribeLogStreams: dispatch failure internal_log_rate_limit=trueI'll try to dig into this later unless someone else gets to it first. It may just be a matter of a missing feature flag.
@jszwedko I wonder if the logging should be improved here, there's also typically a message in the AWS error response, which tells what's actually wrong. vector never logs this. This makes it harder to diagnose any issue with the AWS config, like missing permissions. I haven't had the bandwidth to report a proper issue on this.
Looks like we are still seeing:
2024-09-06T20:47:02.214518Z WARN request{request_id=1}: vector::sinks::util::retries: Retrying after error. error=CloudwatchError::DescribeLogStreams: dispatch failure internal_log_rate_limit=trueI'll try to dig into this later unless someone else gets to it first. It may just be a matter of a missing feature flag.
@jszwedko I wonder if the logging should be improved here, there's also typically a message in the AWS error response, which tells what's actually wrong. vector never logs this. This makes it harder to diagnose any issue with the AWS config, like missing permissions. I haven't had the bandwidth to report a proper issue on this.
I believe "dispatch failure" is an error returned by the SDK before it even makes the request, and so there wouldn't be a response from AWS, but I could be wrong.
FYI dispatch failure internal_log_rate_limit=true doesn't appear to be new / introduced by this PR:
- https://github.com/vectordotdev/vector/issues/19636
- https://github.com/vectordotdev/vector/issues/20314#issuecomment-2124333299
- https://github.com/vectordotdev/vector/issues/21339
FYI
dispatch failure internal_log_rate_limit=truedoesn't appear to be new / introduced by this PR:
It is new in that the integration tests previously didn't hit it, but it does seem like it can potentially be hit in other circumstances too (thanks fro the links!).
rip, it looks like the integration tests are failing. I don't know Rust at all so sorry if this is an obvious statement, but would it be possible to wrap the error being returned with https://docs.rs/aws-sdk-cloudwatch/latest/aws_sdk_cloudwatch/error/struct.DisplayErrorContext.html so we can get more debug output?
rip, it looks like the integration tests are failing. I don't know Rust at all so sorry if this is an obvious statement, but would it be possible to wrap the error being returned with https://docs.rs/aws-sdk-cloudwatch/latest/aws_sdk_cloudwatch/error/struct.DisplayErrorContext.html so we can get more debug output?
There are some details in the CI logs:
thread 'sinks::aws_cloudwatch_logs::integration_tests::cloudwatch_healthcheck' panicked at src/sinks/aws_cloudwatch_logs/integration_tests.rs:461:39:
called `Result::unwrap()` on an `Err` value: DescribeLogGroupsFailed { source: DispatchFailure(DispatchFailure { source: ConnectorError { kind: Other(None), source: InterceptorError { kind: ReadAfterSerialization, interceptor_name: Some("UserAgentInterceptor"), source: Some(MissingApiMetadata) }, connection: Unknown } }) }
I think this needs further debugging.
updated as part of https://github.com/vectordotdev/vector/pull/22687