cloud-provider-azure icon indicating copy to clipboard operation
cloud-provider-azure copied to clipboard

Consistent Use of Context

Open devigned opened this issue 3 years ago • 7 comments

Is your feature request related to a problem?/Why is this needed Much of the context usage in the project is ineffectual. For example, the Azure DiskClient.Get uses a context, but the context provided to the method is largely ineffectual. As you can see here, a call to AttachDisk first calls getContextWithCancel, which simply constructs a context.Background with cancelation. The cancellation is only called in a defer when calling AzureDisk.Get. The cancellation provides no benefit as a way of cancelling processing further down the tree of execution.

Describe the solution you'd like in detail Context should be used to time bound execution hierarchies, and are best when they start from the root of execution. Perhaps, having a context started at the CLI entry point where OS signals are used to establish the initial lifespan of the root context.

Further, context should be exposed upward from any I/O dependent task. If a network request or any other long running activity is called, the context should be required and passed in from higher up in the call hierarchy.

Additional context By setting a consistent approach to passing context, the application will also be able to be more easily instrumented with tracing. For example, Cluster API Provider Azure uses OpenTelemetry to create a distributed trace of the entire reconciliation loop for all resources, https://github.com/kubernetes-sigs/cluster-api-provider-azure/blob/8a44c6d6f7b394b7929c3ae5b0e79730609c6cf1/controllers/azurecluster_controller.go#L114-L120. This makes it super easy to observe what the operator is doing and provides insight when trying to diagnosis issues. In fact, in CAPZ, we are exporting the traces to Application Insights to visualize the running operator. We'd love to be able to something similar in Cloud Provider Azure.

/cc @craiglpeters

devigned avatar May 28 '21 18:05 devigned

Thanks @devigned for reporting the issue and sharing the ideas to provide better observability. Would you like to contribute the OpenTelemetry support?

feiskyer avatar May 30 '21 03:05 feiskyer

Definitely a possibility. I'm discussing this with @ritazh.

First thing I wanted to do is open the issue and start the discussion. I'd suspect an enhancement doc would be proper for such a change.

The major items that would need to be changed:

  • Pipe context.Context from process entry point down through the call hierarchy to any function that may have i/o or an operation that should gracefully cancel.
  • Add OpenTelemetry instrumentation throughout the call hierarchy
  • Document the usage

wdyt?

devigned avatar Jun 01 '21 16:06 devigned

@devigned thanks, LGTM

feiskyer avatar Jun 02 '21 01:06 feiskyer

The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle stale
  • Mark this issue or PR as rotten with /lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

k8s-triage-robot avatar Sep 05 '21 14:09 k8s-triage-robot

@devigned are you still available to work on this?

feiskyer avatar Sep 06 '21 01:09 feiskyer

Yes, but it there are a couple things ahead of this. I think I'll be able to start it toward the end of the month.

devigned avatar Sep 07 '21 14:09 devigned

The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle stale
  • Mark this issue or PR as rotten with /lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

k8s-triage-robot avatar Dec 06 '21 15:12 k8s-triage-robot