kube-mgmt
kube-mgmt copied to clipboard
Sidecar for managing OPA instances in Kubernetes.
kube-mgmt
Policy-based control for Kubernetes deployments.
About
kube-mgmt manages policies / data of Open Policy Agent
instances in Kubernetes.
Use kube-mgmt to:
- Load policies and/or static data into OPA via
ConfigMap. - Replicate Kubernetes resources including CustomResourceDefinitions (CRDs) into OPA.
Deployment Guide
Both OPA and kube-mgmt can be installed using Helm chart.
Follow README to install it into K8s cluster.
Policies and Data loading
kube-mgmt automatically discovers policies and JSON data
stored in ConfigMaps in Kubernetes and loads them into OPA.
Policies or data can be disabled using --enable-policy=false or --enable-data=false flags respectively.
kube-mgmt assumes a ConfigMap contains policy or JSON data if the ConfigMap is:
- Created in a namespace listed in the
--namespacesoption. If you specify--namespaces=*thenkube-mgmtwill look for policies in ALL namespaces. - Labelled with
openpolicyagent.org/policy=regofor policies - Labelled with
openpolicyagent.org/data=opfor JSON data
Label names and their values can be configured using --policy-label, --policy-value, --data-label, --data-value CLI options.
When a ConfigMap has been successfully loaded into OPA,
the openpolicyagent.org/policy-status annotation is set to {"status": "ok"}.
If loading fails for some reason (e.g., because of a parse error), the
openpolicyagent.org/policy-status annotation is set to {"status": "error", "error": ...}
where the error field contains details about the failure.
Data loaded out of ConfigMaps is laid out as follows:
<namespace>/<name>/<key>
For example, if the following ConfigMap was created:
kind: ConfigMap
apiVersion: v1
metadata:
name: hello-data
namespace: opa
labels:
openpolicyagent.org/data: opa
data:
x.json: |
{"a": [1,2,3,4]}
Note: "x.json" may be any key.
You could refer to the data inside your policies as follows:
data.opa["hello-data"]["x.json"].a[0] # evaluates to 1
Note: "opa" is the namespace for the configMap.
You may mock this in a test like other objects: with data.opa as my_mocked_object.
Caching
kube-mgmt can be configured to replicate Kubernetes resources into OPA so that
you can express policies over an eventually consistent cache of Kubernetes
state.
Replication is enabled with the following options:
# Replicate namespace-level resources. May be specified multiple times.
--replicate=<[group/]version/resource>
# Replicate cluster-level resources. May be specified multiple times.
--replicate-cluster=<[group/]version/resource>
Kubernetes resources replicated into OPA are laid out as follows:
<replicate-path>/<resource>/<namespace>/<name> # namespace scoped
<replicate-path>/<resource>/<name> # cluster scoped
<replicate-path>is configurable (via--replicate-path) and defaults tokubernetes.<resource>is the Kubernetes resource plural, e.g.,nodes,pods,services, etc.<namespace>is the namespace of the Kubernetes resource.<name>is the name of the Kubernetes resource.
For example, to search for services with the label "foo" you could write:
some namespace, name
service := data.kubernetes.services[namespace][name]
service.metadata.labels["foo"]
An alternative way to visualize the layout is as single JSON document:
{
"kubernetes": {
"services": {
"default": {
"example-service": {...},
"another-service": {...},
}
}
}
}
}
The example below would replicate Deployments, Services, and Nodes into OPA:
--replicate=apps/v1beta/deployments
--replicate=v1/services
--replicate-cluster=v1/nodes
Custom Resource Definitions (CRDs)
kube-mgmt can also be configured to replicate Kubernetes Custom Resources using the --replicate and --replicate-cluster options. For an example of how OPA can be used to enforce admission control polices on Kubernetes custom resources see Admission Control For Custom Resources
Admission Control
To get started with admission control policy enforcement in Kubernetes 1.9 or later see the Kubernetes Admission Control tutorial. For older versions of Kubernetes, see Admission Control (1.7).
In the Kubernetes Admission Control tutorial, OPA is NOT running with an authorization policy configured and hence clients can read and write policies in OPA. When deploying OPA in an insecure environment, it is recommended to configure authentication and authorization on the OPA daemon. For an example of how OPA can be securely deployed as an admission controller see Admission Control Secure.
OPA API Endpoints and Least-privilege Configuration
kube-mgmt is a privileged component that can load policy and data into OPA.
Other clients connecting to the OPA API only need to query for policy decisions.
To load policy and data into OPA, kube-mgmt uses the following OPA API
endpoints:
PUT v1/policy/<path>- upserting policiesDELETE v1/policy/<path>- deleting policiesPUT v1/data/<path>- upserting dataPATCH v1/data/<path>- updating and removing data
Many users configure OPA with a simple API authorization policy that restricts access to the OPA APIs:
package system.authz
# Deny access by default.
default allow = false
# Allow anonymous access to decision `data.example.response`
#
# NOTE: the specific decision differs depending on your policies.
# NOTE: depending on how callers are configured, they may only require this or the default decision below.
allow {
input.path == ["v0", "data", "example", "response"]
input.method == "POST"
}
# Allow anonymous access to default decision.
allow {
input.path == [""]
input.method == "POST"
}
# This is only used for health check in liveness and readiness probe
allow {
input.path == ["health"]
input.method == "GET"
}
# This is only used for prometheus metrics
allow {
input.path == ["metrics"]
input.method == "GET"
}
# This is used by kube-mgmt to PUT/PATCH against /v1/data and PUT/DELETE against /v1/policies.
#
# NOTE: The $TOKEN value is replaced at deploy-time with the actual value that kube-mgmt will use. This is typically done by an initContainer.
allow {
input.identity == "$TOKEN"
}
Development Guide
This project uses excellent tool Just for buiding.
It is configured by justfile file in root directory.
All available targets can be inspected by running just in command line.
Release procedure
To release a new version - just create GitHub release with corresponding tag, following semantic version converntion.
As soon as tag will be pushed - CI pipeline will build and publish all artifacts.