Authorizers¶
In kcp, a request has four different ways of being admitted:
- It can be made to one of the preconfigured paths that do not require authorization, like
/healthz. - It can be performed by a user in one of the configured always-allow groups, by default
system:masters. - It can pass through the RBAC chain and match configured Roles and ClusterRoles.
- It can be permitted by an external HTTPS webhook backend.
They are related in the following way:
graph TD
start(Request):::state --> main_alt[/one of\]:::or
main_alt --> aaga[Always Allow Groups Auth]
main_alt --> aapa[Always Allow Paths Auth]
main_alt --> rga[Required Groups Auth]
main_alt --> wa[Webhook Auth]
aaga --> decision(Decision):::state
aapa --> decision
subgraph "RBAC"
rga --> wca[Workspace Content Auth]
wca --> scrda[System CRD Auth]
scrda --> mppa[Max. Permission Policy Auth]
mppa --- mppa_alt[/one of\]:::or
mppa_alt --> lpa[Local Policy Auth]
mppa_alt --> gpa[Global Policy Auth]
mppa_alt --> bpa[Bootstrap Policy Auth]
end
lpa --> decision
gpa --> decision
bpa --> decision
wa --> decision
classDef state color:#F77
classDef or fill:none,stroke:none
Always Allow Paths Authorizer¶
Like in vanilla Kubernetes, this authorizer always grants access to the configured URL paths. This is used for the health and liveness checks of kcp.
Always Allow Groups Authorizer¶
This authorizer always permits access if the user is in one of the configured groups. By default this
only includes the system:masters group.
RBAC Chain¶
The primary authorization flow is handled by a sequence of RBAC-based authorizers that a request must satisfy all in order to be granted access.
The following authorizers work together to implement RBAC in kcp:
| Authorizer | Description |
|---|---|
| Workspace content authorizer | validates that the user has access permission to the workspace |
| Required groups authorizer | validates that the user is in the annotation-based list of groups required for a workspace |
| System CRD authorizer | prevents undesired updates to certain core resources, like the status subresource on APIBindings |
| Maximal permission policy authorizer | validates the maximal permission policy RBAC policy in the API exporter workspace |
| Local Policy authorizer | validates the RBAC policy in the workspace that is accessed |
| Global Policy authorizer | validates the RBAC policy in the workspace that is accessed across shards |
| Kubernetes Bootstrap Policy authorizer | validates the RBAC Kubernetes standard policy |
Required Groups Authorizer¶
A authorization.kcp.io/required-groups annotation can be added to a LogicalCluster
to specify which groups a user must be a member of to access the workspace.
The syntax uses semicolons (;) to separate OR conditions and commas (,) to separate AND conditions within each group set.
For example, <group1>,<group2>;<group3> means that a user must be a member of both <group1> AND <group2>, OR be a member of <group3>.
The annotation is copied onto sub-workspaces during workspace creation, but is then not updated automatically if it's changed.
Workspace Content Authorizer¶
The workspace content authorizer checks whether the user is granted access to the workspace.
Access is granted access through verb=access non-resource permission to / inside of the workspace.
The ClusterRole system:kcp:workspace:access is pre-defined which makes it easy
to give a user access through a ClusterRoleBinding inside of the workspace.
For example, to give a user user1 access, create the following ClusterRoleBinding:
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: example-access
subjects:
- kind: User
name: user1
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: system:kcp:workspace:access
To give a user user1 admin access, create the following ClusterRoleBinding:
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: example-admin
subjects:
- kind: User
name: user1
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: cluster-admin
Cross-Workspace¶
Users from other workspaces can be authorized to perform actions using RBAC, e.g. to bind and APIExport.
Service Accounts¶
A service-account defined in a workspace implicitly is granted access to it.
A service-account defined in a different workspace is NOT given access to it.
Note
By default, workspaces are only accessible to a user if they are in Ready phase. Workspaces that are initializing
can be accessed only by users that are granted admin verb on the workspaces/content resource in the
parent workspace.
Service accounts declared within a workspace don't have access to initializing workspaces.
Foreign Users¶
Users can either be global kcp users or users that originate from a specific workspace, e.g. through an OIDC provider configured for that workspace.
Users that do not originate from the workspace the request is being made to and are not global kcp users are only visible as user system:anonymous
with groups system:authenticated and system:cluster:<logical-cluster>, where <logical-cluster> is the name of the logical cluster backing
the workspace they originated from.
System CRD Authorizer¶
This small authorizer simply prevents updates to the status subresource on APIExports or APIBindings. Note that this authorizer does not validate changes to the CustomResourceDefinitions themselves, but to objects from those CRDs instead.
Maximal Permission Policy Authorizer¶
If the requested resource type is part of an API binding, then this authorizer verifies that the request is not exceeding the maximum permission policy of the related API export. Currently, the "local policy" maximum permission policy type is supported.
Local Policy¶
The local maximum permission policy delegates the decision to the RBAC of the related API export.
To distinguish between local RBAC role bindings in that workspace and those for this these maximum permission policy,
every name and group is prefixed with apis.kcp.io:binding:.
Example: Given an APIBinding for type foo declared in workspace consumer that refers to an APIExport declared in workspace provider
and a user user-1 having the group group-1 requesting a create of foo in the default namespace in the consumer workspace,
this authorizer verifies that user-1 is allowed to execute this request by delegating to provider's RBAC using prefixed attributes.
Here, this authorizer prepends the apis.kcp.io:binding: prefix to the username and all groups the user belongs to.
Using prefixed attributes prevents RBAC collisions i.e. if user-1 is granted to execute requests within the provider workspace directly.
For the given example RBAC request looks as follows:
- Username:
apis.kcp.io:binding:user-1 - Groups: [
apis.kcp.io:binding:group-1] - Resource:
foo - Namespace:
default - Workspace:
provider - Verb:
create
The following Role and RoleBinding declared within the provider workspace will grant access to the request:
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
name: foo-creator
clusterName: provider
rules:
- apiGroups:
- foo.api
resources:
- foos
verbs:
- create
---
apiVersion: rbac.authorization.k8s.io/v1
kind: RoleBinding
metadata:
name: user-1-foo-creator
namespace: default
clusterName: provider
subjects:
- kind: User
name: apis.kcp.io:binding:user-1
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: Role
name: foo-creator
Note
The same authorization scheme is enforced when executing the request of a claimed resource via the virtual APIExport API server, i.e. a claimed resource is bound to the same maximal permission policy. Only the actual owner of that resources can go beyond that policy.
TBD: Example
Local Policy Authorizer¶
This authorizer ensures that RBAC rules contained within a workspace are being applied and work just like in a regular Kubernetes cluster.
It is possible to bind to Roles and ClusterRoles in the bootstrap policy from a local policy's
RoleBinding or ClusterRoleBinding, for example the system:kcp:workspace:access ClusterRole exists in the
system:admin logical cluster, but can still be bound from without any other logical cluster.
Global Policy Authorizer¶
This authorizer works identically to the Local Policy Authorizer, just with the difference that it uses a global (i.e. across shards) getter for Roles and RoleBindings.
Bootstrap Policy Authorizer¶
The bootstrap policy authorizer works just like the local authorizer but references RBAC rules
defined in the system:admin system workspace. This workspace is where the classic Kubernetes
RBAC like the cluster-admin ClusterRole is being defined and the policy defined in this workspace
applies to every workspace in a kcp shard.
Webhook Authorizer¶
This authorizer can be enabled by providing the --authorization-webhook-config-file flag to the kcp process
and works identically to how it works in vanilla Kubernetes.
The given configuration file must be of the kubeconfg format and point to an HTTPS server, potentially including certificate information as needed:
apiVersion: v1
kind: Config
clusters:
- name: webhook
cluster:
server: https://localhost:8080/
current-context: webhook
contexts:
- name: webhook
context:
cluster: webhook
The webhook will receive every authorization request made in kcp, including internal ones. This means if the webhook is badly configured, it can even prevent kcp from starting up successfully, but on the other hand this allows a lot of influence over the authorization in kcp.
The webhook will receive JSON-marshalled SubjectAccessReview objects, that (compared to vanilla Kubernetes) include the name of target logical cluster as an extra field, like so:
{
"apiVersion": "authorization.k8s.io/v1beta1",
"kind": "SubjectAccessReview",
"spec": {
"resourceAttributes": {
"namespace": "kittensandponies",
"verb": "get",
"group": "unicorn.example.org",
"resource": "pods"
},
"user": "jane",
"group": [
"group1",
"group2"
],
"extra": {
"authorization.kcp.io/cluster-name": ["root"]
}
}
}
The extra field will contain the logical cluster name (e.g. o43u2gh528rtfg721rg92), not the human-readable path. Webhooks need to resolve the name to a path themselves if necessary.
Note
In the past the cluster name was provided in the authorization.kubernetes.io/cluster-name extra field instead of authorization.kcp.io/cluster-name.
This field is deprecated as of kcp v0.28.3 and will be removed in a future release.
Note
Request payloads can also contain the authentication.kcp.io/cluster-name and authentication.kcp.io/scopes extra fields if the user originates from the workspace the request is made against.
If the users authenticated against another workspace than the target of the request these fields will not be present - instead the user will be seen as system:anonymous with groups system:authenticated and system:cluster:<logical-cluster>, where <logical-cluster> is the name of the logical cluster backing the workspace they authenticated against.
Authorizer Order¶
By default, the authorizers are evaluated in the following order:
- Always Allow Groups Authorizer (
AlwaysAllowGroups) - Always Allow Paths Authorizer (
AlwaysAllowPaths) - RBAC Chain (
RBAC) - Webhook Authorizer (
Webhook)
The order in which authorizers are evaluated can be configured. This allows administrators to customize the authorization flow.
It can be done by setting --authorization-order flag in the kcp server. This flag accepts a comma-separated list of authorizer names, which will be evaluated in the specified order.
Here is an example on how to enable the Webhook to be the first one in the authorizers chain:
Scopes¶
Scopes are a way to limit the access of a user to a specific logical cluster.
Scopes are (optionally) attached to the user identity by setting the
authentication.kcp.io/scopes: cluster:<logical-cluster>,... extra field. The
scope is then checked by the authorizers. For example:
logical-cluster-1,
falling back to be considered as user system:anonymous with groups
system:authenticated and system:cluster:logical-cluster-1 in all other logical clusters.
Each extra field can contain multiple scopes, separated by a comma:
user: user1
groups: ["group1"]
extra:
authentication.kcp.io/scopes:
- cluster:logical-cluster-1,cluster:logical-cluster-2
logical-cluster-1 and
logical-cluster-2, falling back to be considered as user
system:anonymous with groups system:authenticated,
system:cluster:logical-cluster-1 and
system:cluster:logical-cluster-2 in all other logical clusters.
If multiple authentication.kcp.io/scopes values are set, the intersection is
taken:
user: user1
groups: ["group1"]
extra:
authentication.kcp.io/scopes:
- cluster:logical-cluster-1,cluster:logical-cluster-2
- cluster:logical-cluster-2,cluster:logical-cluster-3
logical-cluster-2, falling
back to be considered as user system:anonymous with groups
system:authenticated and cluster:logical-cluster-2 in all other
logical clusters.
The intersection can be empty, in which case it falls back in every logical cluster.
When impersonating a user in a logical cluster, the resulting user identity is scoped to the logical cluster the impersonation is happening in.
A scope mismatch does not invalidate the warrants (see next section) of a user.
Warrants¶
Warrants are a way to grant extra access to a user. It can be limited by the scope of a logical cluster.
A warrant is attached by adding a authorization.kcp.io/warrant extra field to the user identity
with a JSON-encoded user info, and the limiting logical cluster set as authentication.kcp.io/scopes: cluster:<logical-cluster>.
in the embedded user info's extra. The warrant is then checked by the authorizers in the chain of every step
if the primary users is not allowed. For example:
user: user1
groups: ["group1"]
extra:
authorization.kcp.io/warrant: |
{
"user": "user2",
"groups": ["group2"],
"extra": {
"authentication.kcp.io/scopes": "cluster:logical-cluster-1"
}
}
This warrant allows user1 to act under the permissions of user2 in
logical-cluster-1 if user1 is not allowed to act as user2 in the first place.
Note that a warrant only allow to act under the permissions of the warrant user, but not to act as the warrant user itself. E.g. in auditing or admission control, the primary user is still the one that is acting.
Warrants can be nested, i.e. a warrant can contain another warrant.