A tool to use AlibabaCloud RAM credentials to authenticate to a Kubernetes cluster.
If you are an administrator running a Kubernetes cluster on AlibabaCloud, you already need to manage AlibabaCloud RAM credentials to provision and update the cluster. By using AlibabaCloud RAM Authenticator for Kubernetes, you avoid having to manage a separate credential for Kubernetes access.
If you are building a Kubernetes installer on AlibabaCloud, AlibabaCloud RAM Authenticator for Kubernetes can simplify your bootstrap process.
You won't need to somehow smuggle your initial admin credential securely out of your newly installed cluster.
Instead, you can create a dedicated KubernetesAdmin
role at cluster provisioning time and set up Authenticator to allow cluster administrator logins.
Assuming you have a cluster running in AlibabaCloud and you want to add AlibabaCloud RAM Authenticator for Kubernetes support, you need to:
- Create an RAM role you'll use to identify users.
- Run the Authenticator server as a DaemonSet.
- Configure your API server to talk to Authenticator.
- Set up kubectl to use Authenticator tokens.
First, you must create one or more RAM roles that will be mapped to users/groups inside your Kubernetes cluster. The easiest way to do this is to log into the RAM Console:
- Choose the "RAM Roles" / "Create RAM Role" option.
- Select type of trusted entity "Alibaba Cloud Account", Select Trusted Alibaba Cloud Account "Current Alibaba Cloud Account".
- Type in a name at "RAM Role Name" and click "OK"
This will create an RAM role with no permissions that can be assumed by authorized users/roles in your account. Note the AlibabaCloud Resource Name (ARN) of your role, which you will need below.
You can also skip this step and use:
- An existing role (such as a cross-account access role).
- An RAM user (see
mapUsers
below).
The server is meant to run on each of your master nodes as a DaemonSet with host networking so it can expose a localhost port.
For a sample ConfigMap and DaemonSet configuration, see example.yaml
.
If you're building an automated installer, you can also pre-generate the certificate, key, and webhook kubeconfig files easily using ack-ram-authenticator init
.
This command will generate files and place them in the configured output directories.
You can run this on each master node prior to starting the API server. You could also generate them before provisioning master nodes and install them in the appropriate host paths.
If you do not pre-generate files, ack-ram-authenticator server
will generate them on demand.
This works but requires that you restart your Kubernetes API server after installation.
The Kubernetes API integrates with ACK RAM Authenticator for Kubernetes using a token authentication webhook.
When you run ack-ram-authenticator server
, it will generate a webhook configuration file and save it onto the host filesystem.
You'll need to add a single additional flag to your API server configuration:
--authentication-token-webhook-config-file=/etc/kubernetes/ack-ram-authenticator/kubeconfig.yaml
On many clusters, the API server runs as a static pod.
You can add the flag to /etc/kubernetes/manifests/kube-apiserver.yaml
.
Make sure the host directory /etc/kubernetes/ack-ram-authenticator/
is mounted into your API server pod.
You may also need to restart the kubelet daemon on your master node to pick up the updated static pod definition:
systemctl restart kubelet.service
This requires a 1.10+
kubectl
binary to work. If you receivePlease enter Username:
when trying to usekubectl
you need to update to the latestkubectl
Finally, once the server is set up you'll want to authenticate!
You will still need a kubeconfig
that has the public data about your cluster (cluster CA certificate, endpoint address).
The users
section of your configuration, however, should include an exec section (refer to the v1.10 docs)::
# [...]
users:
- name: kubernetes-admin
user:
exec:
apiVersion: client.authentication.k8s.io/v1alpha1
command: ack-ram-authenticator
args:
- "token"
- "-i"
- "CLUSTER_ID"
- "-r"
- "ROLE_ARN"
# no client certificate/key needed here!
This means the kubeconfig
is entirely public data and can be shared across all Authenticator users.
It may make sense to upload it to a trusted public location such as AlibabaCloud OSS.
Make sure you have the ack-ram-authenticator
binary installed.
You can install it with go get -u -v github.com/AliyunContainerService/ack-ram-authenticator/cmd/ack-ram-authenticator
.
To authenticate, run kubectl --kubeconfig /path/to/kubeconfig" [...]
.
kubectl will exec
the ack-ram-authenticator
binary with the supplied params in your kubeconfig which will generate a token and pass it to the apiserver.
The token is valid for 15 minutes and can be reused multiple times.
You can also omit -r ROLE_ARN
to sign the token with your existing credentials without assuming a dedicated role.
This is useful if you want to authenticate as an RAM user directly.
It works using the RAM sts:GetCallerIdentity
API endpoint.
This endpoint returns information about whatever RAM credentials you use to connect to it.
We use this API in a somewhat unusual way by having the Authenticator client generate and pre-sign a request to the endpoint. We serialize that request into a token that can pass through the Kubernetes authentication system.
The token is passed through the Kubernetes API server and into the Authenticator server's /authenticate
endpoint via a webhook configuration.
The Authenticator server validates all the parameters of the pre-signed request to make sure nothing looks funny.
It then submits the request to the real https://sts.aliyuncs.com
server, which validates the client's HMAC signature and returns information about the user.
Now that the server knows the RAM identity of the client, it translates this identity into a Kubernetes user and groups via a simple static mapping.
The Authenticator cluster ID is a unique-per-cluster identifier that prevents certain replay attacks. Specifically, it prevents one Authenticator server (e.g., in a dev environment) from using a client's token to authenticate to another Authenticator server in another cluster.
The cluster ID does need to be unique per-cluster, but it doesn't need to be a secret. Some good choices are:
- A random ID such as from
openssl rand 16 -hex
- The domain name of your Kubernetes API server
Credentials can be specified for use with ack-ram-authenticator
via create file at ~/.acs/credentials, for example:
{
"AcsAccessKeyId": "xxxxxxx",
"AcsAccessKeySecret": "xxxxxxxxxxxxxxxx"
}
if you are using a STS Token, the ~/.acs/credentials file will be like:
{
"AcsAccessKeyId": "xxxxxx",
"AcsAccessKeySecret": "xxxxxx",
"AcsAccessSecurityToken": "xxxxxx"
}
This includes specifying RAM credentials by utilizing a credentials file.
To use ack-ram-authenticator as client, your kubeconfig would be like this:
apiVersion: v1
clusters:
- cluster:
server: ${server}
certificate-authority-data: ${cert}
name: kubernetes
contexts:
- context:
cluster: kubernetes
user: ack
name: ack
current-context: ack
kind: Config
preferences: {}
users:
- name: ack
user:
exec:
apiVersion: client.authentication.k8s.io/v1alpha1
command: ack-ram-authenticator
args:
- "token"
- "-i"
- "mycluster"
If that fails, there are a few possible problems to check for:
-
Make sure your base RAM credentials are available in your shell.
-
Make sure the target role allows your source account access (in the role trust policy).
-
Make sure your source principal (user/role/group) has an RAM policy that allows
sts:AssumeRole
for the target role. -
Make sure you don't have any explicit deny policies attached to your user, group that would prevent the
sts:AssumeRole
.
The client and server have the same configuration format. They can share the same exact configuration file, since there are no secrets stored in the configuration.
# a unique-per-cluster identifier to prevent replay attacks (see above)
clusterID: my-dev-cluster.example.com
# default RAM role to assume for `ack-ram-authenticator token`
defaultRole: acs:ram::000000000000:role/KubernetesAdmin
# server listener configuration
server:
# localhost port where the server will serve the /authenticate endpoint
port: 21362 # (default)
# state directory for generated TLS certificate and private keys
stateDir: /var/ack-ram-authenticator # (default)
# output `path` where a generated webhook kubeconfig will be stored.
generateKubeconfig: /etc/kubernetes/ack-ram-authenticator.kubeconfig # (default)
# each mapRoles entry maps an RAM role to a username and set of groups
# Each username and group can optionally contain template parameters:
# 1) "{{AccountID}}" is the 16 digit ID.
# 2) "{{SessionName}}" is the role session name.
mapRoles:
# statically map acs:ram::000000000000:role/KubernetesAdmin to cluster admin
- roleARN: acs:ram::000000000000:role/KubernetesAdmin
username: kubernetes-admin
groups:
- system:masters
# each mapUsers entry maps an RAM role to a static username and set of groups
mapUsers:
# map user RAM user Alice in 000000000000 to user "alice" in group "system:masters"
- userARN: acs:ram::000000000000:user/Alice
username: alice
groups:
- system:masters
You are welcome to make new issues and pull reuqests.