Management Plane Installation
This page will show you how to install the Tetrate Service Bridge management plane in a production environment.
Before you start, make sure that you've:
✓ Checked the requirements
✓ Checked TSB management plane components
✓ Checked types of certificates and internal certificates requirements
✓ Checked firewall information
✓ If you are upgrading from previous version, also check PostgreSQL backup and restore
✓ Downloaded Tetrate Service Bridge CLI (tctl
)
✓ Synced the Tetrate Service Bridge images
Management Plane Operator
To keep installation simple but still allow a lot of custom configuration
options we have created a management plane operator. The operator will run in
the cluster and bootstraps the management plane as described in a
ManagementPlane Custom Resource. It watches for changes and enacts them. To help
in creating the right Custom Resource Document (CRD) we have added the ability
to our tctl
client to create the base manifests which you can then modify
according to your required set-up. After this you can either apply the manifests
directly to the appropriate clusters or use in your source control operated
clusters.
Operators
If you would like to know more about the inner workings of Operators, and the Operator Pattern, review the Kubernetes documentation
Create the manifest allowing you to install the management plane operator from your private Docker registry:
tctl install manifest management-plane-operator \
--registry <registry-location> > managementplaneoperator.yaml
- Standard
- OpenShift
The managementplaneoperator.yaml
file created by the install manifest command
can be applied directly to the appropriate cluster by using the kubectl client:
kubectl apply -f managementplaneoperator.yaml
After applying the manifest you will see the operator running in the tsb
namespace:
kubectl get pod -n tsb
RedHat Ecosystem Catalog
TSB is certified and listed on RedHat Ecosystem Catalog. TSB can be installed on Openshift platform by following the instructions below or via here.
The managementplaneoperator.yaml
file created by the install manifest command
can be applied directly to the appropriate cluster by using the oc
client:
oc apply -f managementplaneoperator.yaml
After applying the manifest you will see the operator running in the tsb
namespace:
oc get pod -n tsb
Example output:
NAME READY STATUS RESTARTS AGE
tsb-operator-management-plane-d4c86f5c8-b2zb5 1/1 Running 0 8s
Configuring Secrets
The management plane components need some secrets for both internal and external communication purposes. Following are a list of secrets that you need to create.
Secret name | Description |
---|---|
admin-credentials | TSB will create a default admin user with name: admin and this is the password's one way hash for this special account. These credentials are kept outside of your IdP while any other credentials must be stored in your IdP. |
tsb-certs | TLS certificate that has type kubernetes.io/tls . Must have tls.key and tls.cert value. The TLS certificates can be self signed or issued by public CA. |
postgres-credentials | Contains: 1. Postgres username and password. 2. The CA certificate to verify Postgres connections when Postgres is configured to present a self-signed certificate. TLS verification only happens if you set sslMode in Postgres settings to verify-ca or verify-full . See PostgresSettings for more details. 3. Client certificate and private key if Postgres is configured with mutual TLS. |
elastic-credentials | Elasticsearch username and password. |
es-certs | The CA certificate to validate Elasticsearch connections when Elasticsearch is configured to present a self-signed certificate. |
ldap-credentials | Only set if using LDAP as Identity Provider (IdP). Contain LDAP binddn and bindpassword . |
custom-host-ca | Only set if using LDAP as IdP. The CA certificate to validate LDAP connections when LDAP is configured to present a self-signed certificate. |
iam-oidc-client-secret | Only set if using OIDC with any IdP. Contain OIDC client-secret and device-client-secret. |
azure-credentials | Only set if using OIDC with Azure AD as IdP. Client secret to connect to Azure AD for team and user synchronization. |
xcp-central-cert | XCP central mTLS or TLS certificate. The requirement of this secret is depend on whether you use mTLS or JWT with TLS for XCP authentication. Go to Internal certificate requirements for more details. |
mpc-certs | Only set if using mTLS for XCP authentication. Go to Internal certificate requirements for more details. |
Using tctl to Generate Secrets
These secrets can be generated in the correct format by passing them as command-line flags to the tctl
management-plane-secrets command.
XCP central certificate with cert-manager
In case you have installed cert-manager
in the management plane cluster, you
can have tctl
automatically provision certificates for secure communication with
control planes. To do this, add an --xcp-certs
flag to the install manifest management-plane-secrets
command. This will create a certificate Issuer along with xcp-central-cert and mpc-certs certificates.
- OIDC as IdP
- LDAP as IdP
- OIDC as IdP with --xcp-certs flag
The following command will generate managementplane-secrets.yaml
that contains Elasticsearch, Postgres, OIDC and admin credentials along with TSB TLS certificate.
tctl install manifest management-plane-secrets \
--elastic-password <elastic-password> \
--elastic-username <elastic-username> \
--oidc-client-secret "<oidc-client-secret>" \
--postgres-password <postgres-password> \
--postgres-username <postgres-username> \
--tsb-admin-password <tsb-admin-password> \
--tsb-server-certificate "$(cat foo.cert)" \
--tsb-server-key "$(cat foo.key)" > managementplane-secrets.yaml
The following command will generate managementplane-secrets.yaml
that contains Elasticsearch, Postgres, LDAP and admin credentials along with TSB TLS certificate.
tctl install manifest management-plane-secrets \
--elastic-password <elastic-password> \
--elastic-username <elastic-username> \
--ldap-bind-dn <ldap-bind-dn> \
--ldap-bind-password <ldap-bind-password> \
--postgres-password <postgres-password> \
--postgres-username <postgres-username> \
--tsb-admin-password <tsb-admin-password> \
--tsb-server-certificate "$(cat foo.cert)" \
--tsb-server-key "$(cat foo.key)" > managementplane-secrets.yaml
The following command will use cert-manager to generate managementplane-secrets.yaml
that contains XCP central and MPC certificate for XCP secure communication.
tctl install manifest management-plane-secrets \
--elastic-password <elastic-password> \
--elastic-username <elastic-username> \
--oidc-client-secret "<oidc-client-secret>" \
--postgres-password <postgres-password> \
--postgres-username <postgres-username> \
--tsb-admin-password <tsb-admin-password> \
--tsb-server-certificate "$(cat foo.cert)" \
--tsb-server-key "$(cat foo.key)" \
--xcp-certs > managementplane-secrets.yaml
XCP central DNS name
With --xcp-certs
flag, tctl
automatically install required issuer and certificate, XCP central cert will have central.xcp.tetrate.io
as the DNS name.
See the CLI reference
documentation for all available options such as providing CA certificates for
Elasticsearch
, PostgreSQL
and LDAP
. You can also check the bundled explanation from tctl
by running this help command:
tctl install manifest management-plane-secrets --help
Applying secrets
Once you've created your secrets manifest, you can add to source control or apply it to your cluster.
Vault Injection
If you're using Vault
injection for certain components, remove the applicable
secrets from the manifest that you've created before applying it to your
cluster.
- Standard
- OpenShift
kubectl apply -f managementplane-secrets.yaml
oc apply -f managementplane-secrets.yaml
Management Plane Installation
Now you're ready to deploy the management plane.
To deploy the management plane you need to create a ManagementPlane
custom
resource in the Kubernetes cluster that describes the management plane.
Organization name
Organization is a root of the TSB object hierarchy. A TSB Management plane can only have one organization.
To login with tctl
, you will need to specify organization name and it must match with <organization-name>
that you set in the ManagementPlane
CR below. Organization name has to be lowercase to comply with RFC standards.
If not specified, the default value is tetrate
and it cannot be changed after creation.
Below is a ManagementPlane
custom resource (CR) that describes a basic management plane. Save this
managementplane.yaml
and adjust it according to your needs:
XCP authentication
Starting TSB 1.5, default authentication method for XCP central and edge traffic is JWT.
If you want to keep using mTLS you have to set spec.components.xcp.centralAuthModes.mutualTls
to true
.
To migrate from mTLS to JWT follow steps described in here
About ES selfsigned certs
If you are using selfsigned certs for ES, you may encounter an issue with oap pod and oap-deployment not able to start.
To solve this, provide a certificate signed by a well known Certificate Authority and set selfSigned: false
in the ManagementPlane CR.
If your organization acts a signer, then you will need to give the pods privileged access in the container security context.
- OIDC as IdP
- LDAP as IdP
- Using mTLS for XCP Authentication
The following example uses OIDC as identity provider.
apiVersion: install.tetrate.io/v1alpha1
kind: ManagementPlane
metadata:
name: managementplane
namespace: tsb
spec:
hub: <registry-location>
organization: <organization-name>
dataStore:
postgres:
host: <postgres-hostname-or-ip>
port: <postgres-port>
name: <database-name>
telemetryStore:
elastic:
host: <elastic-hostname-or-ip>
port: <elastic-port>
version: <elastic-version>
selfSigned: <is-elastic-use-self-signed-certificate>
protocol: <http or https. default to https if not set>
identityProvider:
oidc:
clientId: <oidc-client-id>
# authorization code flow for TSB UI login
providerConfig:
dynamic:
configurationUri: <oidc-well-known-openid-configuration>
redirectUri: <oidc-callback>
scopes:
- email
- profile
- offline_access
# Customize token issuer to use signing key that you have created before.
tokenIssuer:
jwt:
expiration: 1h
issuers:
- name: https://jwt.tetrate.io
# Signing key stored in Kubernetes secret.
signingKey: private.key
# Kubernetes secret name where you store IAM signing key.
signingKeysSecret: iam-signing-key
components:
internalCertProvider:
certManager:
managed: INTERNAL
If you are not using Azure AD an the OIDC Identity provider, follow the steps in Users Synchronization to see how you can create organizations and sync your users and teams into TSB
The following uses LDAP as the identity provider
apiVersion: install.tetrate.io/v1alpha1
kind: ManagementPlane
metadata:
name: managementplane
namespace: tsb
spec:
hub: <registry-location>
organization: <organization-name>
dataStore:
postgres:
host: <postgres-hostname-or-ip>
port: <postgres-port>
name: <database-name>
telemetryStore:
elastic:
host: <elastic-hostname-or-ip>
port: <elastic-port>
version: <elastic-version>
selfSigned: <is-elastic-use-self-signed-certificate>
protocol: <http or https. default to https if not set>
identityProvider:
ldap:
host: <ldap-hostname-or-ip>
port: <ldap-port>
search:
baseDN: dc=tetrate,dc=io
iam:
matchDN: "cn=%s,ou=People,dc=tetrate,dc=io"
matchFilter: "(&(objectClass=person)(uid=%s))"
sync:
usersFilter: "(objectClass=person)"
groupsFilter: "(objectClass=groupOfUniqueNames)"
membershipAttribute: uniqueMember
# Customize token issuer to use signing key that you have created before.
tokenIssuer:
jwt:
expiration: 1h
issuers:
- name: https://jwt.tetrate.io
# Signing key stored in Kubernetes secret.
signingKey: private.key
# Kubernetes secret name where you store IAM signing key.
signingKeysSecret: iam-signing-key
components:
internalCertProvider:
certManager:
managed: INTERNAL
The following uses LDAP as the identity provider and set XCP authentication to use mTLS.
apiVersion: install.tetrate.io/v1alpha1
kind: ManagementPlane
metadata:
name: managementplane
namespace: tsb
spec:
hub: <registry-location>
organization: <organization-name>
dataStore:
postgres:
host: <postgres-hostname-or-ip>
port: <postgres-port>
name: <database-name>
telemetryStore:
elastic:
host: <elastic-hostname-or-ip>
port: <elastic-port>
version: <elastic-version>
selfSigned: <is-elastic-use-self-signed-certificate>
protocol: <http or https. default to https if not set>
identityProvider:
ldap:
host: <ldap-hostname-or-ip>
port: <ldap-port>
search:
baseDN: dc=tetrate,dc=io
iam:
matchDN: "cn=%s,ou=People,dc=tetrate,dc=io"
matchFilter: "(&(objectClass=person)(uid=%s))"
sync:
usersFilter: "(objectClass=person)"
groupsFilter: "(objectClass=groupOfUniqueNames)"
membershipAttribute: uniqueMember
# Customize token issuer to use signing key that you have created before.
tokenIssuer:
jwt:
expiration: 1h
issuers:
- name: https://jwt.tetrate.io
# Signing key stored in Kubernetes secret.
signingKey: private.key
# Kubernetes secret name where you store IAM signing key.
signingKeysSecret: iam-signing-key
components:
internalCertProvider:
certManager:
managed: INTERNAL
xcp:
centralAuthModes:
# use mTLS for XCP authentication
mutualTls: true
For more information on what each of these sections describes and how to configure them, please check out the following links:
Edit the relevant sections, save your configured custom resource to a file and apply it to your Kubernetes cluster.
- Standard
- OpenShift
kubectl apply -f managementplane.yaml
Before applying it, bear in mind that you will have to allow the service accounts of the different management plane components to your OpenShift Authorization Policies.
oc adm policy add-scc-to-user anyuid -n tsb -z tsb-iam
oc adm policy add-scc-to-user anyuid -n tsb -z tsb-oap
Now you can apply it:
oc apply -f managementplane.yaml
Note: TSB will automatically do this every hour, so this command only needs to be run once after the initial installation.
Verifying Installation
To verify your installation succeeded, log in as the admin user. Try to connect
to the TSB UI or login with the tctl
CLI tool.
The TSB UI is reachable on port 8443 of the external IP as returned by the following command:
- Standard
- OpenShift
kubectl get svc -n tsb envoy
oc get svc -n tsb envoy
To configure tctl
's default config profile to point to your new TSB cluster do
the following:
- Standard
- AWS
tctl config clusters set default --bridge-address $(kubectl get svc -n tsb envoy --output jsonpath='{.status.loadBalancer.ingress[0].ip}'):8443
tctl config clusters set default --bridge-address $(kubectl get svc -n tsb envoy --output jsonpath='{.status.loadBalancer.ingress[0].hostname}'):8443
Now you can log in with tctl
and provide the organization name and admin account credentials.
The tenant field is optional and can be left blank at this point and configured later,
when tenants are added to the platform.
tctl login
Organization: tetrate
Tenant:
Username: admin
Password: *****
Login Successful!
Go to Connect to TSB with tctl for more details on how to configure tctl.