Turandot Tutorial: Self-Contained
Note that this tutorial is complex in that it supports CSAR packages that include container images as artifacts. You may prefer to start with the simpler tutorial.
Start by cloning the Turandot git repository so that you’ll have access to all the example files and helper scripts:
git clone https://github.com/tliron/turandot.git
cd turandot
If you prefer to read code rather than prose, check out the lab/
directory,
where we have scripts that do much of what is explained here.
Requirements
Tools
Download the binary release of Turandot.
Specifically you’ll need the turandot
CLI utility (it’s a self-contained executable).
You will also need the Reposure CLI utility in order to configure the registry that Turandot will use to store TOSCA service templates and artifacts, so download that, too.
Finally, we’ll need the puccini-csar
CLI utility from Puccini
in order to package our CSAR. It’s generally useful to have Puccini available in
order to validate and otherwise work with your TOSCA and CSAR.
A few other tools used by the scripts: podman
(or docker
), pigz
(or gzip
),
zip
, zipinfo
, tar
, jq
.
Kubernetes Cluster
To get started with Turandot you need kubectl
access to a working Kubernetes
cluster.
For development work, Minikube is a quick way to get started. Just make sure to start Minikube with its registry add-on enabled:
minikube start --addons=registry ...
The turandot
and reposure
utilities use the same local configuration you have for
kubectl
, and like kubectl
they can accept a --namespace
argument for selecting
the namespace in which to work. To make commands more succinct in this guide let’s set a
default namespace:
kubectl create namespace workspace
kubectl config set-context --current --namespace=workspace
Installing the Operators
Before you install the operators, note that you may need to specify a cluster-wide role in order to permit them access to the authentication and authorization secrets necessary to connect to the registry (see below), because those secrets may be stored outside of your namespace.
If you indeed need access to secrets for testing you can use “view” role, which already exists in most Kubernetes deployments. It allows read-access to any resource in any namespace.
(For production systems you would want more tightened security, in which case it would be a good idea to create a custom cluster role that would allow access to just the registry secrets and nothing else. Doing so is beyond the scope of this guide.)
To install the operators with the “view” cluster role you can use these commands:
turandot operator install --role=view --site=central --wait -v
reposure operator install --role=view --wait -v
Here we’re also giving this cluster the “central” site identifier. This will be used for multi-cluster policy-based TOSCA service composition.
Note the operators’ container images are downloaded from Docker Hub. Here are direct links for Turandot, Reposure operator, and Reposure surrogate.
The --wait
flag tells the commands to block until the operators are running
successfully. The -v
flag adds more verbosity so you can see what the command is
doing. (You can use -vv
for even more verbosity.)
Configuring the Registry
You will now use Reposure to configure the “default” registry for Turandot.
This can be simple or complex depending on your Kubernetes cluster. The reason it can be challenging is that the Turandot operator does more than just deploy TOSCA, it can also deploy artifacts referred to by your TOSCA, including artifacts of a special type: container images. Container images are downloaded from a registry by the container runtime (CRI-O, Docker, etc.) that runs on each of the cluster’s hosts, and by the container runtime is likely to be configured by delay to require TLS authentication (HTTPS) and may even require authorization.
Reposure comes with built-in support for the built-in registries of a few Kubernetes distributions, making it easy to make use of them. For Minikube:
reposure registry create default --provider=minikube --wait -v
For OpenShift:
reposure registry create default --provider=openshift --wait -v
(For both of these cases we would need to add --role=view
during the operator
installation step above.)
If you’re using neither Minikube nor OpenShift then you must set up your own registry. For production environments you’d likely want to use a robust product, like Harbor or Quay. Alternatively, Reposure comes with a “simple” registry that is suitable for testing and small deployments.
Reposure’s “Simple” Registry
Installing the “simple” registry is simple, but configuring your Kubernetes container runtime to accept it is beyond the scope of this guide. Specifically you would need to allow it to accept your TLS certificate or your custom certificate authority.
However, if you can configure your container runtime to at least accept self-signed
certificates (so-called “insecure” mode, which in Minikube is enabled via the
--insecure-registry
flag),
then Reposure’s “simple” registry can provision such a self-signed certificate for
you.
To do so Reposure relies on cert-manager, which does a lot of the heavy lifting required for provisioning and updating certificates. (The additional challenge of working with TLS certificates in cloud environments is that IP addresses change, so that certificates either have to be updated or tied to a DNS domain name, and then DNS management may be local and custom.)
So, assuming your container runtime is “insecure”, you can start by installing cert-manager:
kubectl apply --filename=https://github.com/cert-manager/cert-manager/releases/download/v1.8.0/cert-manager.yaml
And then install the “simple” registry with self-signed authentication like so:
reposure simple install --authentication --wait -v
Finally, configure the registry:
reposure registry create default --provider=simple --wait -v
(Note that if you are using the “simple” registry with authentication then you don’t
need to install the operators with --role=view
, because the “simple” registry stores
its certificate secrets within its namespace.)
Building the Self-Contained CSAR
Let’s use the included Self-Contained example, which is based on this Kubernetes demo.
First you’ll need to export the container image into a tarball so that you can use it as a TOSCA artifact. The build-csar script will handle that:
examples/self-contained/scripts/save-container-image
Next you’ll use the build-csar script to package the TOSCA topology template, profiles, and artifacts (including the container image tarball you created above) into a CSAR:
examples/self-contained/scripts/build-csar
The CSAR file should now sit in the “dist” directory.
Deploying “Self-Contained”
You can now deploy the CSAR to your cluster:
turandot service deploy self-contained --file=dist/self-contained.csar -v
Follow the logs to see what Turandot is doing:
turandot operator logs --follow
To list all deployed services:
turandot service list
Note that the “Self-Contained” example includes a LoadBalancer Kubernetes service that would
allow you to access its web page from outside the cluster. If your cluster is not configured
with LoadBalancer support then the service will never get an IP address, and the TOSCA “url”
output for your “Self-Contained” service will show http://0.0.0.0:80
, even when successfully
deployed.
If you’re using Minikube, it comes with a primitive ingress solution based on ssh tunneling that can be useful for testing. To run it (it’s blocking, so you might want to do it in a separate terminal session):
minikube tunnel
Once the tunnel is up, the LoadBalancer should get its IP address, and Turandot would soon update the “url” output with the correct URL. You can then use curl or a web browser to access it:
xdg-open $(turandot service output self-contained url)