Skip to content

Latest commit

 

History

History
1935 lines (1632 loc) · 73.7 KB

workshop-definition.md

File metadata and controls

1935 lines (1632 loc) · 73.7 KB

Workshop resource

The Workshop custom resource defines a workshop.

Workshop title and description

Each workshop is required to provide the title and description fields. If the fields are not supplied, the Workshop resource is rejected when you attempt to load it into the Kubernetes cluster.

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-markdown-sample
spec:
  title: Markdown Sample
  description: A sample workshop using Markdown
  content:
    files: github.com/eduk8s/lab-markdown-sample

The title field has a single-line value giving the subject of the workshop.

The description field has a longer description of the workshop.

The following optional information can also be supplied for the workshop:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-markdown-sample
spec:
  title: Markdown Sample
  description: A sample workshop using Markdown
  url: YOUR-GITHUB-URL-FOR-LAB-MARKDOWN-SAMPLE
  difficulty: beginner
  duration: 15m
  vendor: learningcenter.tanzu.vmware.com
  authors:
  - John Smith
  tags:
  - template
  logo: data:image/png;base64,....
  content:
    files: YOUR-GITHUB-URL-FOR-LAB-MARKDOWN-SAMPLE

Where:

  • The url field is the Git repository URL for lab-markdown-sample. For example, https://github.com/eduk8s/lab-markdown-sample. It must be a URL you can go to for more information about the workshop.

  • The difficulty field must indicate who the workshop is targeting. The value must be beginner, intermediate, advanced, or extreme.

  • The duration field gives the expected maximum amount of time the workshop takes to complete. This field only provides informational value and does not police how long a workshop instance lasts. The field format is an integer number with s, m, or h suffix.

  • The vendor field must be a value that identifies the company or organization which the authors are affiliated with. This can be a company or organization name or a DNS hostname under the control of whoever has created the workshop.

  • The authors field must list the people who worked on creating the workshop.

  • The tags field must list labels that help to identify what the workshop is about. This is used in a searchable catalog of workshops.

  • The logo field must be a graphical image provided in embedded data URI format that depicts the workshop topic. The image should be 400 by 400 pixels. This is used in a searchable catalog of workshops.

  • The files field is the Git repository URL for lab-markdown-sample. For example, https://github.com/eduk8s/lab-markdown-sample.

When referring to a workshop definition after it is loaded into a Kubernetes cluster, the value of the name field given in the metadata is used. To experiment with slightly different workshop variations, copy the original workshop definition YAML file and change the value of name. Then make your changes and load them into the Kubernetes cluster.

Downloading workshop content

You can download workshop content when the workshop instance is created. If the amount of content is not too large, the download doesn't increase startup times for the workshop instance. The alternative is to bundle the workshop content in a container image built from the Learning Center workshop base image.

To download workshop content when the workshop instance is started, set the content.files field to the location of the workshop content:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-markdown-sample
spec:
  title: Markdown Sample
  description: A sample workshop using Markdown
  content:
    files: github.com/eduk8s/lab-markdown-sample

The location can be a GitHub or GitLab repository reference, a URL to a tarball hosted on an HTTP server, or a reference to an OCI image artifact on a registry.

In the case of a GitHub or GitLab repository, do not prefix the location with https:// as a symbolic reference is being used and not an actual URL.

The reference format to a GitHub or GitLab repository is similar to that used with kustomize when referencing remote repositories. For example:

  • github.com/organisation/project - Use the workshop content hosted at the root of the GitHub repository. The master or main branch is used.
  • github.com/organisation/project/subdir?ref=develop - Use the workshop content hosted at subdir of the GitHub repository. The develop branch is used.
  • gitlab.com/organisation/project - Use the workshop content hosted at the root of the GitLab repository. The master branch is used.
  • gitlab.com/organisation/project/subdir?ref=develop - Use the workshop content hosted at subdir of the GitLab repository. The develop branch is used.

In the case of a URL to a tarball hosted on a HTTP server, the URL can be in the following formats:

  • https://example.com/workshop.tar - Use the workshop content from the top-level directory of the unpacked tarball.
  • https://example.com/workshop.tar.gz - Use the workshop content from the top-level directory of the unpacked tarball.
  • https://example.com/workshop.tar?path=subdir - Use the workshop content from the specified subdirectory path of the unpacked tarball.
  • https://example.com/workshop.tar.gz?path=subdir - Use the workshop content from the specified subdirectory path of the unpacked tarball.

The tarball referenced by the URL can be uncompressed or compressed.

If using GitHub, instead of using the earlier form for referencing the Git repository containing the workshop content, you can instead use a URL to refer directly to the downloadable tarball for a specific version of the Git repository.

  • https://github.com/organization/project/archive/develop.tar.gz?path=project-develop

When using this form, you must reference the .tar.gz download and cannot use the .zip file. The basename of the tarball file is the branch or commit name. You must specify the path query string parameter where the argument is the project name and branch or commit. The path must be supplied as the contents of the repository is not returned at the root of the archive.

Using GitLab also provides a means of downloading a package as a tarball.

  • https://gitlab.com/organization/project/-/archive/develop/project-develop.tar.gz?path=project-develop

If the GitHub or GitLab repository is private, you can generate a personal access token providing read-only access to the repository and include the credentials in the URL.

  • https://username@token:github.com/organization/project/archive/develop.tar.gz?path=project-develop

As with this method, a full URL is being supplied to request a tarball of the repository and does not refer to the repository itself. You can also reference private enterprise versions of GitHub or GitLab and the repository do not need to be on the public github.com or gitlab.com sites.

The last case references an OCI image artifact stored on a registry. This is not a full container image with the operating system but an image containing only the files making up the workshop content. The URI formats for this are:

  • imgpkg+https://harbor.example.com/organisation/project:version - Use the workshop content from the top-level directory of the unpacked OCI artifact. The registry, in this case, must support https.
  • imgpkg+https://harbor.example.com/organisation/project:version?path=subdir - Use the workshop content from the specified subdirectory path of the unpacked OCI artifact. The registry, in this case, must support https.
  • imgpkg+http://harbor.example.com/organisation/project:version - Use the workshop content from the top-level directory of the unpacked OCI artifact. The registry, in this case, can support only http.
  • imgpkg+http://harbor.example.com/organisation/project:version?path=subdir - Use the workshop content from the specified subdirectory path of the unpacked OCI artifact. The registry, in this case, can support only http.

Instead of the prefix imgpkg+https://, you can use imgpkg://. The registry in this the case must still support https.

For any of the formats, you can supply credentials as part of the URI.

Access to the registry using a secure connection using https must have a valid certificate.

You can create the OCI image artifact using imgpkg from the Carvel toolset. For example, from the top-level directory of the Git repository containing the workshop content, you would run:

imgpkg push -i harbor.example.com/organisation/project:version -f .

In all cases for downloading workshop content, the workshop subdirectory holding the actual workshop content is relocated to /opt/workshop so that it is not visible to a user. If you want other files ignored and not included in what the user can see, you can supply a .eduk8signore file in your repository or tarball and list patterns for the files in it.

Note that the contents of the .eduk8signore file are processed as a list of patterns, and each is applied recursively to subdirectories. To ensure that a file is only ignored if it resides in the root directory, you must prefix it with ./.

./.dockerignore
./.gitignore
./Dockerfile
./LICENSE
./README.md
./kustomization.yaml
./resources

Container image for the workshop

When workshop content is bundled into a container image, the content.image field should specify the image reference identifying the location of the container image to be deployed for the workshop instance:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-markdown-sample
spec:
  title: Markdown Sample
  description: A sample workshop using Markdown
  content:
    image: quay.io/eduk8s/lab-markdown-sample:master

Even though you can download workshop content when the workshop environment is started, you might still want to override the workshop image used as a base. You can do this when you have a custom workshop base image that includes added language runtimes or tools required by specialized workshops.

For example, if running a Java workshop, you can specify the jdk11-environment workshop image, with workshop content still pulled down from GitHub:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-spring-testing
spec:
  title: Spring Testing
  description: Playground for testing Spring development
  content:
    image: dev.registry.tanzu.vmware.com/learning-center/jdk11-environment:latest
    files: github.com/eduk8s-tests/lab-spring-testing

If you want to use the latest version of an image, always include the :latest tag. This is important because the Learning Center Operator looks for version tags:main, :master, :developand:latest. When used, the Operator sets the image pull policy to Alwaysto ensure that a newer version is always pulled if available. Otherwise, the image is cached on the Kubernetes nodes and only pulled when it is initially absent. Any other version tags are always assumed to be unique and are never updated. Be aware of image registries that use a content delivery network (CDN) as front end. When using these image tags, the CDN can still regard them as unique and not do pull through requests to update an image even if it uses a tag of:latest`.

Where special custom workshop base images are available as part of the Learning Center project, you can specify a short name instead of specifying the full location for the image, including the image registry. The Learning Center Operator then fills in the rest of the details:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-spring-testing
spec:
  title: Spring Testing
  description: Playground for testing Spring development
  content:
    image: jdk11-environment:latest
    files: github.com/eduk8s-tests/lab-spring-testing

The short versions of the names which are recognized are:

  • base-environment:* - A tagged version of the base-environment workshop image matched with the current version of the Learning Center Operator.
  • jdk8-environment:* - A tagged version of the jdk8-environment workshop image matched with the current version of the Learning Center Operator.
  • jdk11-environment:* - A tagged version of the jdk11-environment workshop image matched with the current version of the Learning Center Operator.
  • conda-environment:* - A tagged version of the conda-environment workshop image matched with the current version of the Learning Center Operator.

The * variants of the short names map to the most up-to-date version of the image available when the version of the Learning Center Operator was released. That version is guaranteed to work with that version of the Learning Center Operator, whereas the latest version can be newer, with possible incompatibilities.

If required, you can remap the short names in the SystemProfile configuration of the Learning Center Operator. You can map additional short names to your custom workshop base images for use in your deployment of the Learning Center Operator and with any workshop of your own.

Setting environment variables

To set or override environment variables for the workshop instance, you can supply the session.env field:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-markdown-sample
spec:
  title: Markdown Sample
  description: A sample workshop using Markdown
  content:
    files: github.com/eduk8s/lab-markdown-sample
  session:
    env:
    - name: REPOSITORY-URL
      value: YOUR-GITHUB-URL-FOR-LAB-MARKDOWN-SAMPLE

Where:

  • The session.env field is a list of dictionaries with name and value fields.
  • The value field is the Git repository URL for lab-markdown-sample. For example, https://github.com/eduk8s/lab-markdown-sample.

Values of fields in the list of resource objects can reference several predefined parameters. The available parameters are:

  • session_ id - A unique ID for the workshop instance within the workshop environment.
  • session_ namespace - The namespace created for and bound to the workshop instance. This is the namespace unique to the session and where a workshop can create its own resources.
  • environment_ name - The name of the workshop environment. This is the same as the name of the namespace for the workshop environment. Do not rely on their being the same, and use the most appropriate to cope with any potential change.
  • workshop_ namespace - The namespace for the workshop environment. This is the namespace where all deployments of the workshop instances are created and where the service account that the workshop instance runs as exists.
  • service_ account - The name of the service account the workshop instance runs as and which has access to the namespace created for that workshop instance.
  • ingress_ domain - The host domain under which you can create hostnames when creating ingress routes.
  • ingress_ protocol - The protocol (http/https) that is used for ingress routes created for workshops.

The syntax for referencing one of the parameters is $(parameter_name).

The ability to override environment variables using this field must be limited to cases where they are required for the workshop. To set or override a specific workshop environment, set environment variables in the WorkshopEnvironment custom resource for the workshop environment instead.

Overriding the memory available

By default the container the workshop environment runs in is allocated 512Mi. If the editor is enabled, a total of 1Gi is allocated.

Where the purpose of the workshop is mainly aimed at deploying workloads into the Kubernetes cluster, this generally is sufficient. If you are running workloads in the workshop environment container and need more memory, you can override the default by setting memory under session.resources:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-markdown-sample
spec:
  title: Markdown Sample
  description: A sample workshop using Markdown
  content:
    image: quay.io/eduk8s/lab-markdown-sample:master
  session:
    resources:
      memory: 2Gi

Mounting a persistent volume

In circumstances where a workshop needs persistent storage to ensure no loss of work, you can request a persistent volume be mounted into the workshop container if the workshop environment container was killed and restarted:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-markdown-sample
spec:
  title: Markdown Sample
  description: A sample workshop using Markdown
  content:
    image: quay.io/eduk8s/lab-markdown-sample:master
  session:
    resources:
      storage: 5Gi

The persistent volume is mounted on the /home/eduk8s directory. Because this hides any workshop content bundled with the image, an init container is automatically configured and run, which copies the contents of the home directory to the persistent volume before the persistent volume is then mounted on top of the home directory.

Resource budget for namespaces

In conjunction with each workshop instance, a namespace is created for use during the workshop. That is, from the terminal of the workshop, you can deploy dashboard applications into the namespace through the Kubernetes REST API by using tools such as kubectl.

By default, this namespace has whatever limit ranges and resource quotas the Kubernetes cluster can enforce. In most cases, this means there are no limits or quotas.

To control how much resources can be used where no limit ranges and resource quotas are set, or to override any default limit ranges and resource quotas, you can set a resource budget for any namespace created for the workshop instance.

To set the resource budget, set the session.namespaces.budget field.

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-markdown-sample
spec:
  title: Markdown Sample
  description: A sample workshop using Markdown
  content:
    image: quay.io/eduk8s/lab-markdown-sample:master
  session:
    namespaces:
      budget: small

The resource budget sizings and quotas for CPU and memory are:

| Budget    | CPU   | Memory |
|-----------|-------|--------|
| small     | 1000m | 1Gi    |
| medium    | 2000m | 2Gi    |
| large     | 4000m | 4Gi    |
| x-large   | 8000m | 8Gi    |
| xx-large  | 8000m | 12Gi   |
| xxx-large | 8000m | 16Gi   |

A value of 1000m is equivalent to 1 CPU.

Separate resource quotas for CPU and memory are applied for terminating and non-terminating workloads.

Only the CPU and memory quotas are listed in the preceding table, but limits are also in place on the number of resource objects of certain types you can create, including persistent volume claims, replication controllers, services, and secrets.

For each budget type, a limit range is created with fixed defaults. The limit ranges for CPU use on a container are as follows.

Budget Minimum Maximum Request Limit
small 50m 1000m 50m 250m
medium 50m 2000m 50m 500m
large 50m 4000m 50m 500m
x-large 50m 8000m 50m 500m
xx-large 50m 8000m 50m 500m
xxx-large 50m 8000m 50m 500m

Those for memory are:

Budget Minimum Maximum Request Limit
small 32Mi 1Gi 128Mi 256Mi
medium 32Mi 2Gi 128Mi 512Mi
large 32Mi 4Gi 128Mi 1Gi
x-large 32Mi 8Gi 128Mi 2Gi
xx-large 32Mi 12Gi 128Mi 2Gi
xxx-large 32Mi 16Gi 128Mi 2Gi

The request and limit values are the defaults applied to a container when no resources specification is given in a Pod specification.

If a budget sizing for CPU and memory is sufficient, but to override the limit ranges and defaults for request and limit values when none is given in a Pod specification, you can supply overrides in session.namespaces.limits:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-markdown-sample
spec:
  title: Markdown Sample
  description: A sample workshop using Markdown
  content:
    image: quay.io/eduk8s/lab-markdown-sample:master
  session:
    namespaces:
      budget: medium
      limits:
        min:
          cpu: 50m
          memory: 32Mi
        max:
          cpu: 1
          memory: 1Gi
        defaultRequest:
          cpu: 50m
          memory: 128Mi
        default:
          cpu: 500m
          memory: 1Gi

Although all possible properties that can be set are listed in this example, you only need to supply the property for the value you want to override.

If you need more control over limit ranges and resource quotas, you can set the resource budget to custom. This removes any default limit ranges and resource quotas that might be applied to the namespace. You can then specify your own LimitRange and ResourceQuota resources as part of the list of resources created for each session.

Before disabling the quota and limit ranges or contemplating any switch to using a custom set of LimitRange and ResourceQuota resources, consider if that is what is really required.

The default requests defined by these for memory and CPU are fallbacks only. In most cases, instead of changing the defaults, you can specify memory and CPU resources in the Pod template specification of your deployment resources used in the workshop to indicate what the application requires. This allows you to control exactly what the application can use and so fit into the minimum quota required for the task.

This budget setting and the memory values are distinct from the amount of memory the container the workshop environment runs in. To change how much memory is available to the workshop container, set the memory setting under session.resources.

Patching workshop deployment

In order to set or override environment variables, you can provide session.env. To make other changes to the Pod template for the deployment used to create the workshop instance, provide an overlay patch. Such a patch can be used to override the default CPU and memory limit applied to the workshop instance or to mount a volume.

The patches are provided by setting session.patches. The patch is applied to the spec field of the pod template:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-resource-testing
spec:
  title: Resource testing
  description: Play area for testing memory resources
  content:
    files: github.com/eduk8s-tests/lab-resource-testing
  session:
    patches:
      containers:
      - name: workshop
        resources:
          requests:
            memory: "1Gi"
          limits:
            memory: "1Gi"

In this example, the default memory limit of "512Mi" is increased to "1Gi". Although memory is set using a patch in this example, the session.resources.memory field is the preferred way to override the memory allocated to the container the workshop environment is running in.

The patch, when applied, works differently than overlay patches found elsewhere in Kubernetes. Specifically, when patching an array containing a list of objects, a search is performed on the destination array. If an object already exists with the same value for the name field, the item in the source array is overlaid on top of the existing item in the destination array.

If there is no matching item in the destination array, the item in the source array is added to the end of the destination array.

This means an array doesn't outright replace an existing array, but a more intelligent merge of elements in the array is performed.

Creation of session resources

When a workshop instance is created, the workshop dashboard deployment is created in the namespace for the workshop environment. When more than one workshop instance is created under that workshop environment, all those deployments are in the same namespace.

For each workshop instance, a separate empty namespace is created with name corresponding to the workshop session. The workshop instance is configured so that the workshop instance service account can access and create resources in the namespace created for that instance. Each separate workshop instance has its corresponding namespace and cannot see the namespace for another instance.

To pre-create additional resources within the namespace for a workshop instance, you can supply a list of the resources against the session.objects field within the workshop definition. You might use this to add additional custom roles to the service account for the workshop instance when working in that namespace or to deploy a distinct instance of an application for just that workshop instance, such as a private image registry:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-registry-testing
spec:
  title: Registry Testing
  description: Play area for testing image registry
  content:
    files: github.com/eduk8s-tests/lab-registry-testing
  session:
    objects:
    - apiVersion: apps/v1
      kind: Deployment
      metadata:
        name: registry
      spec:
        replicas: 1
        selector:
          matchLabels:
            deployment: registry
        strategy:
          type: Recreate
        template:
          metadata:
            labels:
              deployment: registry
          spec:
            containers:
            - name: registry
              image: registry.hub.docker.com/library/registry:2.6.1
              imagePullPolicy: IfNotPresent
              ports:
              - containerPort: 5000
                protocol: TCP
              env:
              - name: REGISTRY_STORAGE_DELETE_ENABLED
                value: "true"
    - apiVersion: v1
      kind: Service
      metadata:
        name: registry
      spec:
        type: ClusterIP
        ports:
        - port: 80
          targetPort: 5000
        selector:
          deployment: registry

For namespaced resources, it is not necessary to specify the namespace field of the resource metadata. When the namespace field is not present, the resource is automatically created within the session namespace for that workshop instance.

When resources are created, owner references are added, making the WorkshopSession custom resource corresponding to the workshop instance the owner. When the workshop instance is deleted, any resources are automatically deleted.

Values of fields in the list of resource objects can reference several predefined parameters. The available parameters are:

  • session_ id - A unique ID for the workshop instance within the workshop environment.
  • session_ namespace - The namespace created for and bound to the workshop instance. This is the namespace unique to the session and where a workshop can create its own resources.
  • environment_ name - The name of the workshop environment. This is the same as the namespace's name for the workshop environment. Do not rely on their being the same, and use the most appropriate to cope with any potential change.
  • workshop_ namespace - The namespace for the workshop environment. This is the namespace where all deployments of the workshop instances are created and where the service account that the workshop instance runs as exists.
  • service_ account - The name of the service account the workshop instance runs as and which has access to the namespace created for that workshop instance.
  • ingress_ domain - The host domain under which you can create hostnames when creating ingress routes.
  • ingress_ protocol - The protocol (http/https) used for ingress routes created for workshops.

The syntax for referencing one of the parameters is $(parameter_name).

In the case of cluster-scoped resources, it is important that you set the name of the created resource so that it embeds the value of $(session_namespace). This way, the resource name is unique to the workshop instance, and you do not get a clash with a resource for a different workshop instance.

For examples of using the available parameters, see the following sections.

Overriding default role-based access control (RBAC) rules

By default, the service account created for the workshop instance has admin role access to the session namespace created for that workshop instance. This enables the service account to deploy applications to the session namespace and manage secrets and service accounts.

Where a workshop doesn't require admin access for the namespace, you can reduce the level of access it has to edit or view by setting the session.namespaces.role field:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-role-testing
spec:
  title: Role Testing
  description: Play area for testing roles
  content:
    files: github.com/eduk8s-tests/lab-role-testing
  session:
    namespaces:
      role: view

To add additional roles to the service account, such as working with custom resource types added to the cluster, you can add the appropriate Role and RoleBinding definitions to the session.objects field described previously:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-kpack-testing
spec:
  title: Kpack Testing
  description: Play area for testing kpack
  content:
    files: github.com/eduk8s-tests/lab-kpack-testing
  session:
    objects:
    - apiVersion: rbac.authorization.k8s.io/v1
      kind: Role
      metadata:
        name: kpack-user
      rules:
      - apiGroups:
        - build.pivotal.io
        resources:
        - builds
        - builders
        - images
        - sourceresolvers
        verbs:
        - get
        - list
        - watch
        - create
        - delete
        - patch
        - update
    - apiVersion: rbac.authorization.k8s.io/v1
      kind: RoleBinding
      metadata:
        name: kpack-user
      roleRef:
        apiGroup: rbac.authorization.k8s.io
        kind: Role
        name: kpack-user
      subjects:
      - kind: ServiceAccount
        namespace: $(workshop_namespace)
        name: $(service_account)

The subject of a RoleBinding must specify the service account name and namespace where it is contained, both of which are unknown in advance. References to parameters for the workshop namespace and service account for the workshop instance are used when defining the subject.

You can add additional resources by means of session.objects to grant cluster-level roles, which is necessary to grant the service account cluster-admin role:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-admin-testing
spec:
  title: Admin Testing
  description: Play area for testing cluster admin
  content:
    files: github.com/eduk8s-tests/lab-admin-testing
  session:
    objects:
    - apiVersion: rbac.authorization.k8s.io/v1
      kind: ClusterRoleBinding
      metadata:
        name: $(session_namespace)-cluster-admin
      roleRef:
        apiGroup: rbac.authorization.k8s.io
        kind: ClusterRole
        name: cluster-admin
      subjects:
      - kind: ServiceAccount
        namespace: $(workshop_namespace)
        name: $(service_account)

In this case, the name of the cluster role binding resource embeds $(session_namespace) so that its name is unique to the workshop instance and doesn't overlap with a binding for a different workshop instance.

Running user containers as root

In addition to RBAC, which controls what resources a user can create and work with, Pod security policies are applied to restrict what Pods/containers a user deploys can do.

By default, the deployments that a workshop user can create are allowed only to run containers as a non-root user. This means that many container images available on registries such as Docker Hub cannot be used.

If you are creating a workshop where a user must run containers as the root user, you must override the default nonroot security policy and select the anyuid security policy by using the session.namespaces.security.policy setting:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-policy-testing
spec:
  title: Policy Testing
  description: Play area for testing security policies
  content:
    files: github.com/eduk8s-tests/lab-policy-testing
  session:
    namespaces:
      security:
        policy: anyuid

This setting applies to the primary session namespace and any secondary namespaces created.

Creating additional namespaces

For each workshop instance, a primary session namespace is created. Applications can be pre-deployed or deployed into this namespace as part of the workshop.

If you need more than one namespace per workshop instance, you can create secondary namespaces in a a couple of ways.

If the secondary namespaces are to be created empty, you can list the details of the namespaces under the property session.namespaces.secondary:

    apiVersion: learningcenter.tanzu.vmware.com/v1beta1
    kind: Workshop
    metadata:
      name: lab-namespace-testing
    spec:
      title: Namespace Testing
      description: Play area for testing namespaces
      content:
        files: github.com/eduk8s-tests/lab-namespace-testing
      session:
        namespaces:
          role: admin
          budget: medium
          secondary:
          - name: $(session_namespace)-apps
            role: edit
            budget: large
            limits:
              default:
                memory: 512mi

When secondary namespaces are created, by default, the role, resource quotas, and limit ranges are set the same as the primary session namespace. Each namespace, though, has a separate resource budget; it is not shared.

If required, you can override what role, budget, and limits are applied within the entry for the namespace.

Similarly, you can override the security policy for secondary namespaces on a case-by-case basis by adding the security.policy setting under the entry for the secondary namespace.

To create resources in the namespaces you create, create the namespaces by adding an appropriate Namespace resource to session.objects with the definitions of the resources you want to create in the namespaces:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-namespace-testing
spec:
  title: Namespace Testing
  description: Play area for testing namespaces
  content:
    files: github.com/eduk8s-tests/lab-namespace-testing
  session:
    objects:
    - apiVersion: v1
      kind: Namespace
      metadata:
        name: $(session_namespace)-apps

When listing any other resources to be created within the added namespace, such as deployments, ensure that the namespace is set in the metadata of the resource. For example, $(session_namespace)-apps.

To override what role the service account for the workshop instance has in the added namespace, you can set the learningcenter.tanzu.vmware.com/session.role annotation on the Namespace resource:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-namespace-testing
spec:
  title: Namespace Testing
  description: Play area for testing namespaces
  content:
    files: github.com/eduk8s-tests/lab-namespace-testing
  session:
    objects:
    - apiVersion: v1
      kind: Namespace
      metadata:
        name: $(session_namespace)-apps
        annotations:
          learningcenter.tanzu.vmware.com/session.role: view

To have a different resource budget set for the additional namespace, you can add the annotation learningcenter.tanzu.vmware.com/session.budget in the Namespace resource metadata and set the value to the required resource budget:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-namespace-testing
spec:
  title: Namespace Testing
  description: Play area for testing namespaces
  content:
    files: github.com/eduk8s-tests/lab-namespace-testing
  session:
    objects:
    - apiVersion: v1
      kind: Namespace
      metadata:
        name: $(session_namespace)-apps
        annotations:
          learningcenter.tanzu.vmware.com/session.budget: large

To override the limit range values applied corresponding to the budget applied, you can add annotations starting with learningcenter.tanzu.vmware.com/session.limits. for each entry:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-namespace-testing
spec:
  title: Namespace Testing
  description: Play area for testing namespaces
  content:
    files: github.com/eduk8s-tests/lab-namespace-testing
  session:
    objects:
    - apiVersion: v1
      kind: Namespace
      metadata:
        name: $(session_namespace)-apps
        annotations:
          learningcenter.tanzu.vmware.com/session.limits.min.cpu: 50m
          learningcenter.tanzu.vmware.com/session.limits.min.memory: 32Mi
          learningcenter.tanzu.vmware.com/session.limits.max.cpu: 1
          learningcenter.tanzu.vmware.com/session.limits.max.memory: 1Gi
          learningcenter.tanzu.vmware.com/session.limits.defaultrequest.cpu: 50m
          learningcenter.tanzu.vmware.com/session.limits.defaultrequest.memory: 128Mi
          learningcenter.tanzu.vmware.com/session.limits.request.cpu: 500m
          learningcenter.tanzu.vmware.com/session.limits.request.memory: 1Gi

You only must supply annotations for the values you want to override.

If you need more fine-grained control over the limit ranges and resource quotas, set the value of the annotation for the budget to custom and add the LimitRange and ResourceQuota definitions to session.objects.

In this case, you must set the namespace for the LimitRange and ResourceQuota resource to the name of the namespace, e.g., $(session_namespace)-apps so they are only applied to that namespace.

To set the security policy for a specific namespace other than the primary session namespace, you can add the annotation learningcenter.tanzu.vmware.com/session.security.policy in the Namespace resource metadata and set the value to nonroot, anyuid or custom as necessary.

Shared workshop resources

Adding a list of resources to session.objects cause the given resources to be created for each workshop instance, whereas namespaced resources default to being created in the session namespace for a workshop instance.

If you want to have one common shared set of resources created once for the whole workshop environment that is used by all workshop instances, you can list them in the environment.objects field.

For example, this might be used to deploy a single image registry used by all workshop instances. A Kubernetes job is used to import a set of images into the image registry, which is then referenced by the workshop instances.

For namespaced resources, it is not necessary to specify the namespace field of the resource metadata. When the namespace field is not present, the resource is automatically created within the workshop namespace for that workshop environment.

When resources are created, owner references are added, making the WorkshopEnvironment custom resource correspond to the workshop environment of the owner. This means that any resources are also deleted when the workshop environment is deleted.

Values of fields in the list of resource objects can reference a number of predefined parameters. The available parameters are:

  • workshop_ name - The name of the workshop. This is the name of the Workshop definition the workshop environment was created against.
  • environment_ name - The name of the workshop environment. This is the same as the name of the namespace for the workshop environment. Do not rely on their being the same, and use the most appropriate to cope with any potential change.
  • environment_ token - The value of the token that needs to be used in workshop requests against the workshop environment.
  • workshop_ namespace - The namespace for the workshop environment. This is the namespace where all deployments of the workshop instances, and their service accounts, are created. It is the same namespace that shared workshop resources are created.
  • service_ account - The name of a service account you can use when creating deployments in the workshop namespace.
  • ingress_ domain - The host domain under which you can create hostnames when creating ingress routes.
  • ingress_ protocol - The protocol (http/https) used for ingress routes created for workshops.
  • ingress_ secret - The name of the ingress secret stored in the workshop namespace when secure ingress is used.

To create additional namespaces associated with the workshop environment, embed a reference to $(workshop_namespace) in the name of the additional namespaces with an appropriate suffix. Be careful that the suffix doesn't overlap with the range of session IDs for workshop instances.

When creating deployments in the workshop namespace, set the serviceAccountName of the Deployment resource to $(service_account). This ensures the deployment uses a special Pod security policy set up by the Learning Center. If this isn't used and the cluster imposes a more strict default Pod security policy, your deployment might not work, especially if any image runs as root.

Workshop pod security policy

The Pod for the workshop session is set up with a Pod security policy that restricts what can be done from containers in the pod. The nature of the applied Pod security policy is adjusted when enabling support for doing Docker builds. This enables Docker builds inside the sidecar container attached to the workshop container.

If you are customizing the workshop by patching the Pod specification using session.patches to add your own sidecar container, and that sidecar container must run as the root user or needs a custom Pod security policy, you must override the default security policy for the workshop container.

To allow a sidecar container to run as the root user with no extra privileges required, you can override the default nonroot security policy and set it to anyuid:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-policy-testing
spec:
  title: Policy Testing
  description: Play area for testing security policies
  content:
    files: github.com/eduk8s-tests/lab-policy-testing
  session:
    security:
      policy: anyuid

This is a different setting than described previously for changing the security policy for deployments made by a workshop user to the session namespaces. This setting applies only to the workshop container itself.

If you need more fine-grained control of the security policy, you must provide your own resources for defining the Pod security policy and map it, so it is used. The details of the Pod security policy must be in environment.objects and mapped by definitions added to session.objects. For this to be used, you must disable the application of the inbuilt Pod security policies. You can do this by setting session.security.policy to custom:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-policy-testing
spec:
  title: Policy Testing
  description: Play area for testing policy override
  content:
    files: github.com/eduk8s-tests/lab-policy-testing
  session:
    security:
      policy: custom
    objects:
    - apiVersion: rbac.authorization.k8s.io/v1
      kind: RoleBinding
      metadata:
        namespace: $(workshop_namespace)
        name: $(session_namespace)-podman
      roleRef:
        apiGroup: rbac.authorization.k8s.io
        kind: ClusterRole
        name: $(workshop_namespace)-podman
      subjects:
      - kind: ServiceAccount
        namespace: $(workshop_namespace)
        name: $(service_account)
  environment:
    objects:
    - apiVersion: policy/v1beta1
      kind: PodSecurityPolicy
      metadata:
        name: aa-$(workshop_namespace)-podman
      spec:
        privileged: true
        allowPrivilegeEscalation: true
        requiredDropCapabilities:
        - KILL
        - MKNOD
        hostIPC: false
        hostNetwork: false
        hostPID: false
        hostPorts: []
        runAsUser:
          rule: MustRunAsNonRoot
        seLinux:
          rule: RunAsAny
        fsGroup:
          rule: RunAsAny
        supplementalGroups:
          rule: RunAsAny
        volumes:
        - configMap
        - downwardAPI
        - emptyDir
        - persistentVolumeClaim
        - projected
        - secret
    - apiVersion: rbac.authorization.k8s.io/v1
      kind: ClusterRole
      metadata:
        name: $(workshop_namespace)-podman
      rules:
      - apiGroups:
        - policy
        resources:
        - podsecuritypolicies
        verbs:
        - use
        resourceNames:
        - aa-$(workshop_namespace)-podman

By overriding the Pod security policy, you are responsible for limiting what can be done from the workshop Pod. In other words, add only the extra capabilities you need. The Pod security policy is applied only to the Pod the workshop session runs in. It does not affect any Pod security policy applied to service accounts that exist in the session namespace or other namespaces you have created.

There is a better way to set the priority of applied Pod security policies when a default Pod security policy is applied globally by mapping it to the system:authenticated group. This results in priority falling back to the order of the names of the Pod security policies. VMware recommends using aa- as a prefix to the custom Pod security name you create. This ensures it takes precedence over any global default Pod security policy such as restricted, pks-restricted or vmware-system-tmc-restricted, no matter what the name of the global policy default.

Custom security policies for user containers

There is also the option to set the value of the session.namespaces.security.policy setting as custom. This allows for more fine-grained control of the security policy applied to the Pods/containers a user deploys during a session. In this case, you must provide your own resources defining the Pod security policy and map it so it is used.

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-policy-testing
spec:
  title: Policy Testing
  description: Play area for testing policy override
  content:
    files: github.com/eduk8s-tests/lab-policy-testing
  session:
    namespaes:
      security:
        policy: custom
    objects:
    - apiVersion: rbac.authorization.k8s.io/v1
      kind: RoleBinding
      metadata:
        namespace: $(workshop_namespace)
        name: $(session_namespace)-security-policy
      roleRef:
        apiGroup: rbac.authorization.k8s.io
        kind: ClusterRole
        name: $(workshop_namespace)-security-policy
      subjects:
      - kind: Group 
        namespace: $(workshop_namespace)
        name: system:serviceaccounts:$(workshop_namespace)
  environment:
    objects:
    - apiVersion: policy/v1beta1
      kind: PodSecurityPolicy
      metadata:
        name: aa-$(workshop_namespace)-security-policy
      spec:
        privileged: true
        allowPrivilegeEscalation: true
        requiredDropCapabilities:
        - KILL
        - MKNOD
        hostIPC: false
        hostNetwork: false
        hostPID: false
        hostPorts: []
        runAsUser:
          rule: MustRunAsNonRoot
        seLinux:
          rule: RunAsAny
        fsGroup:
          rule: RunAsAny
        supplementalGroups:
          rule: RunAsAny
        volumes:
        - configMap
        - downwardAPI
        - emptyDir
        - persistentVolumeClaim
        - projected
        - secret
    - apiVersion: rbac.authorization.k8s.io/v1
      kind: ClusterRole
      metadata:
        name: $(workshop_namespace)-security-policy
      rules:
      - apiGroups:
        - policy
        resources:
        - podsecuritypolicies
        verbs:
        - use
        resourceNames:
        - aa-$(workshop_namespace)-security-policy

This can also be done on secondary namespaces by either setting the session.namespaces.secondary.security.policy setting to custom or using the learningcenter.tanzu.vmware.com/session.security.policy: custom annotation.

Defining additional ingress points

By default, if running additional background applications, they are only accessible to other processes within the same container. For an application to be accessible to a user through their web browser, an ingress must be created mapping to the port for the application.

You can do this by supplying a list of the ingress points and the internal container port they map to by setting the session.ingresses field in the workshop definition:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    ingresses:
    - name: application
      port: 8080

The form of the hostname used in the URL to access the service is:

$(session_namespace)-application.$(ingress_domain)

Do not use the name of any built-in dashboards, terminal, console, slides, or editor for this name. These are reserved for the corresponding built-in capabilities providing those features.

In addition to specifying ingresses for proxying to internal ports within the same Pod, you can specify a host, protocol, and port corresponding to a separate service running in the Kubernetes cluster:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    ingresses:
    - name: application
      protocol: http
      host: service.namespace.svc.cluster.local
      port: 8080

You can use variables providing information about the current session within the host property if required:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    ingresses:
    - name: application
      protocol: http
      host: service.$(session_namespace).svc.cluster.local
      port: 8080

Available variables are:

  • session_ namespace - The namespace created for and bound to the workshop instance. This is the namespace unique to the session and where a workshop can create its own resources.
  • environment_ name - The name of the workshop environment. Currently, this is the same as the namespace name for the workshop environment. Do not rely on their being the same, and use the most appropriate to cope with any potential change.
  • workshop_ namespace - The namespace for the workshop environment. This is the namespace where all deployments of the workshop instances are created and where the service account that the workshop instance runs as exists.
  • ingress_ domain - The host domain under which you can create hostnames when creating ingress routes.

If the service uses standard http or https ports, you can leave out the port property, and the port is set based on the value of protocol.

When a request is proxied, you can specify additional request headers that must be passed to the service:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    ingresses:
    - name: application
      protocol: http
      host: service.$(session_namespace).svc.cluster.local
      port: 8080
      headers:
      - name: Authorization
        value: "Bearer $(kubernetes_token)"

The value of a header can reference the following variables.

  • kubernetes_ token - The access token of the service account for the current workshop session, used for accessing the Kubernetes REST API.

Accessing any service through the ingress is protected by any access controls enforced by the workshop environment or training portal. If you use the training portal, this should be transparent. Otherwise, supply any login credentials for the workshop when prompted by your web browser.

External workshop instructions

In place of using workshop instructions provided with the workshop content, you can use externally hosted instructions instead. To do this, set sessions.applications.workshop.URL to the URL of an external website:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      workshop:
        url: https://www.example.com/instructions

The external website must be displayed in an HTML iframe, is shown as is, and must provide its page navigation and table of contents if required.

The URL value can reference a number of predefined parameters. The available parameters are:

  • session_ namespace - The namespace created for and bound to the workshop instance. This is the namespace unique to the session and where a workshop can create its own resources.
  • environment_ name - The name of the workshop environment. This is the same as the name of the namespace for the workshop environment. Do not rely on their being the same, and use the most appropriate to cope with any potential change.
  • workshop_ namespace - The namespace for the workshop environment. This is the namespace where all deployments of the workshop instances are created and where the service account that the workshop instance runs as exists.
  • ingress_ domain - The host domain under which you can create hostnames when creating ingress routes.
  • ingress_ protocol - The protocol (http/https) used for ingress routes that are created for workshops.

These could be used, for example, to reference workshops instructions hosted as part of the workshop environment:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      workshop:
        url: $(ingress_protocol)://$(workshop_namespace)-instructions.$(ingress_domain)
  environment:
    objects:
    - ...

In this case environment.objects of the workshop spec must include resources to deploy the application hosting the instructions and expose it through an appropriate ingress.

Disabling workshop instructions

The workshop environment aims to provide instructions for a workshop that users can follow. If you want to use the workshop environment as a development environment or as an administration console that provides access to a Kubernetes cluster, you can disable the display of workshop instructions provided with the workshop content. In this case, only the work area with the terminals, console, and so on is displayed. To disable the display of workshop instructions, add a session.applications.workshop section and set the enabled property to false:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      workshop:
        enabled: false

Enabling the Kubernetes console

By default, the Kubernetes console is not enabled. To enable it and make it available through the web browser, when accessing a workshop, add a session.applications.console section to the workshop definition, and set the enabled property to true:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      console:
        enabled: true

The Kubernetes dashboard provided by the Kubernetes project is used. To use Octant as the console, you can set the vendor property to octant:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      console:
        enabled: true
        vendor: octant

When vendor is not set, kubernetes is assumed.

Enabling the integrated editor

By default, the integrated web-based editor is not enabled. To enable it and make it available through the web browser when accessing a workshop, add a session.applications.editor section to the workshop definition, and set the enabled property to true:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      editor:
        enabled: true

The integrated editor used is based on Visual Studio Code. For more information about the editor, see https://github.com/cdr/code-server in GitHub.

To install additional VS Code extensions, do this from the editor. Alternatively, if building a custom workshop, you can install them from your Dockerfile into your workshop image by running:

code-server --install-extension vendor.extension

Replace vendor.extension with the name of the extension, where the name identifies the extension on the VS Code extensions marketplace used by the editor or provide a pathname to a local .vsix file.

This installs the extensions into $HOME/.config/code-server/extensions.

If downloading extensions yourself and unpacking them or extensions are part of your Git repository, you can instead locate them in the workshop/code-server/extensions directory.

Enabling workshop downloads

You can provide a way for a workshop user to download files as part of the workshop content. Enable this by adding the session.applications.files section to the workshop definition and setting the enabled property to true:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      files:
        enabled: true

The recommended way to provide access to files from workshop instructions is to use the files:download-file clickable action block. This action ensures any file is downloaded to the local machine and is not simply displayed in the browser in place of the workshop instructions.

By default, the user can access any files located under the home directory of the workshop user account. To restrict where the user can download files from, set the directory setting:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      files:
        enabled: true
        directory: exercises

When the specified directory is a relative path, it is evaluated relative to the home directory of the workshop user.

Enabling the test examiner

The test examiner is a feature that allows a workshop to have verification checks that can be triggered from the workshop instructions. The test examiner is deactivated by default. To enable it, add a session.applications.examiner section to the workshop definition and set the enabled property to true:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      examiner:
        enabled: true

Any executable test programs to be used for verification checks must be provided in the workshop/examiner/tests directory.

The test programs must return an exit status of 0 if the test is successful and non zero if it fails. Test programs must not be persistent programs that can run forever.

Clickable actions for the test examiner are used within the workshop instructions to trigger the verification checks. You can configure them to be automatically started when the page of the workshop instructions is loaded.

Enabling session image registry

Workshops using tools such as kpack or tekton that need a place to push container images when built can enable a container image registry. A separate registry is deployed for each workshop session.

The image registry is currently fully usable only if workshops are deployed under a Learning Center. Operator configuration that uses secure ingress. This is because a registry that is not secure is not trusted by the Kubernetes cluster as the source of container images when doing deployments.

To enable the deployment of a registry per workshop session, add a session.applications.registry section to the workshop definition and set the enabled property to true:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      registry:
        enabled: true

The registry mounts a persistent volume for storing images. By default, the size of that persistent volume is 5Gi. To override the size of the persistent volume, add the storage property under the registry section:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      registry:
        enabled: true
        storage: 20Gi

The amount of memory provided to the registry defaults to 768Mi. To increase this, add the memory property under the registry section.

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      registry:
        enabled: true
        memory: 1Gi

The registry is secured with a user name and password unique to the workshop session and should be accessed over a secure connection.

To allow access from the workshop session, the file $HOME/.docker/config.json containing the registry credentials is injected into the workshop session. This is automatically used by tools such as docker.

For deployments in Kubernetes, a secret of type kubernetes.io/dockerconfigjson is created in the namespace and automatically applied to the default service account in the namespace. This means deployments made using the default service account can pull images from the registry without additional configuration. If creating deployments using other service accounts, add configuration to the service account or deployment to add the registry secret for pulling images.

If you need access to the raw registry host details and credentials, they are provided as environment variables in the workshop session. The environment variables are:

  • REGISTRY_ HOST - Contains the hostname for the registry for the workshop session.
  • REGISTRY_ AUTH_FILE - Contains the location of the docker configuration file. Must be the equivalent of $HOME/.docker/config.json.
  • REGISTRY_ USERNAME - Contains the user name for accessing the registry.
  • REGISTRY_ PASSWORD - Contains the password for accessing the registry. This is different for each workshop session.
  • REGISTRY_ SECRET - Contains the name of a Kubernetes secret of type kubernetes.io/dockerconfigjson added to the session namespace, which contains the registry credentials.

The URL for accessing the registry adopts the HTTP protocol scheme inherited from the environment variable INGRESS_PROTOCOL. This is the same HTTP protocol scheme the workshop sessions use.

To use any of the variables as data variables in workshop content, use the same variable name but in lowercase: registry_host, registry_auth_file, registry_username, registry_password and registry_secret.

Enabling ability to use Docker

To build container images in a workshop using docker, you must first enable it. Each workshop session has its own separate Docker daemon instance running in a container.

Enabling support for running docker requires the use of a privileged container for running the Docker daemon. Because of the security implications of providing access to Docker with this configuration, VMware recommends that if you don't trust the people taking the workshop, any workshops that require Docker only be hosted in a disposable Kubernetes cluster that is destroyed after the workshop. You should never enable Docker for workshops hosted on a public service that is always kept running and where arbitrary users could access the workshops.

To enable support for using docker, add a session.applications.docker section to the workshop definition and set the enabled property to true:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      docker:
        enabled: true

The container that runs the Docker daemon mounts a persistent volume to store images that are pulled down or built locally. By default. the size of that persistent volume is 5Gi. To override the size of the persistent volume, add the storage property under the docker section:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      docker:
        enabled: true
        storage: 20Gi

The amount of memory provided to the container running the Docker daemon defaults to 768Mi. To increase this, add the memory property under the registry section:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      docker:
        enabled: true
        memory: 1Gi

The workshop session access to the Docker daemon uses a local Unix socket shared with the container running the Docker daemon. If it uses a local tool to access the socket connection for the Docker daemon directly rather than by running docker, it should use the DOCKER_HOST environment variable to set the location of the socket.

The Docker daemon is only available from within the workshop session and cannot be accessed outside of the Pod by any tools deployed separately to Kubernetes.

Enabling WebDAV access to files

You can access or update local files within the workshop session from the terminal command line or editor of the workshop dashboard. The local files reside in the file system of the container the the workshop session is running in.

To access the files remotely, you can enable WebDAV support for the workshop session.

To enable support for accessing files over WebDAV, add a session.applications.webdav section to the workshop definition, and set the enabled property to true:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      webdav:
        enabled: true

This results in a WebDAV server running within the workshop session environment. A set of credentials is also automatically generated and are available as environment variables. The environment variables are:

  • WEBDAV_ USERNAME - Contains the user name that must be used when authenticating over WebDAV.
  • WEBDAV_ PASSWORD - Contains the password that must be used when authenticating over WebDAV.

To use any environment variables related to the image registry as data variables in workshop content, declare this in the workshop/modules.yaml file in the config.vars section:

config:
  vars:
  - name: WEBDAV_USERNAME
  - name: WEBDAV_PASSWORD

The URL endpoint for accessing the WebDAV server is the same as the workshop session, with /webdav/ path added. This can be constructed from the terminal using:

$INGRESS_PROTOCOL://$SESSION_NAMESPACE.$INGRESS_DOMAIN/webdav/

In workshop content it can be constructed using:

{{ingress_protocol}}://{{session_namespace}}.{{ingress_domain}}/webdav/

You must be able to use WebDAV client support provided by your operating system or by using a standalone WebDAV client, such as CyberDuck.

Using WebDAV can make transferring files to or from the workshop session easier.

Customizing the terminal layout

By default, a single terminal is provided in the web browser when accessing the workshop. If required, you can enable alternate layouts which provide additional terminals. To set the layout, add the session.applications.terminal section and include the layout property with the desired layout:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    applications:
      terminal:
        enabled: true
        layout: split

The options for the layout property are:

  • default - Single terminal.
  • split - Two terminals stacked above each other in ratio 60/40.
  • split/2 - Three terminals stacked above each other in ratio 50/25/25.
  • lower - A single terminal is placed below any dashboard tabs, rather than a tab of its own. The ratio of dashboard tab to terminal is 70/30.
  • none - No terminal is displayed but can still be created from the drop-down menu.

When adding the terminal section, you must include the enabled property and set it to accurate as it is a required field when including the section.

If you do not want a terminal displayed and disable the ability to create terminals from the drop-down menu, set enabled to false.

Adding custom dashboard tabs

Exposed applications, external sites, and additional terminals can be given their own custom dashboard tab. This is done by specifying the list of dashboard panels and the target URL:

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    ingresses:
    - name: application
      port: 8080
    dashboards:
    - name: Internal
      url: "$(ingress_protocol)://$(session_namespace)-application.$(ingress_domain)/"
    - name: External
      url: http://www.example.com

The URL values can reference a number of predefined parameters. The available parameters are:

  • session_ namespace - The namespace created for and bound to the workshop instance. This is the namespace unique to the session and where a workshop can create its own resources.
  • environment_ name - The name of the workshop environment. This is the same as the name of the namespace for the workshop environment. Do not rely on their being the same, and use the most appropriate to cope with any potential change.
  • workshop_ namespace - The namespace for the workshop environment. This is the namespace where all deployments of the workshop instances are created and where the service account that the workshop instance runs as exists.
  • ingress_ domain - The host domain under which you can create hostnames when creating ingress routes.
  • ingress_ protocol - The protocol (http/https) used for ingress routes that are created for workshops.

The URL can reference an external website; however, that website must not prohibit being embedded in an HTML iframe.

If you want to have a custom dashboard tab provide an additional terminal, the url property must use the form terminal:<session>, where <session> is replaced with the name of the terminal session. The name of the terminal session can be any name you choose, but must be restricted to lowercase letters, numbers, and dashes. You should avoid using numeric terminal session names such as "1", "2", and "3" as these are used for the default terminal sessions.

apiVersion: learningcenter.tanzu.vmware.com/v1beta1
kind: Workshop
metadata:
  name: lab-application-testing
spec:
  title: Application Testing
  description: Play area for testing my application
  content:
    image: quay.io/eduk8s-tests/lab-application-testing:master
  session:
    dashboards:
    - name: Example
      url: terminal:example