Skip to content

Latest commit

 

History

History
1010 lines (778 loc) · 31.9 KB

tkgs.md

File metadata and controls

1010 lines (778 loc) · 31.9 KB

Tanzu Kubernetes Grid Service (TKGs)

Table of Content

Prerequisites

vSphere Networking

  • VMware vSphere 7 Enterprise Plus and Tanzu Edition license keys
  • HA and DRS enabled vSphere cluster (fully automated)
  • Shared storage connected to all ESXi Hosts
  • NTP configured and working on all components
  • DNS server configured, reachable, and working
  • 5x free IPs in a row on the management network for the Supervisor Control Plane
  • VDS version 7.0.0
  • Separate and fully routed workload and management network on different VLANs (third Frontend network is recommended)

NSX Networking

  • VMware vSphere 7 Enterprise Plus and Tanzu Edition license keys
  • HA and DRS enabled vSphere cluster (fully automated)
  • Shared storage connected to all ESXi hosts
  • NTP configured and working on all components
  • DNS server configured, reachable, and working
  • 5x free IPs in a row on the management network for the Supervisor Control Plane
  • VDS version 7.0.0
  • at least 1x VLAN and subnet for the overlay network (2x if you don’t have a separate vmnic to spare)
  • NSX-T 3.x with a working overlay network
  • at least 1x NSX Manager and 1x Edge node (LARGE)
  • MTU size of at least 1600
  • 2x /27 subnets for ingress and egress traffic

Licensing

  • Licensing for vSphere with Tanzu

    • VMware vSphere 7 Enterprise Plus with Add-on for Kubernetes license assigned to each host from the Supervisor cluster
    • Tanzu license for the Supervisor cluster
    • You can assign a Tanzu edition license key to multiple Supervisor clusters at a time
  • Prerequisites for Configuring vSphere with Tanzu on a Cluster

    • vSphere Cluster with at least 3 hosts
      • 4 hosts with vSAN
      • HA and DRS (fully automated) is required
    • vSAN File Services for persistent volumes in ReadWriteMany mode
    • User account has the Modify cluster-wide configuration permission to enable Workload Management

Table I: Choose and configure the networking stack

Functionality NSX-T Networking vDS Networking Comment
vSphere Pods Yes No
Tanzu Kubernetes clusters Yes Yes
Embedded Harbor Registry Yes No
Load Balancing Yes Yes, by installing and configuring the NSX Advanced Load Balancer or the HAProxy load balancer. vSphere with Tanzu supports the NSX Advanced Load Balancer with vSphere 7 U2 and later.

Content Library

Domain Activations

  • *.tmc.cloud.vmware.com
  • *.console.cloud.vmware.com
  • *.cloud.vmware.com
  • *.projects.registry.vmware.com
  • *.registry.vmware.com
  • *.registry.pivotal.io
  • *.github.com
  • *.githubusercontent.com
  • *.vmware.com
  • *.docker.io
  • *.docker.com

Binaries

Since vSphere with Tanzu (Workload Management) is a vSphere Cluster feature, it's not necessary to download any additional installation binaries.

CLIs

Kubectl

Install and Set Up kubectl on Linux

  • Install kubectl binary with curl on Linux
curl -LO "https://dl.k8s.io/release/$(curl -L -s https://dl.k8s.io/release/stable.txt)/bin/linux/amd64/kubectl"
  • Download a specific kubectl version
curl -LO https://dl.k8s.io/release/v1.23.0/bin/linux/amd64/kubectl
  • Install using native package management
sudo apt-get update

sudo apt-get install -y apt-transport-https ca-certificates curl

sudo curl -fsSLo /usr/share/keyrings/kubernetes-archive-keyring.gpg https://packages.cloud.google.com/apt/doc/apt-key.gpg

echo "deb [signed-by=/usr/share/keyrings/kubernetes-archive-keyring.gpg] https://apt.kubernetes.io/ kubernetes-xenial main" | sudo tee /etc/apt/sources.list.d/kubernetes.list

sudo apt-get update

sudo apt-get install -y kubectl
  • Autocompletion kubectl
source <(kubectl completion bash)

echo "source <(kubectl completion bash)" >> ~/.bashrc
  • Set alias for kubectl to just k
echo "alias k='kubectl'" >> ~/.bashrc

complete -F __start_kubectl k

Kubectl vSphere

Download and Install the Kubernetes CLI Tools for vSphere

  • Password as an environment variable:
export KUBECTL_VSPHERE_PASSWORD=P@ssw0rd

kubectl vsphere login --vsphere-username administrator@vsphere.local --server=(IP or FQDN)

Networking

Load Balancer Feature Comparison

Supervisor Networking

Tanzu Kubernetes Cluster Networking

The table summarizes Tanzu Kubernetes cluster networking features and their implementation.

Endpoint Provider Description
Pod connectivity Antrea or Calico Container network interface for pods. Antrea uses Open vSwitch. Calico uses the Linux bridge with BGP
Service type: ClusterIP Antrea or Calico Default Kubernetes service type that is only accessible from within the cluster
Service type: NodePort Antrea or Calico Allows external access through a port opened on each worker node by the Kubernetes network proxy
Service type: LoadBalancer NSX-T load balancer, NSX Advanced Load Balancer, HAProxy For NSX-T, one virtual server per service type definition. For NSX Advanced Load Balancer, refer to that section of this documentation Note: Some load balancing features may not be available with HAProxy, such as support for static IPs
Cluster ingress Third-party ingress controller Routing for inbound pod traffic; you can use any third-party ingress controller
Network policy Antrea or Calico Controls what traffic is allowed to and from selected pods and network endpoints. Antrea uses Open vSwitch. Calico uses Linux IP tables

DNS Resolve flow

Tanzu Kubernetes Cluster

Available Virtual Machine Classes

Virtual Machine Classes for Tanzu Kubernetes Clusters

Class CPU Memory (GB) Reserved CPU and Memory
guaranteed-8xlarge 32 128 Yes
best-effort-8xlarge 32 128 No
guaranteed-4xlarge 16 128 Yes
best-effort-4xlarge 16 128 No
guaranteed-2xlarge 8 64 Yes
best-effort-2xlarge 8 64 No
guaranteed-xlarge 4 32 Yes
best-effort-xlarge 4 32 No
guaranteed-large 4 16 Yes
best-effort-large 4 16 No
guaranteed-medium 2 8 Yes
best-effort-medium 2 8 No
guaranteed-small 2 4 Yes
best-effort-small 2 4 No
guaranteed-xsmall 2 2 Yes
best-effort-xsmall 2 2 No

Example Manifest - Advanced

Provisioning Tanzu Kubernetes Clusters Using the Tanzu Kubernetes Grid Service v1alpha2 API

Full Example Manifest file to deploy a Tanzu Kubernetes Cluster:

apiVersion: run.tanzu.vmware.com/v1alpha2
kind: TanzuKubernetesCluster
metadata:
  name: tkgs-v2-cluster-custom
  namespace: tkgs-cluster-ns
spec:
  topology:
    controlPlane:
      replicas: 3
      vmClass: guaranteed-medium
      storageClass: vwt-storage-policy
      volumes:
        - name: etcd
          mountPath: /var/lib/etcd
          capacity:
            storage: 4Gi
      tkr:
        reference:
          name: v1.21.2---vmware.1-tkg.1.ee25d55
    nodePools:
    - name: worker-nodepool-a1
      replicas: 3
      vmClass: guaranteed-large
      storageClass: vwt-storage-policy
      volumes:
        - name: containerd
          mountPath: /var/lib/containerd
          capacity:
            storage: 16Gi
      tkr:
        reference:
          name: v1.21.2---vmware.1-tkg.1.ee25d55
    - name: worker-nodepool-a2
      replicas: 2
      vmClass: guaranteed-medium
      storageClass: vwt-storage-policy
      tkr:
        reference:
          name: v1.21.2---vmware.1-tkg.1.ee25d55
    - name: worker-nodepool-a3
      replicas: 1
      vmClass: guaranteed-small
      storageClass: vwt-storage-policy
      tkr:
        reference:
          name: v1.21.2---vmware.1-tkg.1.ee25d55
  settings:
    storage:
      defaultClass: vwt-storage-policy
    network:
      cni:
        name: antrea
      services:
        cidrBlocks: ["198.53.100.0/16"]
      pods:
        cidrBlocks: ["192.0.5.0/16"]
      serviceDomain: managedcluster.local
      proxy:
        httpProxy: http://<user>:<pwd>@<ip>:<port>
        httpsProxy: http://<user>:<pwd>@<ip>:<port>
        noProxy: [10.246.0.0/16,192.168.144.0/20,192.168.128.0/20]
      trust:
        additionalTrustedCAs:
          - name: CompanyInternalCA-1
            data: LS0tLS1C...LS0tCg==
          - name: CompanyInternalCA-2
            data: MTLtMT1C...MT0tPg==

Example Manifest - Simple

apiVersion: run.tanzu.vmware.com/v1alpha1
kind: TanzuKubernetesCluster
metadata:
  name:
  namespace:
spec:
  distribution:
    version: v1.21.6
  settings:
    network:
      cni:
        name: antrea
      pods:
        cidrBlocks:
        - 193.0.2.0/16
      services:
        cidrBlocks:
        - 195.51.100.0/12
  topology:
    controlPlane:
      class: best-effort-small
      count: 3
      storageClass:
    workers:
      class: best-effort-medium
      count: 3
      storageClass:

Pod Security Policy

Using Pod Security Policies with Tanzu Kubernetes Clusters

Tanzu Kubernetes Grid Service provisions Tanzu Kubernetes clusters with the PodSecurityPolicy Admission Controller enabled. This means that pod security policy is required to deploy workloads. Cluster administrators can deploy pods from their user account to any namespace, and from service accounts to the kube-system namespace. For all other use cases, you must explicitly bind to a PodSecurityPolicy object. Clusters include default pod security policies that you can bind to, or create your own.

apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
  name: psp:privileged
rules:
- apiGroups: ['policy']
  resources: ['podsecuritypolicies']
  verbs:     ['use']
  resourceNames:
  - vmware-system-privileged
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
  name: all:psp:privileged
roleRef:
  kind: ClusterRole
  name: psp:privileged
  apiGroup: rbac.authorization.k8s.io
subjects:
- kind: Group
  name: system:serviceaccounts
  apiGroup: rbac.authorization.k8s.io

Troubleshooting

Support Bundle

Collect a Support Bundle for Tanzu Kubernetes Clusters

To troubleshoot Tanzu Kubernetes cluster errors, you can run a utility to collect a diagnostic log bundle. Obtain the utility from here: KB80949

Show Kubernetes vSphere Objects

  • kubectl get virtualmachineimages
  • kubectl get virtualmachineclasses
  • kubectl get tanzukubernetescluster
    • kubectl get tkc

vCenter Server Appliance Service Troubleshooting

Show service status using vmon-cli

  • List all services: vmon-cli --list
  • Show status of a service: vmon-cli --status
  • List dependencies of a service on other services:
vmon-cli --list-dependencies wcp

cis-license
eam
lookupsvc
sts
trustmanagement
vapi-endpoint
vmware-vpostgres
vpxd
  • Start a service: vmon-cli --start
  • Stop a services: vmon-cli --stop
  • Restart a service: vmon-cli --restart

Show service status using service-control

service-control --help
usage: service-control [-h] [--start] [--stop] [--restart] [--status] [--list] [--list-services] [--list-dependencies] [--vmon-profile VMON_PROFILE] [--all] [--ignore] [--verbose] ...

Perform operation on VMware services. By default the services selected are based on current startup profile. This can be overridden by using --all and --vmon-profile options.

positional arguments:
  services              Services on which to operate

optional arguments:
  -h, --help            show this help message and exit
  --start               Perform start operation on VMware services.
  --stop                Perform stop operation on VMware services.
  --restart             Perform restart operation on a single VMware service.
  --status              Get running status of VMware services.
  --list                List all controllable VMware services. This option will soon be deprecated. Please use --list-services.
  --list-services       Lists all controllable VMware services.
  --list-dependencies   Lists dependencies for a given service.
  --vmon-profile VMON_PROFILE
                        Start/Stop services associated with given profile name.
  --all                 Start/Stop all VMware services i.e. core and the default mgmt services).
  --ignore              Ignore errors. Continue with given operation even if errors occur.
  --verbose             Display detailed information.

DCLI

Check vSphere cluster compatibility.

dcli com vmware vcenter namespacemanagement clustercompatibility list

dcli com vmware vcenter namespacemanagement supportbundle

List/delete integrated Harbor registry

It can happen that disabling/deleting the Integrated Harbor Registry ends in an endless loop and the vSphere objects weren't deleted. In this case, the vCenter dcli can help.

  • List the iHarbor instance:

dcli +show-unreleased +skip +username administrator@jarvis.tanzu +password 'PASSWORD' com vmware vcenter content registries harbor list

  • delete the iHarbor instance using dcli

dcli +show-unreleased +skip +username administrator@jarvis.tanzu +password 'PASSWORD' com vmware vcenter content registries harbor delete --registry Harbor-57b478ea...

SSH to Supervisor Control Plane Node

ssh to your vCenter Server Appliance and execut shell when prompted.

# change into the dir where the pwd decryption file is stored
cd /usr/lib/vmware-wcp/

# execute the file to get the pwd
./decryptK8Pwd.py

Connect to the Supervisor Node using the decrypted password:

Connected to PSQL

Cluster: domain-c7:a4b59b96-9e94-4158-95ff-e1dfeb58c6bf
IP: 10.10.13.30
PWD: DHinqsmNdedRcpMEOvlYtfSROSO6C6Bf4VC9mpzIE6QclX2vQqtJuG4O12FDp4Ox3LVxhYTB81Ntv2DM+g4daJTbBdJKw3nTcO/cFSsH7xCZayQNzWwVMdNYpNo5OSS4Gcnex4nGsauJO5HNz+a+AzJjfkICLLRLwCJHxi3tssk=

Operating with the wcp privileged wcp-vmop-user-domain-cXXXX user

In TKGs, you don't have permissions to e.g. configure, or even power-on, power-off the SupervisorVMs or the TKC-VMs. Not even the administrator@vsphere.local can. Even though, there's a way to use the wcp-vmop-user-domain user in order to execute those operations.

  1. Login into a SupervisorVM (see description above)
  2. execute the foloowing commands:
$ kubectl get secrets wcp-vmop-sa-vc-auth -n vmware-system-vmop -o jsonpath='{.data}' |jq

{
  "password": "PTJ0VC06J15GM01tOEcwfmNd=",
  "username": "d2NwLXZtb3AtdXNlci1kb21haW4tYzIwMzMtN2FjODUzYzItNTIwNy00Y2I0LThhNTItZjQwOGRjNGQ2M2Q4QG1hcms1MC5s="
}
  1. Decrypt the base64 encoded data to get the username and password
$ echo d2NwLXZtb3AtdXNlci1kb21haW4tYzIwMzMtN2FjODUzYzItNTIwNy00Y2I0LThhNTItZjQwOGRjNGQ2M2Q4QG1hcms1MC5sYWI=d2NwLXZtb3AtdXNlci1kb21haW4tYzIwMzMtN2FjODUzYzItNTIwNy00Y2I0LThhNTItZjQwOGRjNGQ2M2Q4QG1hcms1MC5s= | base64 -d

$ echo PTJ0VC06J15GM01tOEcwfmNd= | base64 -d
  1. Login into vSphere using the new credentials, which allows you to execute operations on SupervisorVMs as well as on TKC's

wcp-vmop-sa-vc

SSH into a Tanzu Kubernetes Cluster Nodes using the tkc-ssh-secret

Reference: Docs

  1. Login to the Supervisor-Cluster

kubectl vsphere login --insecure-skip-tls-verify --vsphere-username administrator@mark50.lab --server=mark50.jarvis.tanzu

  1. Get the exact secret-name out of the appropriate vSphere-Namespace
kubectl -n mark50-ns-1 get secrets

NAME                               TYPE                                  DATA   AGE
default-token-j77xs                kubernetes.io/service-account-token   3      121d
mark50-tkc-1-antrea                kubernetes.io/tls                     3      10d
mark50-tkc-1-auth-svc-cert         kubernetes.io/tls                     3      10d
mark50-tkc-1-ca                    Opaque                                2      10d
mark50-tkc-1-ccm-token-6dpjd       kubernetes.io/service-account-token   3      10d
mark50-tkc-1-control-plane-5mmc9   cluster.x-k8s.io/secret               1      10d
mark50-tkc-1-control-plane-jv8xz   cluster.x-k8s.io/secret               1      10d
mark50-tkc-1-control-plane-nnnpp   cluster.x-k8s.io/secret               1      10d
mark50-tkc-1-encryption            Opaque                                1      10d
mark50-tkc-1-etcd                  Opaque                                2      10d
mark50-tkc-1-extensions-ca         kubernetes.io/tls                     3      10d
mark50-tkc-1-kubeconfig            Opaque                                1      10d
mark50-tkc-1-metrics-server-cert   kubernetes.io/tls                     3      10d
mark50-tkc-1-node-zljq9-n8czc      cluster.x-k8s.io/secret               1      122m
mark50-tkc-1-proxy                 Opaque                                2      10d
mark50-tkc-1-pvcsi-token-t7xbb     kubernetes.io/service-account-token   3      10d
mark50-tkc-1-sa                    Opaque                                2      10d
mark50-tkc-1-ssh                   kubernetes.io/ssh-auth                1      10d
mark50-tkc-1-ssh-password          Opaque                                1      10d
mark50-tkc-1-worker-j4v9z-sst4j    cluster.x-k8s.io/secret               1      12s
  1. Get the base64 encoded ssh password for the TKC nodes
k -n mark50-ns-1 get secret mark50-tkc-1-ssh-password -o yaml

apiVersion: v1
data:
  ssh-passwordkey: aTcxWG9rbWVpdkllU01xaEJqRFkzTzVnTUZnanZkRXlmL090Q296cFUy=
kind: Secret
metadata:
  creationTimestamp: "2022-04-29T07:18:50Z"
  name: mark50-tkc-1-ssh-password
  namespace: mark50-ns-1
  ownerReferences:
  - apiVersion: run.tanzu.vmware.com/v1alpha2
    kind: TanzuKubernetesCluster
    name: mark50-tkc-1
    uid: 69cac0d0-fd81-4aa6-8f80-011663b8ecb9
  resourceVersion: "89540146"
  selfLink: /api/v1/namespaces/mark50-ns-1/secrets/mark50-tkc-1-ssh-password
  uid: e61530b9-60d1-4475-8301-ccae21028e8d
type: Opaque
  1. Decode the ssh-passwordkey

echo aTcxWG9rbWVpdkllU01xaEJqRFkzTzVnTUZnanZkRXlmL090Q296cFUy= | base64 -d

  1. ssh into the desired TKC node using e.g. the IP (k -n mark50-ns-1 get virtualmachine -o wide) and the vmware-system-user

SSH into a Tanzu Kubernetes Cluster Nodes using a Jumpbox-Pod

  1. Connect to the Supervisor Cluster.

kubectl vsphere login --insecure-skip-tls-verify --vsphere-username administrator@mark50.lab --server=mark50.jarvis.tanzu

  1. Specify target TKC Namespace and switch context:
export NAMESPACE=

kubectl config use-context $NAMESPACE
  1. View the TKGS-CLUSTER-NAME-ssh secret object:

kubectl get secrets

  1. Create a vSphere Pod using the following jumpbox.yaml

Replace the namespace value YOUR-NAMESPACE with the vSphere Namespace where the target cluster is provisioned. Replace the secretName value YOUR-CLUSTER-NAME-ssh with name of the target cluster.

kubectl create -f - <<EOF
apiVersion: v1
kind: Pod
metadata:
  name: jumpbox
  namespace: YOUR-NAMESPACE     #REPLACE
spec:
  containers:
  - image: "photon:3.0"
    name: jumpbox
    command: [ "/bin/bash", "-c", "--" ]
    args: [ "yum install -y openssh-server; mkdir /root/.ssh; cp /root/ssh/ssh-privatekey /root/.ssh/id_rsa; chmod 600 /root/.ssh/id_rsa; while true; do sleep 30; done;" ]
    volumeMounts:
      - mountPath: "/root/ssh"
        name: ssh-key
        readOnly: true
    resources:
      requests:
        memory: 2Gi
  volumes:
    - name: ssh-key
      secret:
        secretName: YOUR-CLUSTER-NAME-ssh     #REPLACE
EOF
  1. Create an environment variable with the IP address of the target cluster node by running the following set of commands:

kubectl get virtualmachines

export VMNAME=

export VMIP=$(kubectl -n $NAMESPACE get virtualmachine/$VMNAME -o jsonpath='{.status.vmIp}')

kubectl exec -it jumpbox /usr/bin/ssh vmware-system-user@$VMIP

Jumpbox

Docker

Install Docker Engine on Ubuntu

  • Uninstall old versions: sudo apt-get remove docker docker-engine docker.io containerd runc
  • Set up the repository:
sudo apt-get update

sudo apt-get install \
    ca-certificates \
    curl \
    gnupg \
    lsb-release
  • Add Docker’s official GPG key:
curl -fsSL https://download.docker.com/linux/ubuntu/gpg | sudo gpg --dearmor -o /usr/share/keyrings/docker-archive-keyring.gpg
  • Install Docker Engine
sudo apt-get update

sudo apt-get install docker-ce docker-ce-cli containerd.io
  • Manage Docker as a non-root user
sudo groupadd docker

sudo usermod -aG docker $USER

Install Docker Engine on RHEL

  • Uninstall old versions:
sudo yum remove docker \
                  docker-client \
                  docker-client-latest \
                  docker-common \
                  docker-latest \
                  docker-latest-logrotate \
                  docker-logrotate \
                  docker-engine \
                  podman \
                  runc

or

sudo yum remove docker-ce docker-ce-cli containerd.io

sudo rm -rf /var/lib/docker

sudo rm -rf /var/lib/containerd
  • Install using the repository
sudo yum install -y yum-utils

sudo yum-config-manager \
    --add-repo \
    https://download.docker.com/linux/rhel/docker-ce.repo
  • Install Docker Engine
sudo yum install docker-ce docker-ce-cli containerd.io
  • Install a specific version of Docker Engine
yum list docker-ce --showduplicates | sort -r
  • Start Docker
sudo systemctl start docker
  • Manage Docker as a non-root user
sudo groupadd docker

sudo usermod -aG docker $USER

Proxy

General tips:

Special character handling:

Literal backslash characters (\) need to be doubled escape them as shown below.

# export http_proxy=http://DOMAIN\\USERNAME:PASSWORD@SERVER:PORT/
When the username or password uses the @ symbol, add a backslash (\) before the @ – for example:

# export http_proxy=http://DOMAIN\\USERN\@ME:PASSWORD@SERVER:PORT
or

# export http_proxy=http://DOMAIN\\USERNAME:P\@SSWORD@SERVER:PORT

NO_PROXY:

Configure NO_PROXY to ensures that traffic destined to internal addresses won’t get forwarded to the proxy.

PhotonOS

PhotonOS Wiki

Proxy configuration for VMware Photon OS. There are multiplie places in which a proxy can be defined, including in the Kubernetes configuration, or specifically for the tdnf package manager.

vim /etc/sysconfig/proxy

tdnf is using HTTPS as a default!

Ubuntu

Temporary:

Check if Proxy settings are set:

env | grep proxy

Set settings temporary (settings belonging to one shell session!):

export HTTP_PROXY=http://<user>:<pass>@<proxy>:<port>/ export HTTPS_PROXY=http://<user>:<pass>@<proxy>:<port>/ export NO_PROXY=localhost,127.0.0.1,::1

Without user: export HTTP_PROXY=http://SERVER:PORT/

Permanent for All Users:

sudo vi /etc/environment

Update the file with the same information listed above.

Setting Up Proxy for apt

sudo vi /etc/apt/apt.conf

Acquire::http::Proxy "http://[username]:[password]@ [proxy-web-or-IP-address]:[port-number]";
Acquire::https::Proxy "http://[username]:[password]@ [proxy-web-or-IP-address]:[port-number]";

CentOS/RHEL

Check if Proxy settings are set:

echo $http_proxy

Temporary

Without user: export http_proxy=http://SERVER:PORT/

With user: export http_proxy=http://USERNAME:PASSWORD@SERVER:PORT/

With a Domain user: export http_proxy=http://DOMAIN\\USERNAME:PASSWORD@SERVER:PORT/

Permanent:

echo "http_proxy=http://proxy.example.com:3128/" > /etc/environment

Note that unlike a shell script in /etc/profile.d described in the next section, the /etc/environment file is NOT a shell script and applies to all processes without a shell. Source: How to Configure Proxy in CentOS/RHEL/Fedora

Configuring proxy for processes with SHELL

For bash and sh users, add the export line given above into a new file called /etc/profile.d/http_proxy.sh file:

echo "export http_proxy=http://proxy.example.com:3128/" > /etc/profile.d/http_proxy.sh

Setting Up Proxy for yum

vi /etc/yum.conf

proxy=http://proxy.example.com:3128
proxy_username=yum-user
proxy_password=qwerty

Demo App

Hackazon Shop App with Ingress

kubectl -n hackazon apply -f - <<EOF
kind: Deployment
apiVersion: apps/v1
metadata:
  name: hackazon-shop
  labels:
    app: hackazon-shop
spec:
  replicas: 2
  selector:
    matchLabels:
      app: hackazon-shop
  template:
    metadata:
      labels:
        app: hackazon-shop
    spec:
      containers:
      - name: hackazon-shop
        image: projects.registry.vmware.com/tanzu_ese_poc/hackazon:1.0
        ports:
        - name: http
          containerPort: 80
          protocol: TCP
---
kind: Service
apiVersion: v1
metadata:
  name: hackazon-svc-lb
  labels:
    app: hackazon-shop
    svc: hackazon-svc-lb
spec:
  ports:
  - name: http
    port: 80
    targetPort: 80
  selector:
    app: hackazon-shop
  type: LoadBalancer
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
  name: hackazon-ingress
  labels:
    app: hackazon-shop
#  annotations:
#   kubernetes.io/ingress.class: avi
spec:
  rules:
  - host: hackazon.avi.jarvis.tanzu
    http:
      paths:
      - pathType: Prefix
        path: /
        backend:
          service:
            name: hackazon-svc-lb
            port:
              number: 80
EOF

Ghost Web-App with pvc:

kubectl -n ghost create -f - << EOF
apiVersion: apps/v1
kind: Deployment
metadata:
  name: blog
  namespace: ghost
  labels:
    app: blog
spec:
  replicas: 1
  selector:
    matchLabels:
      app: blog
  template:
    metadata:
      labels:
        app: blog
    spec:
      containers:
      - name: blog
        image: projects.registry.vmware.com/tanzu_ese_poc/ghost:4.34.0
        imagePullPolicy: IfNotPresent
        ports:
        - containerPort: 2368
        env:
        - name: url
          value: http://my-blog.corp.local
        volumeMounts:
        - mountPath: /var/lib/ghost/content
          name: content
      volumes:
      - name: content
        persistentVolumeClaim:
          claimName: blog-content
---
apiVersion: v1
kind: Service
metadata:
  labels:
    name: blog
  name: blog
  namespace: ghost
spec:
  ports:
    - port: 80
      targetPort: 2368
  selector:
    app: blog
  type: LoadBalancer
---
kind: PersistentVolumeClaim
apiVersion: v1
metadata:
  name: blog-content
  namespace: ghost
spec:
  accessModes:
    - ReadWriteOnce
  storageClassName: local-path
  resources:
    requests:
      storage: 5Gi
EOF

Helpful Resources