Skip to content

adetalhouet/ocp-ztp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Deploy a OpenShift on libvirt using RHACM ZTP capabilities

The goal is to leverage the latest capabilities from Red Hat Advanced Cluster Management (RHACM) 2.3+ to deploy an OpenShift cluster using the Zero Touch Provisioning on an emulated bare metal environment.

The typical Zero Touch Provisioning flow is meant to work for bare metal environment; but if like me, you don't have a bare metal environment handy, or want to optimize the only server you have, make sure to review the section "Ironic & Metal3".

RHACM works in a hub and spoke manner. So the goal here is to deploy a spoke from the hub cluster.

The overall setup uses the following components:

If you do not have a baremetal cluster, you also need to deploy:

  • Ironic: It is the OpenStack bare metal provisioning tool that uses PXE or BMC to provision and turn on/off machines
  • Metal3: It is the Kubernetes bare metal provisioning tool. Under the hood, it uses Ironic. And above the hood, it provides an operator along with the CRD it supports: BareMetalHost

Let's align on the Zero Touch Provisioning expectation:

  • the overall libvirt environment will be setup manually (although it could easily be automated).
  • once the environment is correctly setup, we will apply the manifests that will automate the spoke cluster creation.

Table of Content

  1. Pre-requisites
  2. Architecture
  3. Install requirements on the hub cluster
  4. Install requirements on the spoke server
  5. Let's deploy the spoke

Pre-requisite

  • Red Hat OpenShift Container Platform 4.8+ for the hub cluster- see here on how to deploy
  • Red Hat Advanced Cluster Management 2.3+ installed on the hub cluster- see here on how to deploy
  • A server with at least 32GB of RAM, 8 CPUs and 120 GB of disk - this is the machine we will use for the spoke. Mine is setup with CentOS 8.4
  • Clone the git repo: git clone https://github.com/adetalhouet/ocp-gitops

Architecture

Architecture

Requirements on the hub cluster

The assumption is the cluster is not deployed on bare metal. If that's the case skip the Ironic and Metal3 portion.

In my case, my hub cluster is deployed in AWS. As it isn't a bare metal cluster, you don't have the Ironic and Metal3 pieces, so we will deploy them ourselves.

Install the Assisted Service

The related manifest for the install are located in the hub folder. The main manifest is 02-assistedserviceconfig.yaml specifying the AgentServiceConfig definition, which defines the base RHCOS image to use for the server installation.

We also create a ClusterImageSet to refer to OpenShift 4.8 version. This will be referenced by the spoke manifest to define what version of OpenShift to install.

Add your private key in the hub/03-assisted-deployment-ssh-private-key.yaml file (use the example), and then apply the folder. The private key will be in the resulting VM, and you will use the corresponding public key to ssh, if needed.

Follow the documention to enable Central Infrastructure Management service.

oc patch hiveconfig hive --type merge -p '{"spec":{"targetNamespace":"hive","logLevel":"debug","featureGates":{"custom":{"enabled":["AlphaAgentInstallStrategy"]},"featureSet":"Custom"}}}'
oc patch provisioning provisioning-configuration --type merge -p '{"spec":{"watchAllNamespaces": true }}'
oc patch provisioning provisioning-configuration --type merge -p '{"spec":{"disableVirtualMediaTLS": true }}'

Everything will be installed in the open-cluster-management namespace.

$ oc apply -k hub
configmap "assisted-service-config" deleted
secret "assisted-deployment-ssh-private-key" deleted
agentserviceconfig.agent-install.openshift.io "agent" deleted
clusterimageset.hive.openshift.io "openshift-v4.8.0" deleted

After view second, check the assisted service has been created

$ oc get pod -n open-cluster-management -l app=assisted-service

Install Ironic and Metal3

The assumption is the cluster is not deployed on bare metal. If that's the case skip this section.

Both Ironic and Metal3 can be installed from the baremetal-operator but experience has proven it is a very opinionated install, and out of the box, doesn't work in my environment, and probably will not work in yours.

So, I pulled all the manifests required for the install, and put them in the metal-provisioner folder.

As Ironic will be the component instructing the remote server to download the ISO, it needs to be configured properly so the remote server can reach back to the underlying Ironic's HTTP server.

The 02-ironic.yaml manifest provides a Service and Route to expose the various services it provides. And it also contains a ConfigMap called ironic-bmo-configmap containing all the configuration bits required for Ironic to work properly. These elements points to my environment, so you need to customize them accordingly, by adjusting the $CLUSTER_NAME.$DOMAIN_NAME in the Route definition and in the ironic-bmo-configmap ConfigMap.

In my case $CLUSTER_NAME.$DOMAIN_NAME = hub-adetalhouet.rhlteco.io

Here is a command to help make that change; make sure to replace $CLUSTER_NAME.$DOMAIN_NAME with yours. If you're on a mac, using gsed instead of sed to use the GNU sed binary.

$ sed -i "s/hub-adetalhouet.rhtelco.io/$CLUSTER_NAME.$DOMAIN_NAME/g" metal-provisioner/02-ironic.yaml

Based on the upstream Ironic image, I had to adjust the start command of the ironic-api and ironic-conductor containers to alter their ironic.conf configuration so it would consume the exposed Route rather than the internal IP. When Ironic using the BMC to configure the server, it will instruct the server to load the boot ISO image from its http server; the Ironic http server must be reachable from the spoke server. In my case, given the hub and the spoke only share public internet as a common network, I had to expose Ironic http server. If you have a private network, the setup would work the same.

In both of these containers, the /etc/ironic/ironic.conf configuration is created at runtime, based on the Jinja template /etc/ironic/ironic.conf.j2; so I modify the template to have the resulting generated config as expected.

$ sed -i "s/{{ env.IRONIC_URL_HOST }}:{{ env.HTTP_PORT }}/{{ env.IRONIC_HTTP_URL }}/g" /etc/ironic/ironic.conf.j2
$ sed -i "s/host = {{ env.IRONIC_URL_HOST }}/host = {{ env.IRONIC_HTTP_URL }}/g" /etc/ironic/ironic.conf.j2

Finally, Ironic uses host network (although not required in our case), so I have granted the metal-provisioner ServiceAccount privileged SCC. And in the ironic-bmo-configmap you need to update the PROVISIONING_INTERFACE to reflect your node interface. This is stupid, because we don't care about this at all in our case, but Ironic will basically take the IP from this interface and use it at many places. Actually, some of the place where it uses the host ip are the places where we made the change in the ironic.conf in the previous section.

Keep in mind, the initial intention of this bare-metal-operator is to work in a BareMetal environment, where it is assumed the PROVISIONING_INTERFACE is on a network that can reach the nodes you would want to either add in the cluster, or provisioned with OpenShift using the ZTP flow.

Have a review of the manifest, and when confident, apply them

$ oc apply -k metal-provisioner
namespace/metal-provisioner created
serviceaccount/metal-provisioner created
clusterrole.rbac.authorization.k8s.io/baremetalhost-role created
clusterrole.rbac.authorization.k8s.io/ironic-scc created
clusterrolebinding.rbac.authorization.k8s.io/baremetalhost-rolebinding created
clusterrolebinding.rbac.authorization.k8s.io/ironic-rolebinding created
configmap/baremetal-operator-ironic created
configmap/ironic-bmo-configmap created
configmap/ironic-htpasswd created
configmap/ironic-inspector-htpasswd created
secret/ironic-auth-config created
secret/ironic-credentials created
secret/ironic-inspector-auth-config created
secret/ironic-inspector-credentials created
secret/ironic-rpc-auth-config created
secret/mariadb-password created
service/ironic created
deployment.apps/baremetal-operator-controller-manager created
deployment.apps/capm3-ironic created
route.route.openshift.io/ironic-api created
route.route.openshift.io/ironic-http created
route.route.openshift.io/ironic-inspector created

Here is an output of what you should expect

oc get all -n metal-provisioner
$ oc get all -n metal-provisioner

NAME                                                         READY   STATUS    RESTARTS   AGE
pod/baremetal-operator-controller-manager-7477d5cd57-2cbmj   2/2     Running   0          20m
pod/capm3-ironic-6cc84ff99c-l5bpt                            5/5     Running   0          20m

NAME             TYPE        CLUSTER-IP    EXTERNAL-IP   PORT(S)                    AGE
service/ironic   ClusterIP   172.30.59.7   <none>        5050/TCP,6385/TCP,80/TCP   20m

NAME                                                    READY   UP-TO-DATE   AVAILABLE   AGE
deployment.apps/baremetal-operator-controller-manager   1/1     1            1           20m
deployment.apps/capm3-ironic                            1/1     1            1           20m

NAME                                                               DESIRED   CURRENT   READY   AGE
replicaset.apps/baremetal-operator-controller-manager-7477d5cd57   1         1         1       20m
replicaset.apps/capm3-ironic-6cc84ff99c                            1         1         1       20m

NAME                                        HOST/PORT                                                            PATH   SERVICES   PORT        TERMINATION   WILDCARD
route.route.openshift.io/ironic-api         ironic-api-metal-provisioner.apps.hub-adetalhouet.rhtelco.io                ironic     api                       None
route.route.openshift.io/ironic-http        ironic-http-metal-provisioner.apps.hub-adetalhouet.rhtelco.io               ironic     httpd                     None
route.route.openshift.io/ironic-inspector   ironic-inspector-metal-provisioner.apps.hub-adetalhouet.rhtelco.io          ironic     inspector                 None

Requirements on the spoke server

I'm assuming you have a blank server, running CentOS 8.4, and login as root.

Install libvirt

Install the required dependencies.

dnf install -y bind-utils libguestfs-tools cloud-init
dnf module install virt -y
dnf install virt-install -y
systemctl enable libvirtd --now

Install and configure Sushy service

Sushy service is a Virtual Redfish BMC emulator for libvirt or OpenStack virtualization. In our case, we will use it for libvirt, in order to add BMC capabilities to libvirt domain. That will enable remote control of the VMs.

dnf install python3 -y
pip3 install sushy-tools

Then you need to configure the service. In my case, I'm binding the sushy service on all my interfaces. But if you have a management interface providing connectivity between the hub and the spoke environments, you should use that interface instead. Also, the port is customizable, and if you have firewall in the way, make sure to open them accordingly.

echo "SUSHY_EMULATOR_LISTEN_IP = u'0.0.0.0'
SUSHY_EMULATOR_LISTEN_PORT = 8000
SUSHY_EMULATOR_SSL_CERT = None
SUSHY_EMULATOR_SSL_KEY = None
SUSHY_EMULATOR_OS_CLOUD = None
SUSHY_EMULATOR_LIBVIRT_URI = u'qemu:///system'
SUSHY_EMULATOR_IGNORE_BOOT_DEVICE = True
# This specifies where to find the boot loader for a UEFI boot. This is what the ZTP process uses.
SUSHY_EMULATOR_BOOT_LOADER_MAP = {
    u'UEFI': {
        u'x86_64': u'/usr/share/OVMF/OVMF_CODE.secboot.fd'
    },
    u'Legacy': {
        u'x86_64': None
    }
}" > /etc/sushy.conf

There is currently an issue with libvirt that basically forces the use of secure boot. Theoritically this can be disabled, but the feature isn't working properly since RHEL 8.3 (so it's the same in CentOS that I'm using). In order to mask the secure feature boot vars, to allow a non-secure boot, the following solution has been suggested:

mkdir -p /etc/qemu/firmware
touch /etc/qemu/firmware/40-edk2-ovmf-sb.json

Now, let's create the sushy service and start it.

echo '[Unit]
Description=Sushy Libvirt emulator
After=syslog.target

[Service]
Type=simple
ExecStart=/usr/local/bin/sushy-emulator --config /etc/sushy.conf
StandardOutput=syslog
StandardError=syslog

[Install]
WantedBy=multi-user.target' > /usr/lib/systemd/system/sushy.service
systemctl start sushy
systemctl enable --now sushy.service

Finally, let's start the built-in firewall and allow traffic on port 8000.

systemctl start firewalld
firewall-cmd --add-port=8000/tcp --permanent
firewall-cmd --add-port=8000/tcp --zone libvirt --permanent

Libvirt setup

Create a pool

When Ironic will use our virtual BMC, emulated by sushy-tools, to load the ISO in the server (VM in our case), sushy-tools will host that image in the default storage pool, so we need to create it accordingly. (I couldn't find a way, yet, to configure the storage pool to use.)

$ mkdir -p /var/lib/libvirt/sno-ztp
$ virsh pool-define-as default --type dir --target /var/lib/libvirt/sno-ztp
$ virsh pool-start default
$ virsh pool-autostart default

$ virsh pool-list
 Name      State    Autostart
-------------------------------
 boot      active   yes
 default   active   yes
 images    active   yes

Create a network

OpenShift Bare Metal install has the following requirements:

  • a proper hostname / domain name mapped to the MAC address of the interface to use for the provisioning
  • a DNS entry for the api.$clusterName.$domainName
  • a DNS entry for the *.apps.$clusterName.$domainName

So we will configure them accordingly in the libvirt network definition, using the built-in dnsmaq capability of libvirt network.

Here is my network definition (libvirt/sno/net.xml)

libvirt/sno/net.xml
<network xmlns:dnsmasq="http://libvirt.org/schemas/network/dnsmasq/1.0">
  <name>sno</name>
  <forward mode='nat'>
    <nat>
      <port start='1024' end='65535'/>
    </nat>
  </forward>
  <bridge name='virbr1' stp='on' delay='0'/>
  <ip address='192.168.123.1' netmask='255.255.255.0'>
    <dhcp>
      <range start='192.168.123.2' end='192.168.123.254'/>
      <host mac="02:04:00:00:00:66" name="sno.lab.adetalhouet" ip="192.168.123.5"/>
    </dhcp>
  </ip>
  <dns>
    <host ip="192.168.123.5"><hostname>api.sno.lab.adetalhouet</hostname></host>
  </dns>
  <dnsmasq:options>
    <dnsmasq:option value="auth-server=sno.lab.adetalhouet,"/>
    <dnsmasq:option value="auth-zone=sno.lab.adetalhouet"/>
    <dnsmasq:option value="host-record=lb.sno.lab.adetalhouet,192.168.123.5"/>
    <dnsmasq:option value="cname=*.apps.sno.lab.adetalhouet,lb.sno.lab.adetalhouet"/>
  </dnsmasq:options>
</network>

Now let's define and start our network

# create the file net.xml with the content above
$ virsh net-define net.xml
$ virsh net-start sno
$ virsh net-autostart sno

$ virsh net-list
 Name      State    Autostart   Persistent
--------------------------------------------
 default   active   yes         yes
 sno       active   no          yes

Create the disk

In order for Assisted Installer to allow the installation of the Single Node OpenShift to happen, one of the requirement is the disk size: it must be at least of 120GB. When creating a disk of 120GB, or even 150GB, for some reason I had issues and the Assisted Service wouldn't allow the installation complaining about the disk size requirepement not being met. So let's create a disk of 200 GB to be sure.

$ qemu-img create -f qcow2 /var/lib/libvirt/sno-ztp/sno.qcow2 200G
Formatting '/var/lib/libvirt/sno-ztp/sno.qcow2', fmt=qcow2 size=214748364800 cluster_size=65536 lazy_refcounts=off refcount_bits=16

Create the VM / libvirt domain

While creating the VM, make sure to adjust RAM and CPU, as well as the network and disk if you've made modification. The interface configured in the domain is the one we pre-defined in the network definition, and we will identify the interface by its mac address. When the VM will boot, it will be able to resolve its hostname through the DNS entry.

(FYI - I spent hours trying to nail down the proper xml definition, more importantly the os bits. When the Assisted Asservice will start the provisioning, it will first start the VM, load the discovery.iso and then restart the VM to boot from the newly added disc. After the restart, the os section will be modified, as Assisted Service will configure an UEFI boot.)

Here is my VM definition (libvirt/sno/vm.xml)

libvirt/sno/vm.xml
<domain type="kvm">
  <name>sno</name>
  <uuid>b6c92bbb-1e87-4972-b17a-12def3948890</uuid>
  <metadata>
    <libosinfo:libosinfo xmlns:libosinfo="http://libosinfo.org/xmlns/libvirt/domain/1.0">
      <libosinfo:os id="http://fedoraproject.org/coreos/stable"/>
    </libosinfo:libosinfo>
  </metadata>
  <memory>33554432</memory>
  <currentMemory>33554432</currentMemory>
  <vcpu>16</vcpu>
  <os>
    <type arch="x86_64" machine="q35">hvm</type>
    <boot dev="hd"/>
    <boot dev="cdrom"/>
  </os>
  <features>
    <acpi/>
    <apic/>
    <smm state='on'/>
    <vmport state="off"/>
  </features>
  <cpu mode="host-passthrough">
    <cache mode="passthrough"/>
  </cpu>
  <clock offset="utc">
    <timer name="rtc" tickpolicy="catchup"/>
    <timer name="pit" tickpolicy="delay"/>
    <timer name="hpet" present="no"/>
  </clock>
  <pm>
    <suspend-to-mem enabled="no"/>
    <suspend-to-disk enabled="no"/>
  </pm>
  <devices>
    <emulator>/usr/libexec/qemu-kvm</emulator>
    <disk type="file" device="disk">
      <driver name="qemu" type="qcow2"/>
      <source file="/var/lib/libvirt/sno-sno.qcow2"/>
      <target dev="vda" bus="virtio"/>
    </disk>
    <controller type="usb" index="0" model="qemu-xhci" ports="15"/>
    <interface type="network">
      <source network="sno"/>
      <mac address="02:04:00:00:00:66"/>
      <model type="virtio"/>
    </interface>
    <console type="pty"/>
    <channel type="unix">
      <source mode="bind"/>
      <target type="virtio" name="org.qemu.guest_agent.0"/>
    </channel>
    <channel type="spicevmc">
      <target type="virtio" name="com.redhat.spice.0"/>
    </channel>
    <input type="tablet" bus="usb"/>
    <graphics type="spice" port="-1" tlsPort="-1" autoport="yes">
      <image compression="off"/>
    </graphics>
    <sound model="ich9"/>
    <video>
      <model type="qxl"/>
    </video>
    <redirdev bus="usb" type="spicevmc"/>
    <redirdev bus="usb" type="spicevmc"/>
    <memballoon model="virtio"/>
    <rng model="virtio">
      <backend model="random">/dev/urandom</backend>
    </rng>
  </devices>
</domain>
Now let's define our domain.
# create the file vm.xml with the content above
virsh define vm.xml
virsh autostart sno

Do not start the VM by yourself, it will be done later in the process, automatically. Moreover, your VM at this point has no CDROM to boot from.

If you have a bridge network, you can add an additional interface to the domain definition. Please see libvirt/sno/vm-bridge-net.xml along with libvirt/sno/bridge-interface.md. And see spoke-ztp/03-nmstateconfig.yaml on how to configure the interface within the resulting VM.

Now the environment is ready, let's create an Single Node OpenShift cluster automagically.

Let's deploy the spoke

We will use all the manifests in the spoke-ztp/ folder. Simply apply the following command:

$ oc apply -k spoke-ztp/
namespace/sno-ztp created
secret/assisted-deployment-pull-secret created
secret/sno-secret created
infraenv.agent-install.openshift.io/sno-ztp-infraenv created
nmstateconfig.agent-install.openshift.io/lab-spoke-adetalhouet created
klusterletaddonconfig.agent.open-cluster-management.io/lab-spoke-adetalhouet created
managedcluster.cluster.open-cluster-management.io/lab-spoke-adetalhouet created
agentclusterinstall.extensions.hive.openshift.io/sno-ztp-clusteragent created
clusterdeployment.hive.openshift.io/sno-ztp-cluster created
baremetalhost.metal3.io/sno-ztp-bmh created

It will take on average 60-ish minutes for the cluster to be ready. That said, to validate the cluster will get deployed properly, few tests you can do.

Let's review the manifests:

  • 00-namespace.yaml creates the namespace where the configuration will be hosted.
  • 01-agentclusterinstall.yaml defines the AgentClusterInstall is responsible for the overall cluster configuration. This is where you specify:
    • the network requirements (clusterNetwork, serviceNetwork, machineNetwork).
    • the OpenShift version to use, by refering to the ClusterImageSet name we created earlier.
    • the overall cluster setup, i.e. how many control and worker node you want. In our case, we deploy a SNO, so only 1 control node.
    • the pub key that goes with the private key setup earlier in the assisted-deployment-ssh-private-key secret
  • 02-clusterdeployment.yaml defines the ClusterDeployment
    • it references the AgentClusterIntall and define the pull-secret to use for the cluster provisioning.
    • this is where you define the baseDomain and the clusterName to use for the spoke cluster
  • 03-nmstateconfig.yaml is required if you are using a bridge network and want to set a static ip. See here along with their doc for more information / use cases.
  • 04-spokeinfraenv.yaml defines the InfraEnv. It is basically a when to customize the intial cluster setup. If you want to add/modify some files for the ignition process, you can. If you want to configure additional networking bits, this is where you can do it as well. Refer to the doc, and here is an example.
  • 05-baremetalhost.yaml defines the BareMetalHost. This is where you provide the information on:
    • how to connect to the server through its BMC
    • the MAC address of the provisioning interface
  • 05-userdata.yaml is an attempt at providing additional information but libvirt isn't liking the way that disk is provided (I didn't dig into this, so here is the error I got if you want to dig into this)
  • 06-assisteddeploymentpullsecret.yaml is the pull-secret to download the images in the spoke cluster.
  • 07-kusterlet.yaml setup the cluster to be imported within RHACM, and have the addon agents installed.
{"level":"info","ts":1628007509.626545,"logger":"provisioner.ironic","msg":"current provision state","host":"sno-ztp~sno-ztp-bmh","lastError":"Deploy step deploy.deploy failed: Redfish exception occurred. Error: Setting power state to power on failed for node 1e87eede-7ddb-4da4-bb7f-2a12037ac323. Error: HTTP POST http://148.251.12.17:8000/redfish/v1/Systems/b6c92bbb-1e87-4972-b17a-12def3948891/Actions/ComputerSystem.Reset returned code 500. Base.1.0.GeneralError: Error changing power state at libvirt URI \"qemu:///system\": internal error: qemu unexpectedly closed the monitor: 2021-08-03T16:18:17.110846Z qemu-kvm: -device isa-fdc,bootindexA=1: Device isa-fdc is not supported with machine type pc-q35-rhel8.2.0 Extended information: [{'@odata.type': '/redfish/v1/$metadata#Message.1.0.0.Message', 'MessageId': 'Base.1.0.GeneralError'}].","current":"deploy failed","target":"active"}
  • 04-assisteddeploymentpullsecret.yaml defines your pull-secret
  • 05-kusterlet.yaml tells RHACM to add this cluster as a managed cluster, and deploy the various addon agents on it.

Few debugging tips

Storage

First, look at the storage pool folder; sometimes ISO upload isn't working properly, and the resulting ISO doesn't have all the data. See the size of both ISO; the expected size, based on my experience, is 107884544. So if the file is not in that range, the VM won't boot properly.

[root@lab sno]# ls -ls /var/lib/libvirt/sno-ztp/
total 25639776
       4 -rw------- 1 qemu qemu        3265 Jul 29 04:04 boot-072b3441-2bd9-4aaf-939c-7e4640e38935-iso-79ad9824-8110-4570-83e3-a8cd6ae9d435.img
  105356 -rw------- 1 qemu qemu   107884544 Aug  1 20:48 boot-961b4d9e-1766-4e38-8a6d-8de54c7a836b-iso-b6c92bbb-1e87-4972-b17a-12def3948890.img
25534416 -rw-r--r-- 1 root root 26146963456 Jul 30 14:36 sno2.qcow2
Network

After a few minutes of having your VM running, it should get its IP from the network. Two ways to validate: Check the network dhcp-leases and ensure the IP has been assigned

[root@lab sno]# virsh net-dhcp-leases sno
 Expiry Time           MAC address         Protocol   IP address         Hostname   Client ID or DUID
----------------------------------------------------------------------------------------------------------
 2021-08-01 22:05:02   02:04:00:00:00:66   ipv4       192.168.123.5/24   sno        01:02:04:00:00:00:66

If so, attempt to SSH using the public key that goes with the private key you configure while installing the Assisted Service. You should be able to ssh properly.

[root@lab sno]# ssh core@192.168.123.5
The authenticity of host '192.168.123.5 (192.168.123.5)' can't be established.
ECDSA key fingerprint is SHA256:N6wy/bQ5YeL01LsLci+IVztzRs8XFVeU4rYJIDGD8SU.
Are you sure you want to continue connecting (yes/no/[fingerprint])? yes
Warning: Permanently added '192.168.123.5' (ECDSA) to the list of known hosts.
Red Hat Enterprise Linux CoreOS 48.84.202107202156-0
  Part of OpenShift 4.8, RHCOS is a Kubernetes native operating system
  managed by the Machine Config Operator (`clusteroperator/machine-config`).

WARNING: Direct SSH access to machines is not recommended; instead,
make configuration changes via `machineconfig` objects:
  https://docs.openshift.com/container-platform/4.8/architecture/architecture-rhcos.html

---
[core@sno ~]$

Here are extra information to expect when you are using a bridge network. The IP of the bridged interface will be manually configured using manifest 03-nmstateconfig.yaml. If you have DHCP for that network, that would work as well. You can validate the created domain has both interface, and you can look for the IP of the private network, if you want to ssh/troubleshoot.

[root@lab sno-pub]# virsh domiflist 21
 Interface   Type      Source   Model    MAC
------------------------------------------------------------
 vnet19      bridge    br0      virtio   00:50:56:01:15:94
 vnet20      network   sno      virtio   52:54:00:73:90:59

[root@lab sno-pub]# virsh domifaddr 21
 Name       MAC address          Protocol     Address
-------------------------------------------------------------------------------
 vnet20     52:54:00:73:90:59    ipv4         192.168.123.248/24
Monitor the Agent

Assuming the above worked, then I suggest you monitor the Agent that was created for your cluster deployment. This will give you an URL you can use to follow the events occuring in your cluster.

The Agent is the bare metal installer agent, it will provide info regarding the bare metal install provisioning.

oc get Agent -n sno-ztp
$ oc get Agent -n sno-ztp
NAME                                   CLUSTER            APPROVED   ROLE     STAGE
b6c92bbb-1e87-4972-b17a-12def3948890   sno-ztp-cluster   true       master   Done

$ oc describe Agent b6c92bbb-1e87-4972-b17a-12def3948890 -n sno-ztp
Name:         b6c92bbb-1e87-4972-b17a-12def3948890
Namespace:    sno-ztp
Labels:       agent-install.openshift.io/bmh=sno-ztp-bmh
              infraenvs.agent-install.openshift.io=sno-ztp-infraenv
Annotations:  <none>
API Version:  agent-install.openshift.io/v1beta1
Kind:         Agent
Metadata:
  Creation Timestamp:  2021-08-01T18:52:44Z
  Finalizers:
    agent.agent-install.openshift.io/ai-deprovision
  Generation:  2
  Resource Version:  27926035
  UID:               a91ff59c-a08c-41ac-9603-1981bec69f70
Spec:
  Approved:  true
  Cluster Deployment Name:
    Name:       sno-ztp-cluster
    Namespace:  sno-ztp
  Role:
Status:
  Bootstrap:  true
  Conditions:
    Last Transition Time:  2021-08-01T18:52:44Z
    Message:               The Spec has been successfully applied
    Reason:                SyncOK
    Status:                True
    Type:                  SpecSynced
    Last Transition Time:  2021-08-01T18:52:44Z
    Message:               The agent's connection to the installation service is unimpaired
    Reason:                AgentIsConnected
    Status:                True
    Type:                  Connected
    Last Transition Time:  2021-08-01T18:52:52Z
    Message:               The agent installation stopped
    Reason:                AgentInstallationStopped
    Status:                True
    Type:                  RequirementsMet
    Last Transition Time:  2021-08-01T18:52:52Z
    Message:               The agent's validations are passing
    Reason:                ValidationsPassing
    Status:                True
    Type:                  Validated
    Last Transition Time:  2021-08-01T19:09:59Z
    Message:               The installation has completed: Done
    Reason:                InstallationCompleted
    Status:                True
    Type:                  Installed
  Debug Info:
    Events URL:  https://assisted-service-open-cluster-management.apps.hub-adetalhouet.rhtelco.io/api/assisted-install/v1/clusters/064d8242-e63a-4ba9-9eb4-aaaa773cdf32/events?api_key=eyJhbGciOiJFUzI1NiIsInR5cCI6IkpXVCJ9.eyJjbHVzdGVyX2lkIjoiMDY0ZDgyNDItZTYzYS00YmE5LTllYjQtYWFhYTc3M2NkZjMyIn0.zWoiIcDcDGY3XfTDij3AktHCocRjNbmB1XFhXJjMrhBO_yZypNRp1OCfKwjuSSkpLGmkhEiZrVAKGrZA7QtA0A&host_id=b6c92bbb-1e87-4972-b17a-12def3948890
    State:       installed
    State Info:  Done
  Inventory:
    Bmc Address:   0.0.0.0
    bmcV6Address:  ::/0
    Boot:
      Current Boot Mode:  uefi
    Cpu:
      Architecture:     x86_64
      Clock Megahertz:  3491
      Count:            16
      Flags:
        fpu
        vme
        de
        pse
        tsc
        msr
        pae
        mce
        cx8
        apic
        sep
        mtrr
        pge
        mca
        cmov
        pat
        pse36
        clflush
        mmx
        fxsr
        sse
        sse2
        ss
        syscall
        nx
        pdpe1gb
        rdtscp
        lm
        constant_tsc
        arch_perfmon
        rep_good
        nopl
        xtopology
        cpuid
        tsc_known_freq
        pni
        pclmulqdq
        vmx
        ssse3
        fma
        cx16
        pdcm
        pcid
        sse4_1
        sse4_2
        x2apic
        movbe
        popcnt
        tsc_deadline_timer
        aes
        xsave
        avx
        f16c
        rdrand
        hypervisor
        lahf_lm
        abm
        cpuid_fault
        invpcid_single
        pti
        ssbd
        ibrs
        ibpb
        stibp
        tpr_shadow
        vnmi
        flexpriority
        ept
        vpid
        ept_ad
        fsgsbase
        tsc_adjust
        bmi1
        avx2
        smep
        bmi2
        erms
        invpcid
        xsaveopt
        arat
        umip
        md_clear
        arch_capabilities
      Model Name:  Intel(R) Xeon(R) CPU E5-1650 v3 @ 3.50GHz
    Disks:
      By Path:     /dev/disk/by-path/pci-0000:00:1f.2-ata-1
      Drive Type:  ODD
      Hctl:        0:0:0:0
      Id:          /dev/disk/by-path/pci-0000:00:1f.2-ata-1
      Installation Eligibility:
        Not Eligible Reasons:
          Disk is removable
          Disk is too small (disk only has 108 MB, but 120 GB are required)
          Drive type is ODD, it must be one of HDD, SSD.
      Io Perf:
      Model:       QEMU_DVD-ROM
      Name:        sr0
      Path:        /dev/sr0
      Serial:      QM00001
      Size Bytes:  107884544
      Smart:       {"json_format_version":[1,0],"smartctl":{"version":[7,1],"svn_revision":"5022","platform_info":"x86_64-linux-4.18.0-305.10.2.el8_4.x86_64","build_info":"(local build)","argv":["smartctl","--xall","--json=c","/dev/sr0"],"exit_status":4},"device":{"name":"/dev/sr0","info_name":"/dev/sr0","type":"scsi","protocol":"SCSI"},"vendor":"QEMU","product":"QEMU DVD-ROM","model_name":"QEMU QEMU DVD-ROM","revision":"2.5+","scsi_version":"SPC-3","device_type":{"scsi_value":5,"name":"CD/DVD"},"local_time":{"time_t":1627844447,"asctime":"Sun Aug  1 19:00:47 2021 UTC"},"temperature":{"current":0,"drive_trip":0}}
      Vendor:      QEMU
      Bootable:    true
      By Path:     /dev/disk/by-path/pci-0000:04:00.0
      Drive Type:  HDD
      Id:          /dev/disk/by-path/pci-0000:04:00.0
      Installation Eligibility:
        Eligible:  true
        Not Eligible Reasons:
      Io Perf:
      Name:        vda
      Path:        /dev/vda
      Size Bytes:  214748364800
      Smart:       {"json_format_version":[1,0],"smartctl":{"version":[7,1],"svn_revision":"5022","platform_info":"x86_64-linux-4.18.0-305.10.2.el8_4.x86_64","build_info":"(local build)","argv":["smartctl","--xall","--json=c","/dev/vda"],"messages":[{"string":"/dev/vda: Unable to detect device type","severity":"error"}],"exit_status":1}}
      Vendor:      0x1af4
    Hostname:      sno
    Interfaces:
      Flags:
        up
        broadcast
        multicast
      Has Carrier:  true
      ipV4Addresses:
        192.168.123.5/24
      ipV6Addresses:
      Mac Address:  02:04:00:00:00:66
      Mtu:          1500
      Name:         enp1s0
      Product:      0x0001
      Speed Mbps:   -1
      Vendor:       0x1af4
    Memory:
      Physical Bytes:  68719476736
      Usable Bytes:    67514548224
    System Vendor:
      Manufacturer:  Red Hat
      Product Name:  KVM
      Virtual:       true
  Progress:
    Current Stage:      Done
    Stage Start Time:   2021-08-01T19:09:59Z
    Stage Update Time:  2021-08-01T19:09:59Z
  Role:                 master
Events:                 <none>
Monitor the ClusterDeployment

The cluster deployment is responsible for the OCP cluster. You can also monitor it, as this is the element that will give you the % of progress of the cluster install.

$ oc describe ClusterDeployments sno-ztp-cluster -n sno-ztp

--[cut]--
status:
  cliImage: >-
    quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5917b18697edb46458d9fd39cefab191c8324561fa83da160f6fdd0b90c55fe0
  conditions:
    - lastProbeTime: '2021-08-01T19:22:30Z'
      lastTransitionTime: '2021-08-01T18:47:08Z'
      message: >-
        The installation is in progress: Finalizing cluster installation.
        Cluster version status: progressing, message: Working towards 4.8.2: 640
        of 676 done (94% complete)
--[/cut]--

Accessing your cluster

After enough time, your cluster should be deployed. In order to get the kubeconfig / kubeadmin password, look at the ClusterDeployment CR, it will contain the secret name where to find the information. Note: the information will be populated only uppon successul deployment.

oc get ClusterDeployments -n sno-ztp
$ oc get ClusterDeployments -n sno-ztp
NAME               PLATFORM          REGION   CLUSTERTYPE   INSTALLED   INFRAID                                VERSION   POWERSTATE    AGE
sno-ztp-cluster    agent-baremetal                          true        064d8242-e63a-4ba9-9eb4-aaaa773cdf32             Unsupported   59m

$ oc describe ClusterDeployments sno-ztp-cluster -n sno-ztp
Name:         sno-ztp-cluster
Namespace:    sno-ztp
Labels:       hive.openshift.io/cluster-platform=agent-baremetal
Annotations:  open-cluster-management.io/user-group: c3lzdGVtOm1hc3RlcnMsc3lzdGVtOmF1dGhlbnRpY2F0ZWQ=
              open-cluster-management.io/user-identity: c3lzdGVtOmFkbWlu
API Version:  hive.openshift.io/v1
Kind:         ClusterDeployment
Metadata:
  Creation Timestamp:  2021-07-30T02:32:25Z
  Finalizers:
    hive.openshift.io/deprovision
    clusterdeployments.agent-install.openshift.io/ai-deprovision
Spec:
  Base Domain:  rhtelco.io
  Cluster Install Ref:
    Group:    extensions.hive.openshift.io
    Kind:     AgentClusterInstall
    Name:     sno-ztp-clusteragent
    Version:  v1beta1
  Cluster Metadata:
    Admin Kubeconfig Secret Ref:
      Name:  sno-ztp-cluster-admin-kubeconfig
    Admin Password Secret Ref:
      Name:      sno-ztp-cluster-admin-password
    Cluster ID:  82f49c11-7fb0-4185-82eb-0eab243fbfd2
    Infra ID:    e85bc2e6-ea53-4e0a-8e68-8922307a0159
  Cluster Name:  lab-spoke-adetalhouet
  Control Plane Config:
    Serving Certificates:
  Installed:  true
  Platform:
    Agent Bare Metal:
      Agent Selector:
        Match Labels:
          Location:  eu/fi
  Pull Secret Ref:
    Name:  assisted-deployment-pull-secret
Status:
  Cli Image:  quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5917b18697edb46458d9fd39cefab191c8324561fa83da160f6fdd0b90c55fe0
  Conditions:
    Last Probe Time:          2021-07-30T03:25:00Z
    Last Transition Time:     2021-07-30T02:32:37Z
    Message:                  The installation is in progress: Finalizing cluster installation. Cluster version status: available, message: Done applying 4.8.0
    Reason:                   InstallationInProgress
    Status:                   False
    Type:                     ClusterInstallCompleted
    Last Probe Time:          2021-07-30T03:28:01Z
    Last Transition Time:     2021-07-30T03:28:01Z
    Message:                  Unsupported platform: no actuator to handle it
    Reason:                   Unsupported
    Status:                   False
    Type:                     Hibernating
    Last Probe Time:          2021-07-30T03:28:01Z
    Last Transition Time:     2021-07-30T03:28:01Z
    Message:                  ClusterSync has not yet been created
    Reason:                   MissingClusterSync
    Status:                   True
    Type:                     SyncSetFailed
    Last Probe Time:          2021-07-30T03:28:01Z
    Last Transition Time:     2021-07-30T03:28:01Z
    Message:                  Get "https://api.lab-spoke-adetalhouet.rhtelco.io:6443/api?timeout=32s": dial tcp 148.251.12.17:6443: connect: connection refused
    Reason:                   ErrorConnectingToCluster
    Status:                   True
    Type:                     Unreachable
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  Platform credentials passed authentication check
    Reason:                   PlatformAuthSuccess
    Status:                   False
    Type:                     AuthenticationFailure
    Last Probe Time:          2021-07-30T02:32:37Z
    Last Transition Time:     2021-07-30T02:32:37Z
    Message:                  The installation has not failed
    Reason:                   InstallationNotFailed
    Status:                   False
    Type:                     ClusterInstallFailed
    Last Probe Time:          2021-07-30T02:41:00Z
    Last Transition Time:     2021-07-30T02:41:00Z
    Message:                  The cluster requirements are met
    Reason:                   ClusterAlreadyInstalling
    Status:                   True
    Type:                     ClusterInstallRequirementsMet
    Last Probe Time:          2021-07-30T02:32:37Z
    Last Transition Time:     2021-07-30T02:32:37Z
    Message:                  The installation is waiting to start or in progress
    Reason:                   InstallationNotStopped
    Status:                   False
    Type:                     ClusterInstallStopped
    Last Probe Time:          2021-07-30T03:28:01Z
    Last Transition Time:     2021-07-30T03:28:01Z
    Message:                  Control plane certificates are present
    Reason:                   ControlPlaneCertificatesFound
    Status:                   False
    Type:                     ControlPlaneCertificateNotFound
    Last Probe Time:          2021-07-30T02:32:37Z
    Last Transition Time:     2021-07-30T02:32:37Z
    Message:                  Images required for cluster deployment installations are resolved
    Reason:                   ImagesResolved
    Status:                   False
    Type:                     InstallImagesNotResolved
    Last Probe Time:          2021-07-30T02:32:37Z
    Last Transition Time:     2021-07-30T02:32:37Z
    Message:                  InstallerImage is resolved.
    Reason:                   InstallerImageResolved
    Status:                   False
    Type:                     InstallerImageResolutionFailed
    Last Probe Time:          2021-07-30T02:32:37Z
    Last Transition Time:     2021-07-30T02:32:37Z
    Message:                  The installation has not failed
    Reason:                   InstallationNotFailed
    Status:                   False
    Type:                     ProvisionFailed
    Last Probe Time:          2021-07-30T02:32:37Z
    Last Transition Time:     2021-07-30T02:32:37Z
    Message:                  The installation is waiting to start or in progress
    Reason:                   InstallationNotStopped
    Status:                   False
    Type:                     ProvisionStopped
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  no ClusterRelocates match
    Reason:                   NoMatchingRelocates
    Status:                   False
    Type:                     RelocationFailed
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  Condition Initialized
    Reason:                   Initialized
    Status:                   Unknown
    Type:                     AWSPrivateLinkFailed
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  Condition Initialized
    Reason:                   Initialized
    Status:                   Unknown
    Type:                     AWSPrivateLinkReady
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  Condition Initialized
    Reason:                   Initialized
    Status:                   Unknown
    Type:                     ActiveAPIURLOverride
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  Condition Initialized
    Reason:                   Initialized
    Status:                   Unknown
    Type:                     DNSNotReady
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  Condition Initialized
    Reason:                   Initialized
    Status:                   Unknown
    Type:                     DeprovisionLaunchError
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  Condition Initialized
    Reason:                   Initialized
    Status:                   Unknown
    Type:                     IngressCertificateNotFound
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  Condition Initialized
    Reason:                   Initialized
    Status:                   Unknown
    Type:                     InstallLaunchError
    Last Probe Time:          2021-07-30T02:32:25Z
    Last Transition Time:     2021-07-30T02:32:25Z
    Message:                  Condition Initialized
    Reason:                   Initialized
    Status:                   Unknown
    Type:                     RequirementsMet
  Install Started Timestamp:  2021-07-30T02:41:00Z
  Install Version:            4.8.0
  Installed Timestamp:        2021-07-30T02:41:00Z
  Installer Image:            quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eb3e6c54c4e2e07f95a9af44a5a1839df562a843b4ac9e1d5fb5bb4df4b4f7d6
Events:                       <none>

Some post deploy action

As I have a server with only one Interface and no console port access, I couldn't create a bridge interface for libvirt. So the poor man solution is to use iptables to forward the traffic hitting my public IP port 443 to my private VM IP.

They are more fancing way to do this.

when the host is stopped
/sbin/iptables -D FORWARD -o virbr1 -p tcp -d 192.168.123.5 --dport 443 -j ACCEPT
/sbin/iptables -t nat -D PREROUTING -p tcp --dport 443 -j DNAT --to 192.168.123.5:443

/sbin/iptables -D FORWARD -o virbr1 -p tcp -d 192.168.123.5 --dport 6443 -j ACCEPT
/sbin/iptables -t nat -D PREROUTING -p tcp --dport 6443 -j DNAT --to 192.168.123.5:6443
when is host is up
/sbin/iptables -I FORWARD -o virbr1 -p tcp -d 192.168.123.5 --dport 443 -j ACCEPT
/sbin/iptables -t nat -I PREROUTING -p tcp --dport 443 -j DNAT --to 192.168.123.5:443

/sbin/iptables -I FORWARD -o virbr1 -p tcp -d 192.168.123.5 --dport 6443 -j ACCEPT
/sbin/iptables -t nat -I PREROUTING -p tcp --dport 6443 -j DNAT --to 192.168.123.5:6443
backup

https://www.cyberciti.biz/faq/how-to-install-kvm-on-centos-8-headless-server/

https://wiki.libvirt.org/page/Networking#Forwarding_Incoming_Connections

https://www.itix.fr/blog/deploy-openshift-single-node-in-kvm/

oc patch provisioning provisioning-configuration --type merge -p '{"spec":{"watchAllNamespaces": true}}'
oc patch hiveconfig hive --type merge -p '{"spec":{"targetNamespace":"hive","logLevel":"debug","featureGates":{"custom":{"enabled":["AlphaAgentInstallStrategy"]},"featureSet":"Custom"}}}'

About

Deploy OCP using Zero Touch Provisioning on libvirt

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages