This quickstart will get you up and running using Static IPs. This is assuming you're on bare metal or in another environment without cloud integration.
Create a machine/vm with the following minimum configuration.
- CentOS/RHEL 7 or 8
- 50GB HD
- 4 CPUs
- 8 GB of RAM
In this example, I'll be using the following.
- CentOS 8
- 50GB HD
- 4CPUs
- 8 GB of RAM
- IP - 192.168.7.77
- NetMask - 255.255.255.0
- Default Gateway - 192.168.7.1
- DNS Server - 8.8.8.8
After the helper node is installed; login to it
NOTE If using RHEL 7 - you need to enable the
rhel-7-server-rpms
and therhel-7-server-extras-rpms
repos. If you're using RHEL 8 you will need to enablerhel-8-for-x86_64-baseos-rpms
,rhel-8-for-x86_64-appstream-rpms
, andansible-2.9-for-rhel-8-x86_64-rpms
Install EPEL
yum -y install https://dl.fedoraproject.org/pub/epel/epel-release-latest-$(rpm -E %rhel).noarch.rpm
Install ansible
and git
and clone this repo
yum -y install ansible git
git clone https://github.com/RedHatOfficial/ocp4-helpernode
cd ocp4-helpernode
Create the vars-static.yaml file with the IP addresss that will be assigned to the masters/workers/boostrap. The IP addresses need to be right since they will be used to create your DNS server.
cp docs/examples/vars-static.yaml .
vi vars-static.yaml
NOTE See the
vars.yaml
documentaion page for more info about what it does.
Run the playbook to setup your helper node (using -e staticips=true
to flag to ansible that you won't be installing dhcp/tftp)
ansible-playbook -e @vars-static.yaml -e staticips=true tasks/main.yml
After it is done run the following to get info about your environment and some install help
/usr/local/bin/helpernodecheck
Now you can start the installation process. Create an install dir.
mkdir ~/ocp4
cd ~/ocp4
Create a place to store your pull-secret
mkdir -p ~/.openshift
Visit try.openshift.com and select "Bare Metal". Download your pull secret and save it under ~/.openshift/pull-secret
# ls -1 ~/.openshift/pull-secret
/root/.openshift/pull-secret
This playbook creates an sshkey for you; it's under ~/.ssh/helper_rsa
. You can use this key or create/user another one if you wish.
# ls -1 ~/.ssh/helper_rsa
/root/.ssh/helper_rsa
⚠️ If you want you use your own sshkey, please modify~/.ssh/config
to reference your key instead of the one deployed by the playbook
Next, create an install-config.yaml
file.
⚠️ Make sure you update if your filenames or paths are different.
cat <<EOF > install-config.yaml
apiVersion: v1
baseDomain: example.com
compute:
- hyperthreading: Enabled
name: worker
replicas: 0
controlPlane:
hyperthreading: Enabled
name: master
replicas: 3
metadata:
name: ocp4
networking:
clusterNetworks:
- cidr: 10.254.0.0/16
hostPrefix: 24
networkType: OpenShiftSDN
serviceNetwork:
- 172.30.0.0/16
platform:
none: {}
pullSecret: '$(< ~/.openshift/pull-secret)'
sshKey: '$(< ~/.ssh/helper_rsa.pub)'
EOF
Create the installation manifests
openshift-install create manifests
Edit the manifests/cluster-scheduler-02-config.yml
Kubernetes manifest file to prevent Pods from being scheduled on the control plane machines by setting mastersSchedulable
to false
.
🚨 Skip this step if you're installing a compact cluster
$ sed -i 's/mastersSchedulable: true/mastersSchedulable: false/g' manifests/cluster-scheduler-02-config.yml
It should look something like this after you edit it.
$ cat manifests/cluster-scheduler-02-config.yml
apiVersion: config.openshift.io/v1
kind: Scheduler
metadata:
creationTimestamp: null
name: cluster
spec:
mastersSchedulable: false
policy:
name: ""
status: {}
Next, generate the ignition configs
openshift-install create ignition-configs
Finally, copy the ignition files in the ignition
directory for the websever
cp ~/ocp4/*.ign /var/www/html/ignition/
restorecon -vR /var/www/html/
chmod o+r /var/www/html/ignition/*.ign
⚠️ Read all the instructions before attempting to install RHCOS!
Boot into your instance using the RHCOS ISO Installer
Once booted; press tab
on the boot menu
Add your staticips and coreos options. Here is an example of what I used for my bootstrap node. (type this ALL IN ONE LINE ...I only used linebreaks here for ease of readability...but type it all in one line)
If installing 4.5 and earlier, you need
coreos.inst.image_url=http://192.168.7.77:8080/install/bios.raw.gz
ip=192.168.7.20::192.168.7.1:255.255.255.0:bootstrap.ocp4.example.com:enp1s0:none
nameserver=192.168.7.77
coreos.inst.install_dev=vda
coreos.live.rootfs_url=http://192.168.7.77:8080/install/rootfs.img
coreos.inst.ignition_url=http://192.168.7.77:8080/ignition/bootstrap.ign
^ Do this for ALL of your VMs!!!
NOTE Using
ip=...
syntax will set the host with a static IP you provided persistantly accross reboots. The syntax isip=<ipaddress>::<defaultgw>:<netmask>:<hostname>:<iface>:none
. To set the DNS server usenameserver=<dnsserver>
. You can usenameserver=
multiple times.
Boot/install the VMs in the following order
- Bootstrap
- Masters
- Workers
On your laptop/workstation visit the status page
firefox http://192.168.7.77:9000
You'll see the bootstrap turn "green" and then the masters turn "green", then the bootstrap turn "red". This is your indication that you can continue.
Manually booting into the ISO and typing in the kernel parameters for ALL nodes can be cumbersome. You MAY want to opt to use Chuckers' ISO maker. I've written a little how to for the HelperNode.
The boostrap VM actually does the install for you; you can track it with the following command.
openshift-install wait-for bootstrap-complete --log-level debug
Once you see this message below...
DEBUG OpenShift Installer v4.2.0-201905212232-dirty
DEBUG Built from commit 71d8978039726046929729ad15302973e3da18ce
INFO Waiting up to 30m0s for the Kubernetes API at https://api.ocp4.example.com:6443...
INFO API v1.13.4+838b4fa up
INFO Waiting up to 30m0s for bootstrapping to complete...
DEBUG Bootstrap status: complete
INFO It is now safe to remove the bootstrap resources
...you can continue....at this point you can delete the bootstrap server.
First, login to your cluster
export KUBECONFIG=/root/ocp4/auth/kubeconfig
Your install may be waiting for worker nodes to get approved. Normally the machineconfig node approval operator
takes care of this for you. However, sometimes this needs to be done manually. Check pending CSRs with the following command.
oc get csr
You can approve all pending CSRs in "one shot" with the following
oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs oc adm certificate approve
You may have to run this multiple times depending on how many workers you have and in what order they come in. Keep a watch
on these CSRs
watch oc get csr
In order to setup your registry, you first have to set the managementState
to Managed
for your cluster
oc patch configs.imageregistry.operator.openshift.io cluster --type merge --patch '{"spec":{"managementState":"Managed"}}'
For PoCs, using emptyDir
is okay (to use PVs follow this doc)
oc patch configs.imageregistry.operator.openshift.io cluster --type merge --patch '{"spec":{"storage":{"emptyDir":{}}}}'
If you need to expose the registry, run this command
oc patch configs.imageregistry.operator.openshift.io/cluster --type merge -p '{"spec":{"defaultRoute":true}}'
To finish the install process, run the following
openshift-install wait-for install-complete
Note: You can watch the operators running with
oc get clusteroperators
in another window with awatch
to see it progress
The OpenShift 4 web console will be running at https://console-openshift-console.apps.{{ dns.clusterid }}.{{ dns.domain }}
(e.g. https://console-openshift-console.apps.ocp4.example.com
)
- Username: kubeadmin
- Password: the output of
cat /root/ocp4/auth/kubeadmin-password
If you didn't install the latest release, then just run the following to upgrade.
oc adm upgrade --to-latest
Scale the router if you need to
oc patch --namespace=openshift-ingress-operator --patch='{"spec": {"replicas": 3}}' --type=merge ingresscontroller/default
Your install should be done! You're a UPI master!