Skip to content

mhenriks/kubevirtci

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Getting Started with a multi-node Kubernetes Provider

Download this repo

git clone https://github.com/kubevirt/kubevirtci.git
cd kubevirtci

Start multi node k8s cluster with 2 nics

export KUBEVIRT_PROVIDER=k8s-1.13.3 KUBEVIRT_NUM_NODES=2 KUBEVIRT_NUM_SECONDARY_NICS=1
make cluster-up

Stop k8s cluster

make cluster-down

Use provider's kubectl client with kubectl.sh wrapper script

cluster-up/kubectl.sh get nodes
cluster-up/kubectl.sh get pods --all-namespaces

Use your own kubectl client by defining the KUBECONFIG environment variable

export KUBECONFIG=$(cluster-up/kubeconfig.sh)

kubectl get nodes
kubectl apply -f <some file>

SSH into a node

cluster-up/ssh.sh node01

Attach to node console with socat and pty (non okd4 providers)

# Get node01 container id
node01_id=$(docker ps |grep node01 |awk '{print $1}')

# Install socat
docker exec $node01_id yum install -y socat

# Attach to node01 console
docker exec -it $node01_id socat - /dev/pts/0

Getting Started with multi-node OKD Provider

Download this repo

git clone https://github.com/kubevirt/kubevirtci.git
cd kubevirtci

Start okd cluster (pre-configured with a master and worker node)

export KUBEVIRT_PROVIDER=okd-4.1
# export OKD_CONSOLE_PORT=443  # Uncomment to access OKD console
make cluster-up

Stop okd cluster

make cluster-down

Use provider's OC client with oc.sh wrapper script

cluster-up/oc.sh get nodes
cluster-up/oc.sh get pods --all-namespaces

Use your own OC client by defining the KUBECONFIG environment variable

export KUBECONFIG=$(cluster-up/kubeconfig.sh)

oc get nodes
oc apply -f <some file>

SSH into master

cluster-up/ssh.sh master-0

SSH into worker

cluster-up/ssh.sh worker-0

Connect to the container (with KUBECONFIG exported)

make connect

In order to check newly created provider run, this will point to the local created provider upon cluster-up

export KUBEVIRTCI_PROVISION_CHECK=1

To access the OKD UI from the host running docker, remember to export OKD_CONSOLE_PORT=443 before make cluster-up. You should find out the IP address of the OKD docker container

clusterip=$(docker inspect $(docker ps | grep "kubevirtci/$KUBEVIRT_PROVIDER" | awk '{print $1}') | jq -r '.[0].NetworkSettings.IPAddress' )

and make it known in /etc/hosts via

cat << EOF >> /etc/hosts
$clusterip console-openshift-console.apps.test-1.tt.testing
$clusterip oauth-openshift.apps.test-1.tt.testing
EOF

Now you can browse to https://console-openshift-console.apps.test-1.tt.testing and log in by picking the htpasswd_provider option. The credentials are admin/admin.

To access the OKD UI from a remote client, forward incoming port 433 into the OKD cluster on the host running kubevirtci:

$ nic=em1  # the interface facing your remote client
$ sudo iptables -t nat -A PREROUTING -p tcp -i $nic --dport 443 -j DNAT --to-destination $clusterip

On your remote client host, point the cluster fqdn to the host running kubevirtci

kubevirtci_ip=a.b.c.d  # put here the ip address of the host running kubevirtci
cat << EOF >> /etc/hosts
$kubevirtci_ip console-openshift-console.apps.test-1.tt.testing
$kubevirtci_ip oauth-openshift.apps.test-1.tt.testing
EOF

About

Contains cluster definitions and client tools to quickly spin up and destroy ephemeral and scalable k8s and ocp clusters for testing

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Shell 59.4%
  • Go 24.5%
  • Roff 12.0%
  • Python 2.8%
  • Other 1.3%