FreeBSD virtual environment management and repository

2020-10 upd: we reached the first fundraising goal and rented a server in Hetzner for development! Thank you for donating !

Attention! Current pages describe CBSD version 13.0.x. If you are using an older version, please update first.

Attention! I apologize for the automatic translation of this text. You can improve it by sending me a more correct version of the text or fix html pages via GITHUB repository.

CBSD k8s module: deploying a kubernetes cluster


The module consists of two components - an image of a Linux-based virtual machine with kubernetes, prepared as part of the k8s-bhyve prject and scripts, that configure and launch the cluster

Installing K8S module

Installation of the module is done by the regular module script from the GitHub repository of the project:

cbsd module mode=install k8s

The module uses a pre-formed image, which must first be obtained:

cbsd fetch_iso name=cloud-kubernetes-20 dstdir=default cloud=1 conv2zvol=1

Activate the module through the config file and re-initialize CBSD:

echo 'k8s.d' >> ~cbsd/etc/modules.conf
cbsd initenv

If the cbsd k8s command exists, then the module is ready to work.

K8S cluster initialization

To initialize a new K8S cluster, use the command: cbsd k8s mode=init:

cbsd k8s mode=init k8s_name=k1

where k1 - name of the cluster profile.

The number of master and worker nodes is regulated by the number of IP addresses that you assign through the init_masters_ips and init_nodes_ips parameters

In addition, 1 IP address is assigned as an API endpoint, via the vip=, virtual IP parameter

You can assign fixed addresses for masters and workers, or get them automatically from the CBSD pool by setting DHCP as addresses, for example

cbsd k8s mode=init k8s_name=k1 init_masters_ips="DHCP DHCP DHCP" init_nodes_ips="DHCP DHCP DHCP" vip=DHCP

As a result of this command, you will get a cluster named, consisting of 3 masters and 3 workers with automatically assigned IP addresses.

Other possible arguments and their description:

k8s_nameprofile name, short unique ID, for example: k1
vpcuse CBSD VPC in which to deploy the cluster
clustername of Kubernetes cluster, by default:
master_hostnamehostname for master node
k8s_verwhich version of K8S to use
etcd_verwhich version of ETCD to use
flannel_verwhich version of flannel to use
init_masters_ipslist of IP addresses for master nodes. The number of IP determines the number of masters
init_nodes_ipslist of IP addresses for worker nodes. The number of IP determines the number of worker
vipIP of VRRP, for cluster API Endpoint
ip4_gwIP address, used by VM as default gateway, by default:
dns_ipIP address for the internal DNS server in the kuberenetes network
coredns_enableinstall CoreDNS service?
ingress_hosthostname of the Ingress service
kubelet_masterIt controls whether the master node can also execute worker functions and run containers, by default - yes: 1
pv_enableUse PV ? By default: 0
master_interfacespecify an alternative uplink interface of the VM master nodes. default: auto
worker_interfacespecify an alternative uplink interface of the VM worker nodes. default: auto
master_vm_ramMaster node configuration, amount of RAM. By default: 2g
master_vm_cpusMaster node configuration, amount of vCPU. By default: 1
master_vm_imgsizeMaster node configuration, disk space. By default: 20g
worker_vm_ramWorker node configuration, amount of RAM.. By default: 2g
worker_vm_cpusWorker node configuration, amount of vCPU. By default: 1
worker_vm_imgsizeWorker node configuration, disk space. By default: 20g

Initializing CBSDfile

When the k8s cluster is initialized, you must generate a CBSDfile to start and stop the cluster. To do this, use the command: k8s mode=init_upfile:

Two files will be generated in the current working directory - CBSDfile and bootstrap.config. This is all you need to start the cluster.

cbsd k8s mode=init_upfile k8s_name=k1

K8S cluster launch

Once in the directory where CBSDfile and bootstrap.config are generated, run the command: cbsd up:

cbsd up

Upon completion of initialization, the system will import kubeconfig into working directory.

You can copy it to another host or manage the cluster via kube and helm commands from your host system.

Destroying of the K8S cluster

Once in the directory where CBSDfile and bootstrap.config are generated, run the command: cbsd destroy:

cbsd destroy


PV is configured with the pv_enable=1 option and the corresponding pv_spec-* parameters.

Attention: The current version will automatically configure the NFS server, which entails a complete generation of /etc/exports and modification of /etc/rc.conf files, followed by the launch of the appropriate services.

Creation of two independent K8S clusters of different configurations with PV-NFS on single host