Quick start
This tutorial will guide you to a working Ceph setup using the default configuration.
Prerequisites
-
Cluster is already managed by Project Syn
-
At least 3 adequately sized nodes (minimum 8 vCPU, 16GB RAM) are provisioned for the storage cluster
-
The nodes are labeled with
node-role.kubernetes.io/storage=''
and tainted withstoragenode=True:NoSchedule
-
-
Block storage volumes (
volumeMode=Block
) are available in the cluster -
The block storage volumes are associated with a StorageClass
-
A working
commodore
command is available locally. See the Running Commodore documentation for details.
Steps
-
Enable component
rook-ceph
c-cluster-id.ymlapplications: - rook-ceph
-
Configure backing storageclass for the Ceph cluster
c-cluster-id.ymlparameters: rook_ceph: ceph_cluster: block_storage_class: <block-storage-class> (1)
1 Specify the name of a storage class which provides volumes with volumeMode=Block
. -
If your block storage volumes are provisioned dynamically, specify the volume size
c-cluster-id.ymlparameters: rook_ceph: ceph_cluster: block_volume_size: 200Gi (1)
1 If you are using pre-provisioned block storage volumes, you can leave this parameter at its default value of 1
. -
If your block storage volumes are backed by SSDs (or better), tune Ceph for SSD backing storage.
c-cluster-id.ymlparameters: rook_ceph: ceph_cluster: tune_fast_device_class: true
-
Compile and push the cluster catalog
commodore catalog compile <c-cluster-id> --push -i