From 2531a88ce148dd244d7ff81c47508855e8530ab9 Mon Sep 17 00:00:00 2001 From: Nikolai Rodionov Date: Wed, 24 Jul 2024 13:17:35 +0200 Subject: [PATCH] Update the rook cluster --- .../values.rook-ceph-cluster.yaml | 108 ++++-------------- 1 file changed, 21 insertions(+), 87 deletions(-) diff --git a/values/badhouseplants/values.rook-ceph-cluster.yaml b/values/badhouseplants/values.rook-ceph-cluster.yaml index 453302a..7b8d50e 100644 --- a/values/badhouseplants/values.rook-ceph-cluster.yaml +++ b/values/badhouseplants/values.rook-ceph-cluster.yaml @@ -42,69 +42,6 @@ cephFileSystems: csi.storage.k8s.io/node-stage-secret-namespace: "{{ .Release.Namespace }}" csi.storage.k8s.io/fstype: ext4 -cephBlockPools: [] - # - name: ceph-blockpool - # # see https://github.com/rook/rook/blob/master/Documentation/CRDs/Block-Storage/ceph-block-pool-crd.md#spec for available configuration - # spec: - # failureDomain: host - # replicated: - # size: 3 - # # Enables collecting RBD per-image IO statistics by enabling dynamic OSD performance counters. Defaults to false. - # # For reference: https://docs.ceph.com/docs/latest/mgr/prometheus/#rbd-io-statistics - # # enableRBDStats: true - # storageClass: - # enabled: true - # name: ceph-block - # isDefault: false - # reclaimPolicy: Delete - # allowVolumeExpansion: true - # volumeBindingMode: "Immediate" - # mountOptions: [] - # # - discard - # allowedTopologies: [] - # parameters: - # imageFormat: "2" - # imageFeatures: layering - # csi.storage.k8s.io/provisioner-secret-name: rook-csi-rbd-provisioner - # csi.storage.k8s.io/provisioner-secret-namespace: "{{ .Release.Namespace }}" - # csi.storage.k8s.io/controller-expand-secret-name: rook-csi-rbd-provisioner - # csi.storage.k8s.io/controller-expand-secret-namespace: "{{ .Release.Namespace }}" - # csi.storage.k8s.io/node-stage-secret-name: rook-csi-rbd-node - # csi.storage.k8s.io/node-stage-secret-namespace: "{{ .Release.Namespace }}" - # csi.storage.k8s.io/fstype: ext4 -cephObjectStores: [] - # - name: ceph-objectstore - # spec: - # metadataPool: - # failureDomain: host - # replicated: - # size: 3 - # dataPool: - # failureDomain: host - # erasureCoded: - # dataChunks: 2 - # codingChunks: 1 - # preservePoolsOnDelete: true - # gateway: - # port: 80 - # resources: - # limits: - # cpu: "150m" - # memory: "256Mi" - # requests: - # cpu: "50m" - # memory: "128Mi" - # instances: 1 - # priorityClassName: system-cluster-critical - # storageClass: - # enabled: true - # name: ceph-bucket - # reclaimPolicy: Delete - # volumeBindingMode: "Immediate" - # parameters: - # region: us-east-1 - # ingress: - # enabled: false cephClusterSpec: dashboard: enabled: true @@ -123,13 +60,18 @@ cephClusterSpec: mon: count: 1 resources: - mgr: - limits: - cpu: "400m" - memory: "512Mi" + api: requests: - cpu: "200m" - memory: "256Mi" + cpu: "500m" + memory: "512Mi" + limits: + memory: "512Mi" + mgr: + requests: + cpu: "500m" + memory: "512Mi" + limits: + memory: "512Mi" mon: limits: cpu: "2000m" @@ -138,32 +80,24 @@ cephClusterSpec: cpu: "300m" memory: "128Mi" osd: - limits: - cpu: "400m" - memory: "1280Mi" requests: - cpu: "200m" - memory: "256Mi" + cpu: "500m" + memory: "512Mi" + limits: + memory: "1280Mi" + #limits: + # cpu: "400m" + # memory: "1280Mi" + #requests: + # cpu: "200m" + # memory: "256Mi" cephFileSystemVolumeSnapshotClass: enabled: true name: ceph-filesystem isDefault: true deletionPolicy: Delete - annotations: {} - labels: {} - # see https://rook.io/docs/rook/v1.10/Storage-Configuration/Ceph-CSI/ceph-csi-snapshot/#cephfs-snapshots for available configuration - parameters: {} -cephBlockPoolsVolumeSnapshotClass: - enabled: true - name: ceph-block - isDefault: false - deletionPolicy: Delete - annotations: {} - labels: {} - # see https://rook.io/docs/rook/v1.10/Storage-Configuration/Ceph-CSI/ceph-csi-snapshot/#rbd-snapshots for available configuration - parameters: {} ingress: # -- Enable an ingress for the ceph-dashboard dashboard: