Update the rook cluster

This commit is contained in:
Nikolai Rodionov 2024-07-24 13:17:35 +02:00
parent 9ec7041ea9
commit 2531a88ce1
Signed by: allanger
GPG Key ID: 0AA46A90E25592AD

View File

@ -42,69 +42,6 @@ cephFileSystems:
csi.storage.k8s.io/node-stage-secret-namespace: "{{ .Release.Namespace }}" csi.storage.k8s.io/node-stage-secret-namespace: "{{ .Release.Namespace }}"
csi.storage.k8s.io/fstype: ext4 csi.storage.k8s.io/fstype: ext4
cephBlockPools: []
# - name: ceph-blockpool
# # see https://github.com/rook/rook/blob/master/Documentation/CRDs/Block-Storage/ceph-block-pool-crd.md#spec for available configuration
# spec:
# failureDomain: host
# replicated:
# size: 3
# # Enables collecting RBD per-image IO statistics by enabling dynamic OSD performance counters. Defaults to false.
# # For reference: https://docs.ceph.com/docs/latest/mgr/prometheus/#rbd-io-statistics
# # enableRBDStats: true
# storageClass:
# enabled: true
# name: ceph-block
# isDefault: false
# reclaimPolicy: Delete
# allowVolumeExpansion: true
# volumeBindingMode: "Immediate"
# mountOptions: []
# # - discard
# allowedTopologies: []
# parameters:
# imageFormat: "2"
# imageFeatures: layering
# csi.storage.k8s.io/provisioner-secret-name: rook-csi-rbd-provisioner
# csi.storage.k8s.io/provisioner-secret-namespace: "{{ .Release.Namespace }}"
# csi.storage.k8s.io/controller-expand-secret-name: rook-csi-rbd-provisioner
# csi.storage.k8s.io/controller-expand-secret-namespace: "{{ .Release.Namespace }}"
# csi.storage.k8s.io/node-stage-secret-name: rook-csi-rbd-node
# csi.storage.k8s.io/node-stage-secret-namespace: "{{ .Release.Namespace }}"
# csi.storage.k8s.io/fstype: ext4
cephObjectStores: []
# - name: ceph-objectstore
# spec:
# metadataPool:
# failureDomain: host
# replicated:
# size: 3
# dataPool:
# failureDomain: host
# erasureCoded:
# dataChunks: 2
# codingChunks: 1
# preservePoolsOnDelete: true
# gateway:
# port: 80
# resources:
# limits:
# cpu: "150m"
# memory: "256Mi"
# requests:
# cpu: "50m"
# memory: "128Mi"
# instances: 1
# priorityClassName: system-cluster-critical
# storageClass:
# enabled: true
# name: ceph-bucket
# reclaimPolicy: Delete
# volumeBindingMode: "Immediate"
# parameters:
# region: us-east-1
# ingress:
# enabled: false
cephClusterSpec: cephClusterSpec:
dashboard: dashboard:
enabled: true enabled: true
@ -123,13 +60,18 @@ cephClusterSpec:
mon: mon:
count: 1 count: 1
resources: resources:
mgr: api:
limits:
cpu: "400m"
memory: "512Mi"
requests: requests:
cpu: "200m" cpu: "500m"
memory: "256Mi" memory: "512Mi"
limits:
memory: "512Mi"
mgr:
requests:
cpu: "500m"
memory: "512Mi"
limits:
memory: "512Mi"
mon: mon:
limits: limits:
cpu: "2000m" cpu: "2000m"
@ -138,32 +80,24 @@ cephClusterSpec:
cpu: "300m" cpu: "300m"
memory: "128Mi" memory: "128Mi"
osd: osd:
limits:
cpu: "400m"
memory: "1280Mi"
requests: requests:
cpu: "200m" cpu: "500m"
memory: "256Mi" memory: "512Mi"
limits:
memory: "1280Mi"
#limits:
# cpu: "400m"
# memory: "1280Mi"
#requests:
# cpu: "200m"
# memory: "256Mi"
cephFileSystemVolumeSnapshotClass: cephFileSystemVolumeSnapshotClass:
enabled: true enabled: true
name: ceph-filesystem name: ceph-filesystem
isDefault: true isDefault: true
deletionPolicy: Delete deletionPolicy: Delete
annotations: {}
labels: {}
# see https://rook.io/docs/rook/v1.10/Storage-Configuration/Ceph-CSI/ceph-csi-snapshot/#cephfs-snapshots for available configuration
parameters: {}
cephBlockPoolsVolumeSnapshotClass:
enabled: true
name: ceph-block
isDefault: false
deletionPolicy: Delete
annotations: {}
labels: {}
# see https://rook.io/docs/rook/v1.10/Storage-Configuration/Ceph-CSI/ceph-csi-snapshot/#rbd-snapshots for available configuration
parameters: {}
ingress: ingress:
# -- Enable an ingress for the ceph-dashboard # -- Enable an ingress for the ceph-dashboard
dashboard: dashboard: