Enable ReadWriteMany PVC Support in Additional Namespaces¶

The default general cephfs-provisioner storage class is enabled for the default, kube-system, and kube-public namespaces. To enable an additional namespace, for example for an application-specific namespace, a modification to the configuration (Helm overrides) of the cephfs-provisioner service is required.

About this task

The following example illustrates the configuration of three additional application-specific namespaces to access the cephfs-provisioner cephfs storage class.

Note

Due to limitations with templating and merging of overrides, the entire storage class must be redefined in the override when updating specific values.

Procedure

  1. List installed Helm chart overrides for the platform-integ-apps.

    ~(keystone_admin)]$ system helm-override-list platform-integ-apps
    +--------------------+----------------------+
    | chart name         | overrides namespaces |
    +--------------------+----------------------+
    | ceph-pools-audit   | ['kube-system']      |
    | cephfs-provisioner | ['kube-system']      |
    | rbd-provisioner    | ['kube-system']      |
    +--------------------+----------------------+
    
  2. Review existing overrides for the cephfs-provisioner chart. You will refer to this information in the following step.

    ~(keystone_admin)]$ system helm-override-show platform-integ-apps cephfs-provisioner kube-system
    
    +--------------------+------------------------------------------------------+
    | Property           | Value                                                |
    +--------------------+------------------------------------------------------+
    | attributes         | enabled: true                                        |
    |                    |                                                      |
    | combined_overrides | classdefaults:                                       |
    |                    |   adminId: admin                                     |
    |                    |   adminSecretName: ceph-secret-admin                 |
    |                    |   monitors:                                          |
    |                    |   - 192.168.204.2:6789                               |
    |                    | csiConfig:                                           |
    |                    | - cephFS:                                            |
    |                    |     subvolumeGroup: csi                              |
    |                    |   clusterID: 6d273112-f2a6-4aec-8727-76b690274c60    |
    |                    |   monitors:                                          |
    |                    |   - 192.168.204.2:6789                               |
    |                    | provisioner:                                         |
    |                    |   replicaCount: 1                                    |
    |                    |   snapshotter:                                       |
    |                    |     enabled: true                                    |
    |                    | snapshotClass:                                       |
    |                    |   clusterID: 6d273112-f2a6-4aec-8727-76b690274c60    |
    |                    |   provisionerSecret: ceph-pool-kube-cephfs-data      |
    |                    | storageClasses:                                      |
    |                    | - additionalNamespaces:                              |
    |                    |   - default                                          |
    |                    |   - kube-public                                      |
    |                    |   chunk_size: 64                                     |
    |                    |   clusterID: 6d273112-f2a6-4aec-8727-76b690274c60    |
    |                    |   controllerExpandSecret: ceph-pool-kube-cephfs-data |
    |                    |   crush_rule_name: storage_tier_ruleset              |
    |                    |   data_pool_name: kube-cephfs-data                   |
    |                    |   fs_name: kube-cephfs                               |
    |                    |   metadata_pool_name: kube-cephfs-metadata           |
    |                    |   name: cephfs                                       |
    |                    |   nodeStageSecret: ceph-pool-kube-cephfs-data        |
    |                    |   provisionerSecret: ceph-pool-kube-cephfs-data      |
    |                    |   replication: 1                                     |
    |                    |   userId: ceph-pool-kube-cephfs-data                 |
    |                    |   userSecretName: ceph-pool-kube-cephfs-data         |
    |                    |   volumeNamePrefix: pvc-volumes-                     |
    |                    | csiConfig:                                           |
    |                    | - cephFS:                                            |
    |                    |     subvolumeGroup: csi                              |
    |                    |   clusterID: 6d273112-f2a6-4aec-8727-76b690274c60    |
    |                    |   monitors:                                          |
    |                    |   - 192.168.204.2:6789                               |
    |                    | provisioner:                                         |
    |                    |   replicaCount: 1                                    |
    |                    |                                                      |
    | name               | cephfs-provisioner                                   |
    | namespace          | kube-system                                          |
    | system_overrides   | classdefaults:                                       |
    |                    |   adminId: admin                                     |
    |                    |   adminSecretName: ceph-secret-admin                 |
    |                    |   monitors: ['192.168.204.2:6789']                   |
    |                    | csiConfig:                                           |
    |                    | - cephFS: {subvolumeGroup: csi}                      |
    |                    |   clusterID: !!binary |                              |
    |                    |     NmQyNzMxMTItZjJhNi00YWVjLTg3MjctNzZiNjkwMjc0YzYw |
    |                    |   monitors: ['192.168.204.2:6789']                   |
    |                    | provisioner:                                         |
    |                    |   replicaCount: 1                                    |
    |                    |   snapshotter: {enabled: true}                       |
    |                    | snapshotClass:                                       |
    |                    |   clusterID: !!binary |                              |
    |                    |     NmQyNzMxMTItZjJhNi00YWVjLTg3MjctNzZiNjkwMjc0YzYw |
    |                    |   provisionerSecret: ceph-pool-kube-cephfs-data      |
    |                    | storageClasses:                                      |
    |                    | - additionalNamespaces: [default, kube-public]       |
    |                    |   chunk_size: 64                                     |
    |                    |   clusterID: !!binary |                              |
    |                    |     NmQyNzMxMTItZjJhNi00YWVjLTg3MjctNzZiNjkwMjc0YzYw |
    |                    |   controllerExpandSecret: ceph-pool-kube-cephfs-data |
    |                    |   crush_rule_name: storage_tier_ruleset              |
    |                    |   data_pool_name: kube-cephfs-data                   |
    |                    |   fs_name: kube-cephfs                               |
    |                    |   metadata_pool_name: kube-cephfs-metadata           |
    |                    |   name: cephfs                                       |
    |                    |   nodeStageSecret: ceph-pool-kube-cephfs-data        |
    |                    |   provisionerSecret: ceph-pool-kube-cephfs-data      |
    |                    |   replication: 1                                     |
    |                    |   userId: ceph-pool-kube-cephfs-data                 |
    |                    |   userSecretName: ceph-pool-kube-cephfs-data         |
    |                    |   volumeNamePrefix: pvc-volumes-                     |
    |                    | csiConfig:                                           |
    |                    | - cephFS: {subvolumeGroup: csi}                      |
    |                    |   clusterID: !!binary |                              |
    |                    |     NmQyNzMxMTItZjJhNi00YWVjLTg3MjctNzZiNjkwMjc0YzYw |
    |                    |   monitors: ['192.168.204.2:6789']                   |
    |                    | provisioner: {replicaCount: 1}                       |
    |                    |                                                      |
    | user_overrides     | None                                                 |
    +--------------------+------------------------------------------------------+
    
  3. Create an overrides yaml file defining the new namespaces.

    In this example, create the file /home/sysadmin/update-namespaces.yaml with the following content:

    ~(keystone_admin)]$ cat <<EOF > ~/update-namespaces.yaml
    storageClasses:
    - additionalNamespaces: [default, kube-public, new-app, new-app2, new-app3]
      chunk_size: 64
      claim_root: /pvc-volumes
      crush_rule_name: storage_tier_ruleset
      data_pool_name: kube-cephfs-data
      fs_name: kube-cephfs
      metadata_pool_name: kube-cephfs-metadata
      name: cephfs
      replication: 2
      userId: ceph-pool-kube-cephfs-data
      userSecretName: ceph-pool-kube-cephfs-data
    EOF
    
  4. Apply the overrides file to the chart.

    ~(keystone_admin)]$ system helm-override-update  --values /home/sysadmin/update-namespaces.yaml platform-integ-apps cephfs-provisioner kube-system
    +----------------+----------------------------------------------+
    | Property       | Value                                        |
    +----------------+----------------------------------------------+
    | name           | cephfs-provisioner                           |
    | namespace      | kube-system                                  |
    | user_overrides | storageClasses:                              |
    |                | - additionalNamespaces:                      |
    |                |   - default                                  |
    |                |   - kube-public                              |
    |                |   - new-app                                  |
    |                |   - new-app2                                 |
    |                |   - new-app3                                 |
    |                |   chunk_size: 64                             |
    |                |   claim_root: /pvc-volumes                   |
    |                |   crush_rule_name: storage_tier_ruleset      |
    |                |   data_pool_name: kube-cephfs-data           |
    |                |   fs_name: kube-cephfs                       |
    |                |   metadata_pool_name: kube-cephfs-metadata   |
    |                |   name: cephfs                               |
    |                |   replication: 2                             |
    |                |   userId: ceph-pool-kube-cephfs-data         |
    |                |   userSecretName: ceph-pool-kube-cephfs-data |
    +----------------+----------------------------------------------+
    
  5. Confirm that the new overrides have been applied to the chart.

    The following output has been edited for brevity.

    ~(keystone_admin)]$ system helm-override-show platform-integ-apps cephfs-provisioner kube-system
    +--------------------+---------------------------------------------+
    | Property           | Value                                       |
    +--------------------+---------------------------------------------+
    | user_overrides     | storageClasses:                             |
    |                    | - additionalNamespaces:                     |
    |                    |   - default                                 |
    |                    |   - kube-public                             |
    |                    |   - new-app                                 |
    |                    |   - new-app2                                |
    |                    |   - new-app3                                |
    |                    |   chunk_size: 64                            |
    |                    |   claim_root: /pvc-volumes                  |
    |                    |   crush_rule_name: storage_tier_ruleset     |
    |                    |   data_pool_name: kube-cephfs-data          |
    |                    |   fs_name: kube-cephfs                      |
    |                    |   metadata_pool_name: kube-cephfs-metadata  |
    |                    |   name: cephfs                              |
    |                    |   replication: 2                            |
    |                    |   userId: ceph-pool-kube-cephfs-data        |
    |                    |   userSecretName: ceph-pool-kube-cephfs-data|
    +--------------------+---------------------------------------------+
    
  6. Apply the overrides.

    1. Run the application-apply command.

      ~(keystone_admin)]$ system application-apply platform-integ-apps
      +---------------+--------------------------------------+
      | Property      | Value                                |
      +---------------+--------------------------------------+
      | active        | True                                 |
      | app_version   | 1.0-62                               |
      | created_at    | 2022-12-14T04:14:08.878186+00:00     |
      | manifest_file | fluxcd-manifests                     |
      | manifest_name | platform-integ-apps-fluxcd-manifests |
      | name          | platform-integ-apps                  |
      | progress      | None                                 |
      | status        | applying                             |
      | updated_at    | 2022-12-14T04:58:58.543295+00:00     |
      +---------------+--------------------------------------+
      
    2. Monitor progress using the application-list command.

      ~(keystone_admin)]$ system application-list
      +--------------------------+---------+-------------------------------------------+------------------+----------+-----------+
      | application              | version | manifest name                             | manifest file    | status   | progress  |
      +--------------------------+---------+-------------------------------------------+------------------+----------+-----------+
      | platform-integ-apps      | 1.0-62  | platform-integ-apps-fluxcd-manifests      | fluxcd-manifests | applied  | completed |
      +--------------------------+---------+-------------------------------------------+------------------+----------+-----------+
      

    You can now create and mount PVCs from the default RBD provisioner’s general storage class, from within these application-specific namespaces.