From a14fee99d8c235706d19f48280b93ff0f62347a8 Mon Sep 17 00:00:00 2001 From: Mateo Florido Date: Fri, 17 Jan 2025 13:53:28 -0500 Subject: [PATCH] docs: Fix snap storage --- docs/src/snap/howto/storage/ceph.md | 6 +++--- docs/src/snap/howto/storage/cloud.md | 4 ++-- docs/src/snap/howto/storage/storage.md | 5 +++++ 3 files changed, 10 insertions(+), 5 deletions(-) diff --git a/docs/src/snap/howto/storage/ceph.md b/docs/src/snap/howto/storage/ceph.md index 5d5f33e42..baa4affe7 100644 --- a/docs/src/snap/howto/storage/ceph.md +++ b/docs/src/snap/howto/storage/ceph.md @@ -90,7 +90,7 @@ data: config.json: |- [ { - "clusterID": "fsid 6d5c12c9-6dfb-445a-940f-301aa7de0f29", + "clusterID": "6d5c12c9-6dfb-445a-940f-301aa7de0f29", "monitors": [ "10.0.0.136:6789", ] @@ -200,7 +200,7 @@ Create the ceph-csi provisioner and node plugins: ``` wget https://raw.githubusercontent.com/ceph/ceph-csi/master/deploy/rbd/kubernetes/csi-rbdplugin-provisioner.yaml -: + kubectl apply -f csi-rbdplugin-provisioner.yaml wget https://raw.githubusercontent.com/ceph/ceph-csi/master/deploy/rbd/kubernetes/csi-rbdplugin.yaml @@ -227,7 +227,7 @@ metadata: name: csi-rbd-sc provisioner: rbd.csi.ceph.com parameters: - clusterID: b9127830-b0cc-4e34-aa47-9d1a2e9949a8 + clusterID: 6d5c12c9-6dfb-445a-940f-301aa7de0f29 pool: kubernetes imageFeatures: layering csi.storage.k8s.io/provisioner-secret-name: csi-rbd-secret diff --git a/docs/src/snap/howto/storage/cloud.md b/docs/src/snap/howto/storage/cloud.md index 365333a69..bbe309b9c 100644 --- a/docs/src/snap/howto/storage/cloud.md +++ b/docs/src/snap/howto/storage/cloud.md @@ -20,7 +20,7 @@ AWS APIs. The policies provided here are quite open and should be scoped down based on your security requirements. You will most likely want to create a role for your instance. You can call this -role "k8s-control-plane" or "k8s-worker". Then, define and attach the following +role `k8s-control-plane` or `k8s-worker`. Then, define and attach the following policies to the role. Once the role is created with the required policies, attach the role to the instance. @@ -131,7 +131,7 @@ For a worker node: A cluster using the AWS cloud provider needs to label existing nodes and resources with a ClusterID or the kube-controller-manager will not start. Add the following tag to your instance, making sure to replace the placeholder id -with your own (this can simply be "k8s" or "my-k8s-cluster"). +with your own (this can simply be `k8s` or `my-k8s-cluster`). ``` kubernetes.io/cluster/=owned diff --git a/docs/src/snap/howto/storage/storage.md b/docs/src/snap/howto/storage/storage.md index 667abc3f4..d526de393 100644 --- a/docs/src/snap/howto/storage/storage.md +++ b/docs/src/snap/howto/storage/storage.md @@ -4,6 +4,11 @@ cluster, especially for single-node support. This guide walks you through enabling and configuring this feature. +```{warning} +Local storage will not survive node removal, so it may not be suitable for +certain setups, such as production multi-node clusters. +``` + ## Prerequisites This guide assumes the following: