Docs: update ceph troubleshooting with luminious data
This PS updates the output from the troubleshooting guide to reflect the expected output and commands for ceph luminious. Change-Id: I8351d782d3b1ddbf2fc121b7096548e1ba38f7bd
This commit is contained in:
parent
cc84b98da5
commit
52cb1c4f43
@ -16,17 +16,27 @@ can be done with the following Ceph command:
|
|||||||
|
|
||||||
::
|
::
|
||||||
|
|
||||||
admin@kubenode01:~$ kubectl exec -t -i ceph-mon-0 -n ceph -- ceph status
|
admin@kubenode01:~$ MON_POD=$(kubectl get --no-headers pods -n=ceph -l="application=ceph,component=mon" | awk '{ print $1; exit }')
|
||||||
cluster 046de582-f8ee-4352-9ed4-19de673deba0
|
admin@kubenode01:~$ kubectl exec -n ceph ${MON_POD} -- ceph -s
|
||||||
health HEALTH_OK
|
cluster:
|
||||||
monmap e3: 3 mons at {ceph-mon-392438295-6q04c=10.25.65.131:6789/0,ceph-mon-392438295-ksrb2=10.25.49.196:6789/0,ceph-mon-392438295-l0pzj=10.25.79.193:6789/0}
|
id: 06a191c7-81bd-43f3-b5dd-3d6c6666af71
|
||||||
election epoch 6, quorum 0,1,2 ceph-mon-392438295-ksrb2,ceph-mon-392438295-6q04c,ceph-mon-392438295-l0pzj
|
health: HEALTH_OK
|
||||||
fsmap e5: 1/1/1 up {0=mds-ceph-mds-2810413505-gtjgv=up:active}
|
|
||||||
osdmap e23: 5 osds: 5 up, 5 in
|
services:
|
||||||
flags sortbitwise
|
mon: 1 daemons, quorum att.port.direct
|
||||||
pgmap v22012: 80 pgs, 3 pools, 12712 MB data, 3314 objects
|
mgr: att.port.direct(active)
|
||||||
101 GB used, 1973 GB / 2186 GB avail
|
mds: cephfs-1/1/1 up {0=mds-ceph-mds-68c9c76d59-zqc55=up:active}
|
||||||
80 active+clean
|
osd: 1 osds: 1 up, 1 in
|
||||||
|
rgw: 1 daemon active
|
||||||
|
|
||||||
|
data:
|
||||||
|
pools: 11 pools, 208 pgs
|
||||||
|
objects: 352 objects, 464 MB
|
||||||
|
usage: 62467 MB used, 112 GB / 173 GB avail
|
||||||
|
pgs: 208 active+clean
|
||||||
|
|
||||||
|
io:
|
||||||
|
client: 253 B/s rd, 39502 B/s wr, 1 op/s rd, 8 op/s wr
|
||||||
admin@kubenode01:~$
|
admin@kubenode01:~$
|
||||||
|
|
||||||
Use one of your Ceph Monitors to check the status of the cluster. A
|
Use one of your Ceph Monitors to check the status of the cluster. A
|
||||||
@ -43,10 +53,9 @@ To verify that your deployment namespace has a client key:
|
|||||||
|
|
||||||
::
|
::
|
||||||
|
|
||||||
admin@kubenode01: $ kubectl get secret -n openstack
|
admin@kubenode01: $ kubectl get secret -n openstack pvc-ceph-client-key
|
||||||
NAME TYPE DATA AGE
|
NAME TYPE DATA AGE
|
||||||
default-token-nvl10 kubernetes.io/service-account-token 3 7d
|
pvc-ceph-client-key kubernetes.io/rbd 1 8h
|
||||||
pvc-ceph-client-key kubernetes.io/rbd 1 6m
|
|
||||||
|
|
||||||
Without this, your RBD-backed PVCs will never reach the ``Bound`` state. For
|
Without this, your RBD-backed PVCs will never reach the ``Bound`` state. For
|
||||||
more information, see how to `activate namespace for ceph <../install/multinode.html#activating-control-plane-namespace-for-ceph>`_.
|
more information, see how to `activate namespace for ceph <../install/multinode.html#activating-control-plane-namespace-for-ceph>`_.
|
||||||
@ -70,15 +79,16 @@ correctly:
|
|||||||
::
|
::
|
||||||
|
|
||||||
admin@kubenode01:~$ kubectl describe storageclass/general
|
admin@kubenode01:~$ kubectl describe storageclass/general
|
||||||
Name: general
|
Name: general
|
||||||
IsDefaultClass: No
|
IsDefaultClass: No
|
||||||
Annotations: <none>
|
Annotations: <none>
|
||||||
Provisioner: kubernetes.io/rbd
|
Provisioner: ceph.com/rbd
|
||||||
Parameters: adminId=admin,adminSecretName=pvc-ceph-conf-combined-storageclass,adminSecretNamespace=ceph,monitors=ceph-mon.ceph:6789,pool=rbd,userId=admin,userSecretName=pvc-ceph-client-key
|
Parameters: adminId=admin,adminSecretName=pvc-ceph-conf-combined-storageclass,adminSecretNamespace=ceph,imageFeatures=layering,imageFormat=2,monitors=ceph-mon.ceph.svc.cluster.local:6789,pool=rbd,userId=admin,userSecretName=pvc-ceph-client-key
|
||||||
No events.
|
ReclaimPolicy: Delete
|
||||||
|
Events: <none>
|
||||||
admin@kubenode01:~$
|
admin@kubenode01:~$
|
||||||
|
|
||||||
The parameters are what we're looking for here. If we see parameters
|
The parameters are what we're looking for here. If we see parameters
|
||||||
passed to the StorageClass correctly, we will see the
|
passed to the StorageClass correctly, we will see the
|
||||||
``ceph-mon.ceph:6789`` hostname/port, things like ``userid``, and
|
``ceph-mon.ceph.svc.cluster.local:6789`` hostname/port, things like ``userid``,
|
||||||
appropriate secrets used for volume claims.
|
and appropriate secrets used for volume claims.
|
||||||
|
Loading…
Reference in New Issue
Block a user