Docs: update ceph troubleshooting with luminious data

This PS updates the output from the troubleshooting guide to reflect
the expected output and commands for ceph luminious.

Change-Id: I8351d782d3b1ddbf2fc121b7096548e1ba38f7bd
This commit is contained in:
portdirect 2017-12-28 09:23:55 -05:00
parent cc84b98da5
commit 52cb1c4f43

View File

@ -16,17 +16,27 @@ can be done with the following Ceph command:
:: ::
admin@kubenode01:~$ kubectl exec -t -i ceph-mon-0 -n ceph -- ceph status admin@kubenode01:~$ MON_POD=$(kubectl get --no-headers pods -n=ceph -l="application=ceph,component=mon" | awk '{ print $1; exit }')
cluster 046de582-f8ee-4352-9ed4-19de673deba0 admin@kubenode01:~$ kubectl exec -n ceph ${MON_POD} -- ceph -s
health HEALTH_OK cluster:
monmap e3: 3 mons at {ceph-mon-392438295-6q04c=10.25.65.131:6789/0,ceph-mon-392438295-ksrb2=10.25.49.196:6789/0,ceph-mon-392438295-l0pzj=10.25.79.193:6789/0} id: 06a191c7-81bd-43f3-b5dd-3d6c6666af71
election epoch 6, quorum 0,1,2 ceph-mon-392438295-ksrb2,ceph-mon-392438295-6q04c,ceph-mon-392438295-l0pzj health: HEALTH_OK
fsmap e5: 1/1/1 up {0=mds-ceph-mds-2810413505-gtjgv=up:active}
osdmap e23: 5 osds: 5 up, 5 in services:
flags sortbitwise mon: 1 daemons, quorum att.port.direct
pgmap v22012: 80 pgs, 3 pools, 12712 MB data, 3314 objects mgr: att.port.direct(active)
101 GB used, 1973 GB / 2186 GB avail mds: cephfs-1/1/1 up {0=mds-ceph-mds-68c9c76d59-zqc55=up:active}
80 active+clean osd: 1 osds: 1 up, 1 in
rgw: 1 daemon active
data:
pools: 11 pools, 208 pgs
objects: 352 objects, 464 MB
usage: 62467 MB used, 112 GB / 173 GB avail
pgs: 208 active+clean
io:
client: 253 B/s rd, 39502 B/s wr, 1 op/s rd, 8 op/s wr
admin@kubenode01:~$ admin@kubenode01:~$
Use one of your Ceph Monitors to check the status of the cluster. A Use one of your Ceph Monitors to check the status of the cluster. A
@ -43,10 +53,9 @@ To verify that your deployment namespace has a client key:
:: ::
admin@kubenode01: $ kubectl get secret -n openstack admin@kubenode01: $ kubectl get secret -n openstack pvc-ceph-client-key
NAME TYPE DATA AGE NAME TYPE DATA AGE
default-token-nvl10 kubernetes.io/service-account-token 3 7d pvc-ceph-client-key kubernetes.io/rbd 1 8h
pvc-ceph-client-key kubernetes.io/rbd 1 6m
Without this, your RBD-backed PVCs will never reach the ``Bound`` state. For Without this, your RBD-backed PVCs will never reach the ``Bound`` state. For
more information, see how to `activate namespace for ceph <../install/multinode.html#activating-control-plane-namespace-for-ceph>`_. more information, see how to `activate namespace for ceph <../install/multinode.html#activating-control-plane-namespace-for-ceph>`_.
@ -70,15 +79,16 @@ correctly:
:: ::
admin@kubenode01:~$ kubectl describe storageclass/general admin@kubenode01:~$ kubectl describe storageclass/general
Name: general Name: general
IsDefaultClass: No IsDefaultClass: No
Annotations: <none> Annotations: <none>
Provisioner: kubernetes.io/rbd Provisioner: ceph.com/rbd
Parameters: adminId=admin,adminSecretName=pvc-ceph-conf-combined-storageclass,adminSecretNamespace=ceph,monitors=ceph-mon.ceph:6789,pool=rbd,userId=admin,userSecretName=pvc-ceph-client-key Parameters: adminId=admin,adminSecretName=pvc-ceph-conf-combined-storageclass,adminSecretNamespace=ceph,imageFeatures=layering,imageFormat=2,monitors=ceph-mon.ceph.svc.cluster.local:6789,pool=rbd,userId=admin,userSecretName=pvc-ceph-client-key
No events. ReclaimPolicy: Delete
Events: <none>
admin@kubenode01:~$ admin@kubenode01:~$
The parameters are what we're looking for here. If we see parameters The parameters are what we're looking for here. If we see parameters
passed to the StorageClass correctly, we will see the passed to the StorageClass correctly, we will see the
``ceph-mon.ceph:6789`` hostname/port, things like ``userid``, and ``ceph-mon.ceph.svc.cluster.local:6789`` hostname/port, things like ``userid``,
appropriate secrets used for volume claims. and appropriate secrets used for volume claims.