Update "ceph -s" command output

Since the version of ceph bump to L, the output should
be updated.

Change-Id: I4c05b9b7423369a43dcb5855a2f0ef016ddbe488
This commit is contained in:
wu.chunyang 2018-05-09 22:06:21 +08:00
parent 3f100310ee
commit c65861c356

View File

@ -248,14 +248,20 @@ indicates a healthy cluster:
docker exec ceph_mon ceph -s docker exec ceph_mon ceph -s
cluster 5fba2fbc-551d-11e5-a8ce-01ef4c5cf93c cluster:
health HEALTH_OK id: f2ed6c00-c043-4e1c-81b6-07c512db26b1
monmap e1: 1 mons at {controller=10.0.0.128:6789/0} health: HEALTH_OK
election epoch 2, quorum 0 controller
osdmap e18: 2 osds: 2 up, 2 in services:
pgmap v27: 64 pgs, 1 pools, 0 bytes data, 0 objects mon: 1 daemons, quorum 172.16.31.121
68676 kB used, 20390 MB / 20457 MB avail mgr: poc12-01(active)
64 active+clean osd: 4 osds: 4 up, 4 in; 5 remapped pgs
data:
pools: 4 pools, 512 pgs
objects: 0 objects, 0 bytes
usage: 432 MB used, 60963 MB / 61395 MB avail
pgs: 512 active+clean
If Ceph is run in an **all-in-one** deployment or with less than three storage If Ceph is run in an **all-in-one** deployment or with less than three storage
nodes, further configuration is required. It is necessary to change the default nodes, further configuration is required. It is necessary to change the default