Quantcast
Channel: Proxmox Support Forum
Viewing all articles
Browse latest Browse all 171654

Ceph remain in HEALTH_WARN after osd remove

$
0
0
I have a cluster of 3 servers with cepth storage over 9 disks (3 each server).


One osd is going down/out and so I "remove" it, after that system start to rebalance data over the remainig osd but after some hours rebalance is stopping with 1 page stuck unclean:


Code:

    # ceph -s
    cluster 79796df2-0dc6-4a2d-8c63-5be76c25c12b
    health HEALTH_WARN 1 pgs backfilling; 1 pgs stuck unclean; recovery 710/1068853 objects degraded (0.066%)
    monmap e3: 3 mons at {0=10.10.10.1:6789/0,1=10.10.10.2:6789/0,2=10.10.10.3:6789/0}, election epoch 50, quorum 0,1,2 0,1,2
    osdmap e1251: 8 osds: 8 up, 8 in
      pgmap v16360830: 1216 pgs, 5 pools, 1383 GB data, 347 kobjects
            4048 GB used, 2652 GB / 6701 GB avail
            710/1068853 objects degraded (0.066%)
                1215 active+clean
                  1 active+remapped+backfilling
  client io 34820 kB/s rd, 126 kB/s wr, 335 op/s

This the osd tree:


Code:


    #ceph osd tree
    # id    weight    type name    up/down    reweight
    -1    6.56    root default
    -2    2.46        host proxmox00
    0    0.82            osd.0    up    1   
    1    0.82            osd.1    up    1   
    2    0.82            osd.2    up    1   
    -3    2.46        host proxmox01
    3    0.82            osd.3    up    1   
    4    0.82            osd.4    up    1   
    5    0.82            osd.5    up    1   
    -4    1.64        host proxmox02
    6    0.82            osd.6    up    1   
    8    0.82            osd.8    up    1

Because the cluster does not return in the state HEALT_OK?

Viewing all articles
Browse latest Browse all 171654

Trending Articles



<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>