i am facig a ceph-osd issue on our Openstack infra. We upgraded from stein to train (ceph mimic to nautilaus) since then the ceph cluster is showing BlueFS spillover warning and performance went down a lot. Irealised that since the ceph cluster deployment we misconfigured the Bluefs db and wal sizes 1GB and 100 Mo). By reading different articles apparently this should be 30GB. The only clean way i came up with is to destroy each osd at time and recreate it.
This is 4 node cluster with each 12 disks (4 NVMe and 8 spinning). 1 NVMe is dedicated for journaling on each node. Ceph-osd and Nova-Compute are collocated.
Before starting the process and since this is a charmed deployment any of you guys have an experience doing that ?
if we run :
- juju run-action --wait ceph-osd/0 osd-out
- juju remove-unit ceph-osd/0
juju add-unit ceph-osd --to 12
Is this it ? or am i missing something ? will this affect nova-compute ?
I would greatly appreciate any feedback on this !