Forums

Home / Forums

You need to log in to create posts and topics. Login · Register

My MON Service crash

one of my mon service is crash.

probably db corruption:
mon / MonitorDBStore.h: 306: FAILED assert (0 == "failed to write to db")

This server is mon and osd at the same time

I would like to know the best practice for changing a mon.
I unplug and reinstall?

Did I remove osd discs before?

This mon02, if I create a new server and call it mon05, and try
add it as new mon server, does the petasan cluster accept?
It's only 3 mon servers, right?

 


 

um de meus mon service is crash.

provavelmente db corrupcao:
mon/MonitorDBStore.h: 306: FAILED assert(0 == "failed to write to db")

Esse servidor e mon e osd ao mesmo tempo

Gostaria de saber qual a melhor pratica para trocar um mon.
Desligo e reinstalo?

Removo os discos do osd antes?

Esse mon o mon02, se eu criar um novo servidor e chamalo de mon05, e tentar
adicionar ele como novo servidor mon, o cluster petasan aceita?
Sao apenas 3 mon servers, correto?

For v1.5 and 2.0

Replace the OS disk
Install using installer, you need to specify the same hostname, same management interface, same management ip of what you had specified before.
When deploying the node: choose "Replace Management Node", join any of the 2 existing management nodes, click next, next
When done, reboot the system to restart old OSDs (this is a bug)

done.

but the new node only enters the cluster when
it's clean / ok, right?

View post on imgur.com


root@san01:~# ceph -w --cluster cloud
cluster 9f99e76f-1f50-4aa3-b876-dbf194a3cadf
health HEALTH_WARN
676 pgs backfill_wait
3 pgs backfilling
679 pgs degraded
679 pgs stuck unclean
679 pgs undersized
recovery 229202/698840 objects degraded (32.797%)
recovery 194120/698840 objects misplaced (27.777%)
1 mons down, quorum 0,2 san01,san03
monmap e3: 3 mons at {san01=10.0.10.1:6789/0,san02=10.0.10.2:6789/0,san03=10.0.10.3:6789/0}
election epoch 882, quorum 0,2 san01,san03
osdmap e36286: 11 osds: 8 up, 8 in; 679 remapped pgs
flags sortbitwise,require_jewel_osds
pgmap v20312235: 1000 pgs, 1 pools, 1314 GB data, 330 kobjects
1740 GB used, 8176 GB / 9916 GB avail
229202/698840 objects degraded (32.797%)
194120/698840 objects misplaced (27.777%)
676 active+undersized+degraded+remapped+wait_backfill
321 active+clean
3 active+undersized+degraded+remapped+backfilling
recovery io 112 MB/s, 29 objects/s
client io 3973 kB/s rd, 468 kB/s wr, 31 op/s rd, 98 op/s wr

----------------------------------------------------------------------------------

 

The deployment should complete quickly. Are you using 1.5 or 2,0 ?

1.5.

 

The cluster is OK.

 

thanks