Ceph ile ilgili yazdığımız bildiriye bağlantıdan ulaşabilirsiniz.
You can access our report about Ceph from the link.
sudo chown -R ceph:ceph /var/lib/ceph sudo chmod -R 777 /var/lib/ceph /etc/rc.local sudo chown ceph:ceph /dev/journal1 sudo chown ceph:ceph /dev/journal2 sudo chown ceph:ceph /dev/journal3... chmod +x /etc/rc.local sudo reboot
Run the commandline bottom, after activated all osd on the osd-nodes.
grep osd/ceph- /proc/mounts
The output of the commandline below is like;
/dev/sdc1 /var/lib/ceph/osd/ceph-0 xfs rw,noatime,attr2,inode64,noquota 0 0 /dev/sdd1 /var/lib/ceph/osd/ceph-1 xfs rw,noatime,attr2,inode64,noquota 0 0 /dev/sde1 /var/lib/ceph/osd/ceph-2 xfs rw,noatime,attr2,inode64,noquota 0 0 ...
Add this output to /etc/fstab file. If you see all osds are up now, then reboot osd nodes. Osds will be all down and up again.
You can also try to add below codes to /etc/rc.local and reboot.
sudo start ceph-osd id=0 (id of the osd) sudo start ceph-osd id=1 sudo start ceph-osd id=2 sudo start ceph-osd id=...