canonical / microceph

Ceph for a one-rack cluster and appliances

Home Page:https://snapcraft.io/microceph

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

second MDS fails to deploy

slapcat opened this issue · comments

Creating one filesystem works fine, but when a second is created, it remains in a degraded/offline state because an MDS fails to come up for it:

# ceph -s
  cluster:
    id:     05e0f755-cb2a-410e-b565-98793ae95192
    health: HEALTH_ERR
            1 filesystem is offline
            1 filesystem is online with fewer MDS than max_mds

After updating both filesystems to have max_mds = 2, the status changes and the count of MDS increases to 2, but notes 1 is failed:

  cluster:
    id:     05e0f755-cb2a-410e-b565-98793ae95192
    health: HEALTH_ERR
            1 filesystem is degraded
            1 filesystem has a failed mds daemon
            1 filesystem is offline
            2 filesystems are online with fewer MDS than max_mds
 
  services:
    mon: 1 daemons, quorum meatwad (age 6m)
    mgr: meatwad(active, since 6m)
    mds: 1/2 daemons up (1 failed)

What is the correct way to increase the MDS count and/or use multiple filesystems with Microceph?

Hey @slapcat! It appears you only have 1 MicroCeph node in the ceph cluster. Adding one more node would automatically spin up extra instances of mon/mgr/mds trio.

Thanks for the explanation! I don't plan on adding more nodes at the moment, so I guess I can use one cephfs with subvolumes instead.

Feel free to close this issue if you don't need it.

@slapcat you can also spin-up an lxd container to add as another Microceph node. 😁