for i in `rados -p cephfs_metadata ls`;do rados -p cephfs_metadata rm $i; done
执行结束后再运行rados -p cephfs_metadata ls 应该没有打印信息。
运行命令查看 fs 当前状态如下
1 2 3 4 5 6 7 8 9
root@storage0:/# ceph fs status cephfs - 0 clients ====== RANK STATE MDS ACTIVITY DNS INOS DIRS CAPS 0 failed POOL TYPE USED AVAIL cephfs_metadata metadata 0 12.2T cephfs_data data 8299M 12.2T cephfs_recovery - 0 clients
看到 metadata 部分 USED 大小为 0,说明元数据已经全部丢失了。
此时运行 ceph -s 发现集群状态并没有马上发生改变,只用当 mds service 停止一个或者 fs fail 时,状态才会发生改变。
root@storage0:/home/admin# ceph mds stat cephfs:0/1 2 up:standby, 1 damaged
下一步将尝试进行恢复,在已有数据池没有问题的情况下。
2. 创建备份文件系统
这里的目的是从 data pool 中恢复一个可用的备份文件系统,两个 FS 只有 Metadata pool 是不相同的。
首先将现有的文件系统停止,避免过程中对数据池发生进一步改写。
1 2 3 4
root@storage0:/home/admin# ceph fs ls name: cephfs, metadata pool: cephfs_metadata, data pools: [cephfs_data ] root@storage0:/home/admin# ceph fs fail cephfs cephfs marked not joinable; MDS cannot join the cluster. All MDS ranks marked failed.
接下来创建一个恢复文件系统,其中元数据是新建的用来存储恢复的元数据,数据池仍然是原有的数据池。
1 2 3 4 5 6 7
root@storage0:/home/admin# ceph osd pool create cephfs_recovery_meta 256 pool 'cephfs_recovery_meta' created root@storage0:/home/admin# ceph fs new cephfs_recovery cephfs_recovery_meta cephfs_data --recover --allow-dangerous-metadata-overlay Pool 'cephfs_data' (id '1') has pg autoscale mode 'on' but is not marked as bulk. Consider setting the flag by running # ceph osd pool set cephfs_data bulk true new fs with metadata pool 3 and data pool 1
此时运行命令可以看到两个 fs 的状态
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
root@storage0:/# ceph fs status cephfs - 0 clients ====== RANK STATE MDS ACTIVITY DNS INOS DIRS CAPS 0 failed POOL TYPE USED AVAIL cephfs_metadata metadata 0 12.2T cephfs_data data 8299M 12.2T cephfs_recovery - 0 clients =============== RANK STATE MDS ACTIVITY DNS INOS DIRS CAPS 0 active cephfs.storage0.dkvwoh Reqs: 0 /s 11 14 12 0 POOL TYPE USED AVAIL cephfs_recovery_meta metadata 96.0k 11.8T cephfs_data data 8299M 12.2T STANDBY MDS cephfs.storage1.qcyalm MDS version: ceph version 18.2.0 (5dd24139a1eada541a3bc16b6941c5dde975e26d) reef (stable)
3. 恢复原 FS
首先确保整个过程文件系统处于关闭状态
1 2 3 4
root@storage0:/# ceph fs fail cephfs cephfs marked not joinable; MDS cannot join the cluster. All MDS ranks marked failed. root@storage0:/# ceph fs set cephfs joinable false cephfs marked not joinable; MDS cannot join as newly active.
root@storage0:/# ceph fs set cephfs joinable true cephfs marked joinable; MDS may join as newly active. root@storage0:/# ceph tell mds.0 scrub start / recursive repair 2023-12-07T01:14:44.711+0000 7f9631ffb700 0 client.17146 ms_handle_reset on v2:10.0.1.3:6848/641236932 2023-12-07T01:14:44.751+0000 7f9631ffb700 0 client.17150 ms_handle_reset on v2:10.0.1.3:6848/641236932 { "return_code": 0, "scrub_tag": "ae8e7687-5bb6-40a6-b999-dbf1b6d88f84", "mode": "asynchronous" }
此时再看集群状态发现原集群状态已经恢复
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
root@storage0:/# ceph fs status cephfs - 1 clients ====== RANK STATE MDS ACTIVITY DNS INOS DIRS CAPS 0 active cephfs.flexbj-storage1.qcyalm Reqs: 0 /s 12 15 13 1 POOL TYPE USED AVAIL cephfs_metadata metadata 112k 12.2T cephfs_data data 8299M 12.2T STANDBY MDS =============== RANK STATE MDS ACTIVITY DNS INOS DIRS CAPS 0 active cephfs.storage0.dkvwoh Reqs: 0 /s 11 14 12 0 POOL TYPE USED AVAIL cephfs_recovery_meta metadata 96.0k 11.8T cephfs_data data 8299M 12.2T MDS version: ceph version 18.2.0 (5dd24139a1eada541a3bc16b6941c5dde975e26d) reef (stable)
4. 删除备份 FS
运行如下命令进行删除即可
1 2 3 4 5
root@storage0:/# ceph fs fail cephfs_recovery cephfs_recovery marked not joinable; MDS cannot join the cluster. All MDS ranks marked failed. root@storage0:/# ceph fs rm cephfs_recovery --yes-i-really-mean-it root@storage0:/# ceph osd pool rm cephfs_recovery_meta cephfs_recovery_meta --yes-i-really-really-mean-it pool 'cephfs_recovery_meta' removed