ceph create file system and mount
ceph create file system
개요
- ceph version 은 15.2 입니다.
create replicated pool
ceph orch apply mds cephfs --placement="ceph0,ceph1,ceph2"
FS_NAME=cephfs_repl_data
FS_META=cephfs_repl_meta
PG_CNT=8
ceph osd pool create ${FS_NAME} ${PG_CNT}
ceph osd pool create ${FS_META} ${PG_CNT}
ceph fs new cephfs ${FS_META} ${FS_NAME}
[root@ceph-test mnt]# ceph fs status
cephfs - 0 clients
======
RANK STATE MDS ACTIVITY DNS INOS
0 active cephfs.ceph1.witvio Reqs: 0 /s 10 13
POOL TYPE USED AVAIL
cephfs_repl_meta metadata 1536k 410G
cephfs_repl_data data 0 410G
STANDBY MDS
cephfs.ceph2.iplnaa
cephfs.ceph0.fpvsgg
MDS version: ceph version 15.2.3 (d289bbdec69ed7c1f516e0a093594580a76b78d0) octopus (stable)
[root@ceph-test mnt]# ceph auth get client.admin | grep key
exported keyring for client.admin
key = AQDk/NleJfSOExAAKHtFWBmEDdNCLc/WGLFUaQ==
mount.ceph ceph1,ceph0,ceph2:/ /mnt/cephfs-repl -o name=admin,secret=AQDk/NleJfSOExAAKHtFWBmEDdNCLc/WGLFUaQ==
erasure code pool
- ceph 의 file system 은 multi file system 을 지원하지 않는다.
- 사실 테스트 해보지를 않았다.
# deploy mds
ceph orch apply mds cephfs --placement="ceph0,ceph1,ceph2"
FS_NAME=cephfs_era_data
FS_META=cephfs_era_meta
PG_CNT=8
ceph osd pool create ${FS_NAME} erasure
ceph osd pool create ${FS_META} ${PG_CNT}
ceph osd pool create cephfs_data ${PG_CNT}
# ceph fs 의 erasure code pool 을 사용하기 위하여는 해당 옵션을 enable 하여야 한다.
ceph osd pool set ${FS_NAME} allow_ec_overwrites true
ceph osd pool application enable ${FS_NAME} cephfs
ceph fs new cephfs ${FS_META} ${FS_NAME} --force
# ceph status 의 mds 의 active 가 앞으로 오게한다.
mount.ceph ceph1,ceph0,ceph2:/ /mnt/cephfs-era -o name=admin,secret=AQDk/NleJfSOExAAKHtFWBmEDdNCLc/WGLFUaQ==
Last modified on 2020-06-07