1.环境准备
服务器配置
四台服务器使用Centos7.6操作系统,admin为管理节点 192.168.30.15 admin 192.168.30.16 storage1 192.168.30.17 storage2 192.168.30.18 storage3
admin节点配置yum仓库配置
vim /etc/yum.repos.d/ceph.repo [Ceph] name=Ceph packages for $basearch baseurl=https://mirrors.aliyun.com/ceph/rpm-mimic/el7/x86_64/ enabled=1 gpgcheck=1 type=rpm-md gpgkey=https://download.ceph.com/keys/release.asc priority=1 [Ceph-noarch] name=Ceph noarch packages baseurl=https://mirrors.aliyun.com/ceph/rpm-mimic/el7/noarch/ enabled=1 gpgcheck=1 type=rpm-md gpgkey=https://download.ceph.com/keys/release.asc priority=1 [ceph-source] name=Ceph source packages baseurl=https://mirrors.aliyun.com/ceph/rpm-mimic/el7/SRPMS/ enabled=1 gpgcheck=1 type=rpm-md gpgkey=https://download.ceph.com/keys/release.asc priority=1
hosts添加解析
vim /etc/hosts 192.168.30.15 admin 192.168.30.16 storage1 192.168.30.17 storage2 192.168.30.18 storage3
配置ssh免密码登陆
ssh-keygen ssh-copy-id storage1 ssh-copy-id storage2 ssh-copy-id storage3
2.部署ceph集群
#更新软件源并安装ceph-deploy管理工具 yum clean all && yum list yum install python-setuptools ceph-deploy -y #创建集群配置文件目录 mkdir /etc/ceph && cd /etc/ceph #初始化monitor节点,准备创建集群 ceph-deploy new admin #配置ceph.conf配置文件,默认副本数为3,修改副本数改为2 osd_pool_default_size = 2 #安装ceph软件 ceph-deploy install admin storage1 storage2 storage3 #生成monitor检测集群使用的秘钥 ceph-deploy mon create-initial #分发配置文件到集群每个节点 ceph-deploy admin storage1 storage2 storage3 #配置mgr,用于管理集群 ceph-deploy mgr storage1 storage2 storage3 #使用ceph -s命令进行验证,集群搭建完成 [root@admin ceph]# ceph -s cluster: id: eae1fd09-7410-446a-bb50-08717bc335ee health: HEALTH_OK services: mon: 1 daemons, quorum admin mgr: storage1(active), standbys: storage2, storage3 osd: 3 osds: 3 up, 3 in rgw: 3 daemons active data: pools: 6 pools, 288 pgs objects: 221 objects, 2.2 KiB usage: 3.0 GiB used, 57 GiB / 60 GiB avail pgs: 288 active+clean #部署rgw用来提供对象存储 ceph-deploy rgw storage1 storage2 storage3 #部署mds用来提供cephfs【暂时未使用到】 ceph-deploy mds create storage1 storage2 storage3 #向集群添加osd ceph-deploy osd create storage1 --data /dev/sdb ceph-deploy osd create storage2 --data /dev/sdb ceph-deploy osd create storage3 --data /dev/sdb #使用ceph -s命令查看osd状态
3.挂载cephfs
挂载: ceph-fuse -m 192.168.30.15:6789 /opt/ -c ceph.client.admin.keyring 卸载: umount /opt/