前言:
目前我们对“centos安装parted”都比较关注,同学们都想要知道一些“centos安装parted”的相关内容。那么小编同时在网上收集了一些关于“centos安装parted””的相关文章,希望我们能喜欢,同学们快快来学习一下吧!一、ceph安装1 前期准备
本次安装环境为:
ceph1(集群命令分发管控,提供磁盘服务集群)
CentOS7.5
192.168.3.61
ceph2(提供磁盘服务集群)
CentOs7.5
192.168.3.62
ceph3(提供磁盘服务集群)
CentOs7.5
192.168.3.63
2 编辑hosts文件,增加以下内容
# ceph1-master and ceph1-osd1192.168.3.61 ceph1# ceph2-osd2192.168.3.62 ceph2# ceph3-osd3192.168.3.63 ceph33 ceph是通ssh下发指令。
首先配置管理节点可以无密码访问存储节点,生成密钥,并建立touch authorized_keys文件(在ceph1操作)
ssh-keygen -t rsa # 生成公钥及私钥touch /root/.ssh/authorized_keyscat /root/.ssh/id_rsa.pub > /root/.ssh/authorized_keyschmod 700 /root/.sshchmod 644 /root/.ssh/authorized_keys# 将密钥拷贝到其他主机ssh-copy-id ceph2ssh-copy-id ceph3# 关闭防火墙及安全选项,所有节点egrep -v "^$|^#" /etc/selinux/configsystemctl stop firewalldsystemctl disable firewalld4 ceph各节点安装ceph-deploy
(各节点都要安装)
安装EPEL软件包仓库,并安装
yum install -y wgetmv /etc/yum.repos.d/CentOS-Base.repo /etc/yum.repos.d/CentOS-Base.repo.bakwget -O /etc/yum.repos.d/CentOS-Base.repo -Uvh
替换ceph.repo 服务器
sed -i 's#htt.*://download.ceph.com#; /etc/yum.repos.d/ceph.repo
安装 ceph-deploy和python-setuptools
yum updateyum install ceph-deploy --disablerepo=epelyum install python-setuptools5 安装NTP服务器
centos7默认的ntp是chronyd,系统默认自己安装,只需要把ceph1改成阿里云ntp,其他机器改成ceph1并启动服务。
ceph1:
vi /etc/chrony.conf# Use public servers from the pool.ntp.org project.# Please consider joining the pool ().#server 0.centos.pool.ntp.org iburst#server 1.centos.pool.ntp.org iburst#server 2.centos.pool.ntp.org iburst#server 3.centos.pool.ntp.org iburstserver ntp1.aliyun.com iburst.......allow 192.168.3.0/24
ceph2,ceph3:
vi /etc/chrony.conf# Use public servers from the pool.ntp.org project.# Please consider joining the pool ().#server 0.centos.pool.ntp.org iburst#server 1.centos.pool.ntp.org iburst#server 2.centos.pool.ntp.org iburst#server 3.centos.pool.ntp.org iburstserver ceph1 iburst........
所有节点,重启ntp服务:
systemctl restart chronyd systemctl enable chronyd6 搭建ceph mon
在ceph-master节点上创建ceph-install目录
mkdir /ceph-install && cd /ceph-install
在ceph-master的ceph-install上创建MON节点。
ceph-deploy new ceph1 ceph2 ceph3
新创建的集群被自动命名为ceph,集群创建完成后,在my-cluster目录中会产生ceph配置文件、monitor秘钥文件和log文件,具体如下:
ls .ceph.conf ceph-deploy-ceph.log ceph.mon.keyring
ceph.conf默认配置如下:
[root@ceph1 ceph-install]# cat ceph.conf[global]fsid = 319d2cce-b087-4de9-bd4a-13edc7644abcmon_initial_members = ceph1, ceph2, ceph3mon_host = 192.168.3.61,192.168.3.62,192.168.3.63auth_cluster_required = cephxauth_service_required = cephxauth_client_required = cephx
在ceph1用ceph-deploy安装,注意ceph-deploy已经默认安装N版本了,将repo环境变量更改成阿里云的地址再安装
export CEPH_DEPLOY_REPO_URL= CEPH_DEPLOY_GPG_URL=
在ceph1进入ceph-install目录,用deploy安装
ceph-deploy install ceph1 ceph2 ceph3
初始化monitors节点并收集keys
[root@ceph1 ceph-install]# ceph-deploy mon create-initial
初始化完成后,在ceph-install中将会新增多个秘钥文件,具体如下:
[root@ceph1 ceph-install]# lsceph.bootstrap-mds.keyring ceph.bootstrap-osd.keyring ceph.client.admin.keyring ceph-deploy-ceph.logceph.bootstrap-mgr.keyring ceph.bootstrap-rgw.keyring ceph.conf ceph.mon.keyring
将ceph-install目录的keyring拷贝到/etc/ceph/下面,其他节点也拷贝过去。
cp -p /ceph-install/*keyring /etc/ceph/scp /ceph-install/*keyring root@ceph2:/etc/ceph/scp /ceph-install/*keyring root@ceph3:/etc/ceph/
查看集群健康情况
# ceph -s cluster: id: 319d2cce-b087-4de9-bd4a-13edc7644abc health: HEALTH_OK services: mon: 3 daemons, quorum ceph1,ceph2,ceph3 mgr: no daemons active osd: 0 osds: 0 up, 0 in data: pools: 0 pools, 0 pgs objects: 0 objects, 0B usage: 0B used, 0B / 0B avail pgs:7 创建OSD存储节点
每个节点都执行nvme分区:
# nvme0n1作为journal盘,画成4个分区parted /dev/nvme0n1 mklabel gptparted /dev/nvme0n1 mkpart primary 0% 25%parted /dev/nvme0n1 mkpart primary 26% 50%parted /dev/nvme0n1 mkpart primary 51% 75%parted /dev/nvme0n1 mkpart primary 76% 100%
显示如下:
[root@ceph2 ~]# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTsda 8:0 0 200G 0 disk ├─sda1 8:1 0 1G 0 part /boot└─sda2 8:2 0 153.9G 0 part ├─centos-root 253:0 0 150G 0 lvm / └─centos-swap 253:1 0 3.9G 0 lvm [SWAP]sdb 8:16 0 300G 0 disk sdc 8:32 0 300G 0 disk sdd 8:48 0 300G 0 disk nvme0n1 259:0 0 100G 0 disk ├─nvme0n1p1 259:1 0 25G 0 part ├─nvme0n1p2 259:2 0 24G 0 part ├─nvme0n1p3 259:3 0 24G 0 part └─nvme0n1p4 259:4 0 24G 0 part
创建osd,在ceph1执行
ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdb --journal /dev/nvme0n1p1 ceph1ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdc --journal /dev/nvme0n1p2 ceph1ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdd --journal /dev/nvme0n1p3 ceph1ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdb --journal /dev/nvme0n1p1 ceph2ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdc --journal /dev/nvme0n1p2 ceph2ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdd --journal /dev/nvme0n1p3 ceph2ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdb --journal /dev/nvme0n1p1 ceph3ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdc --journal /dev/nvme0n1p2 ceph3ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdd --journal /dev/nvme0n1p3 ceph3
把配置文件和admin 秘钥到管理节点和ceph节点
ceph-deploy --overwrite-conf admin ceph1 ceph2 ceph3
安装mgr :
ceph-deploy mgr create ceph1 ceph2 ceph3
执行命令查询:
# ceph -s cluster: id: 319d2cce-b087-4de9-bd4a-13edc7644abc health: HEALTH_OK services: mon: 3 daemons, quorum ceph1,ceph2,ceph3 mgr: ceph1(active), standbys: ceph3, ceph2 osd: 9 osds: 9 up, 9 in data: pools: 0 pools, 0 pgs objects: 0 objects, 0B usage: 968MiB used, 2.63TiB / 2.64TiB avail pgs: # ceph osd treeID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 2.63507 root default -3 0.87836 host ceph1 0 hdd 0.29279 osd.0 up 1.00000 1.00000 1 hdd 0.29279 osd.1 up 1.00000 1.00000 2 hdd 0.29279 osd.2 up 1.00000 1.00000 -5 0.87836 host ceph2 3 hdd 0.29279 osd.3 up 1.00000 1.00000 4 hdd 0.29279 osd.4 up 1.00000 1.00000 5 hdd 0.29279 osd.5 up 1.00000 1.00000 -7 0.87836 host ceph3 6 hdd 0.29279 osd.6 up 1.00000 1.00000 7 hdd 0.29279 osd.7 up 1.00000 1.00000 8 hdd 0.29279 osd.8 up 1.00000 1.000008 ceph1整体记录
[root@ceph1 ceph-install]# history 1 ip add 2 poweroff 3 ip add 4 ip add 5 cd /etc/sysconfig/network-scripts/ 6 ls -ls 7 mv ifcfg-ens3 ifcfg-eth0 8 vi ifcfg-eth0 9 vi /etc/default/grub 10 grub2-mkconfig -o /boot/grub2/grub.cfg 11 reboot 12 ip add 13 poweroff 14 ip add 15 lsblk 16 poweroff 17 hostnamectl set-hostname ceph1 18 exit 19 ip add 20 vi /etc/sysconfig/network-scripts/ifcfg-eth0 21 ifdown eth0 && ifup eth0 22 vi /etc/hosts 23 scp /etc/hosts root@ceph2:/etc/hosts 24 scp /etc/hosts root@ceph3:/etc/hosts 25 ping ceph1 26 ping ceph2 27 ping ceph3 28 ssh-keygen -t rsa 29 touch /root/.ssh/authorized_keys 30 cat id_rsa.pub > authorized_keys 31 cat /root/.ssh/id_rsa.pub > /root/.ssh/authorized_keys 32 chmod 700 /root/.ssh 33 chmod 644 /root/.ssh/authorized_keys 34 ssh-copy-id ceph2 35 ssh-copy-id ceph3 36 ssh ceph1 37 ssh ceph2 38 ssh ceph3 39 egrep -v "^$|^#" /etc/selinux/config 40 systemctl stop firewalld 41 systemctl disable firewalld 42 lsblk 43 poweroff 44 exit 45 cd /etc/yum.repos.d/ 46 ls 47 cd .. 48 vi /etc/yum.conf 49 cd /etc/yum.repos.d/ 50 ls 51 mv /etc/yum.repos.d/CentOS-Base.repo /etc/yum.repos.d/CentOS-Base.repo.bak 52 wget -O /etc/yum.repos.d/CentOS-Base.repo 53 mv /etc/yum.repos.d/CentOS-Base.repo.bak /etc/yum.repos.d/CentOS-Base.repo 54 yum install -y wget 55 mv /etc/yum.repos.d/CentOS-Base.repo /etc/yum.repos.d/CentOS-Base.repo.bak 56 wget -O /etc/yum.repos.d/CentOS-Base.repo 57 rpm -Uvh 58 cat ceph.repo 59 sed -i 's#htt.*://download.ceph.com#; /etc/yum.repos.d/ceph.repo 60 cat ceph.repo 61 ssh ceph1 62 ssh ceph2 63 ssh ceph3 64 systemctl status chronyd.service 65 vi /etc/chrony.conf 66 systemctl restart chronyd 67 systemctl enable chronyd 68 chronyc sources 69 vi /etc/chrony.conf 70 systemctl restart chronyd 71 chronyc sources 72 vi /etc/chrony.conf 73 systemctl restart chronyd 74 vi /etc/chrony.conf 75 systemctl restart chronyd 76 chronyc sources 77 systemctl status firewalld 78 yum install -y ansible 79 ls 80 cat CentOS-Base.repo 81 vi epel.repo 82 yum install -y ansible 83 vi /etc/chrony.conf 84 systemctl restart chronyd 85 cd /root 86 cd /etc/ansible/ 87 ls 88 vi hosts 89 ls 90 cp -p hosts hosts.bak 91 echo 1 > hosts 92 vi hosts 93 ansible -i hosts -m ping all 94 ansible -i hosts -m shell -a 'date' 95 ansible -i hosts -m shell -a 'date' all 96 chronyc sources 97 systemctl restart chronyd 98 cd /etc/yum.repos.d/ 99 ls 100 scp epel.repo root@:/etc/yum.repos.d/ 101 scp epel.repo root@ceph2:/etc/yum.repos.d/ 102 scp epel.repo root@ceph3:/etc/yum.repos.d/ 103 yum update 104 yum install ceph-deploy --disablerepo=epel 105 yum install python-setuptools -y 106 mkdir /ceph-install && cd /ceph-install 107 ceph-deploy new ceph1 ceph2 ceph3 108 ls 109 cat ceph.conf 110 export CEPH_DEPLOY_REPO_URL= 111 export CEPH_DEPLOY_GPG_URL= 112 ceph-deploy install ceph1 ceph2 ceph3 113 ceph-deploy mon create-initial 114 ls 115 cp -p /ceph-install/*keyring /etc/ceph/ 116 scp /ceph-install/*keyring root@ceph2:/etc/ceph/ 117 scp /ceph-install/*keyring root@ceph3:/etc/ceph/ 118 ceph -s 119 cat ceph.conf 120 netstat -lpnt 121 yum search netstat 122 ss -lpnt 123 yum install -y net-tools.x86_64 124 yum search netstat 125 netstat -lpnt 126 poweroff 127 ceph -s 128 ls / 129 cd /ceph-install/ 130 ls 131 ceph-deploy osd create --help 132 lsblk 133 parted /dev/nvme0n1 mklabel gpt 134 parted /dev/nvme0n1 mkpart primary 0% 25% 135 lsblk 136 cat /etc/fstab 137 ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdb --journal /dev/nvme0n1p1 ceph1 138 lsblk 139 cd /var/lib/ceph/osd/ceph-0 140 ls 141 history 142 ceph osd tree 143 parted /dev/nvme0n1 mkpart primary 26% 50% 144 parted /dev/nvme0n1 mkpart primary 51% 75% 145 lsblk 146 parted /dev/nvme0n1 mkpart primary 76% 100% 147 lsblk 148 cd / 149 cd /ceph-install/ 150 ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdc --journal /dev/nvme0n1p2 ceph1 151 ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdd --journal /dev/nvme0n1p3 ceph1 152 ceph -s 153 ceph-deploy --overwrite-conf admin ceph1 ceph2 ceph3 154 ceph-deploy mgr create ceph1 ceph2 ceph3 155 ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdb --journal /dev/nvme0n1p1 ceph2 156 ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdc --journal /dev/nvme0n1p2 ceph2 157 ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdd --journal /dev/nvme0n1p3 ceph2 158 ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdb --journal /dev/nvme0n1p1 ceph3 159 ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdc --journal /dev/nvme0n1p2 ceph3 160 ceph-deploy osd create --filestore --fs-type xfs --data /dev/sdd --journal /dev/nvme0n1p3 ceph3 161 lsblk 162 ceph -s 163 ceph osd tree 164 history9 优化配置文件ceph.conf并下发
cat ceph.conf [global]fsid = 319d2cce-b087-4de9-bd4a-13edc7644abcmon_initial_members = ceph1, ceph2, ceph3mon_host = 192.168.3.61,192.168.3.62,192.168.3.63auth_cluster_required = cephxauth_service_required = cephxauth_client_required = cephxmon max pg per osd = 3000mon_allow_pool_delete = truerbd_default_features = 1public_network = 192.168.3.0/24[client]rbd cache = truerbd cache size = 536870912rbd cache max dirty = 26738688rbd cache max dirty age = 15[osd]filestore min sync interval = 10filestore max sync interval = 15filestore queue max ops = 25000filestore queue max bytes = 1048576000filestore queue committing max ops = 50000filestore queue committing max bytes = 10485760000filestore fd cache size = 1024filestore op threads = 32journal max write bytes = 1073714824journal max write entries = 10000journal queue max ops = 50000journal queue max bytes = 10485760000
下发配置:
ceph-deploy --overwrite-conf config push ceph{1..3}二、ceph集成社区版openstack1 集成cinder
(1) 在cinder-volume节点安装ceph client
yum install -y ceph-common
注意:glance要安装python-rbd
cinder-volume节点安装ceph成功后,会生成/etc/ceph/这个目录,里面有一个文件。
# ls -ls /etc/ceph/总用量 44 -rwxr-xr-x. 1 root root 2910 10月 31 2018 rbdmap
(2) ceph提前建立pool volumes,并创建cinder账号和授权;
ceph1:
# ceph osd pool create volumes 128# ceph auth get-or-create client.cinder mon 'allow r' osd 'allow class-read object_prefix rbd_children, allow rwx pool=volumes, allow rwx pool=volumes-cache, allow rwx pool=vms, allow rwx pool=vms-cache, allow rx pool=images, allow rx pool=images-cache'[client.cinder] key = AQC0Ma5favxeBRAA4qIdM04lzflv7VF7guntJQ==# ceph auth get-or-create client.glance mon 'allow r' osd 'allow class-read object_prefix rbd_children, allow rwx pool=images'[client.glance] key = AQA5Mq5f7GGqORAAu7b2EyHTjuBDUm0e45QqSg==
在ceph1 /etc/ceph/创建2个认证文件,内容为上一步auth创建最后两行的账号及密钥内容:
# cat ceph.client.cinder.keyring[client.cinder] key = AQC0Ma5favxeBRAA4qIdM04lzflv7VF7guntJQ==# cat ceph.client.glance.keyring[client.glance] key = AQA5Mq5f7GGqORAAu7b2EyHTjuBDUm0e45QqSg==
将生成的文件发送到运行glance-api和cinder-volume的节点上,glance-api通常运行在控制节点上,本例的cinder-volume也运行在控制节点上。
# scp ceph.client.cinder.keyring root@192.168.3.10:/etc/ceph/# scp ceph.client.glance.keyring root@192.168.3.10:/etc/ceph/
在控制节点和存储节点更改权限:
# chown glance:glance ceph.client.glance.keyring# chown cinder:cinder ceph.client.cinder.keyring
运行nova-compute服务节点需要用到client.cinder的秘钥文件,并需将其传递到计算节点,本例子计算与存储为同一节点,所以略。运行nova-compute节点需要将client,cinder用户的秘钥文件存储到libvirt中,当基于ceph后端的cinder卷被attach到虚拟机实例时,libvirt需要用到秘钥文件以访问ceph集群。
............
在运行nova-compute的计算节点上将秘钥文件添加到libvirt中:
# uuidgen26966ae6-d92f-43d4-b89f-c7327973df36# cat > secret.xml <<EOF> <secret ephemeral='no' private='no'>> <uuid>26966ae6-d92f-43d4-b89f-c7327973df36</uuid>> <usage type='ceph'>> <name>client.cinder secret</name>> </usage>> </secret>> EOF# virsh secret-define --file secret.xml生成 secret 26966ae6-d92f-43d4-b89f-c7327973df36virsh secret-set-value --secret 26966ae6-d92f-43d4-b89f-c7327973df36 --base64 AQC0Ma5favxeBRAA4qIdM04lzflv7VF7guntJQ==# AQC0Ma5favxeBRAA4qIdM04lzflv7VF7guntJQ==为ceph.client.cinder.keyring的key
(3) 将ceph-server服务器的ceph.conf拷贝过来cinder-volume及compute节点/etc/ceph这个目录下。
ceph1:
scp /etc/ceph/ceph.conf root@192.168.3.10:/etc/ceph/
本例子controller及compute为同一台主机。
(4) 编辑cinder-volume节点,在/etc/cinder/cinder-volume添加以下信息
[DEFAULT]enabled_backends = lvm,rbd-1glance_api_version = 2[rbd-1]volume_driver = cinder.volume.drivers.rbd.RBDDrivervolume_backend_name = rbd-1rbd_pool = volumesrbd_ceph_conf = /etc/ceph/ceph.confrbd_flatten_volume_from_snapshot = falserbd_max_clone_depth = 5rbd_store_chunk_size = 4rados_connect_timeout = 5rbd_user = cinderrbd_secret_uuid = 26966ae6-d92f-43d4-b89f-c7327973df36report_discard_supported = Trueimage_upload_use_cinder_backend = False
(5) 创建cinder type类型rbd-1
# cinder type-create rbd-1# cinder type-key rbd-1 set volume_backend_name=rbd-1
查询tpye是否生效
# openstack volume type list+--------------------------------------+-------------+-----------+| ID | Name | Is Public |+--------------------------------------+-------------+-----------+| 946d519d-a747-4c63-b083-437a3cb495e9 | rbd-1 | True || 162afe89-b5ed-4ede-b90d-02a354b8253a | __DEFAULT__ | True |+--------------------------------------+-------------+-----------+
创建ceph块存储
# openstack volume create disk1113 --type rbd-1 --size 1+---------------------+--------------------------------------+| Field | Value |+---------------------+--------------------------------------+| attachments | [] || availability_zone | nova || bootable | false || consistencygroup_id | None || created_at | 2020-11-13T07:51:46.000000 || description | None || encrypted | False || id | 4bda7201-dbd5-4274-b375-619f16a7d402 || migration_status | None || multiattach | False || name | disk1113 || properties | || replication_status | None || size | 1 || snapshot_id | None || source_volid | None || status | creating || type | rbd-1 || updated_at | None || user_id | 607785c930c44219b6462b1e2bff7dcc |+---------------------+--------------------------------------+[root@controller cinder]# openstack volume list+--------------------------------------+-----------------------------------------------------+-----------+------+---------------------------------+| ID | Name | Status | Size | Attached to |+--------------------------------------+-----------------------------------------------------+-----------+------+---------------------------------+| 4bda7201-dbd5-4274-b375-619f16a7d402 | disk1113 | available | 1 | |2 集成glance
(1) 在ceph创建pool : images
# ceph osd pool create images 128
(2) 在openstack控制节点操作
修改 /etc/glance/glance-api.conf
...........[glance_store]stores = file,http,rbd#default_store = file#filesystem_store_datadir = /var/lib/glance/images/default_store = rbdrbd_store_user = glancerbd_store_pool = imagesrbd_store_chunk_size = 8rbd_store_ceph_conf = /etc/ceph/ceph.conf............
重启glance服务
systemctl restart openstack-glance-api.service
上传镜像
# glance image-create --name cirros.raw --disk-format raw --container-format bare --file cirros.raw --store rbd --progress[=============================>] 100%+------------------+----------------------------------------------------------------------------------+| Property | Value |+------------------+----------------------------------------------------------------------------------+| checksum | ba3cd24377dde5dfdd58728894004abb || container_format | bare || created_at | 2020-11-16T06:15:49Z || disk_format | raw || id | 1430bac6-3e89-476f-b971-a4ac2e3c4083 || min_disk | 0 || min_ram | 0 || name | cirros.raw || os_hash_algo | sha512 || os_hash_value | b795f047a1b10ba0b7c95b43b2a481a59289dc4cf2e49845e60b194a911819d3ada03767bbba4143 || | b44c93fd7f66c96c5a621e28dff51d1196dae64974ce240e || os_hidden | False || owner | e2c91481258f417c83ffb4ea9d7a2339 || protected | False || size | 46137344 || status | active || tags | [] || updated_at | 2020-11-16T06:15:54Z || virtual_size | Not available || visibility | shared |+------------------+----------------------------------------------------------------------------------+
(3) ceph1查看:
# rbd ls images1430bac6-3e89-476f-b971-a4ac2e3c4083
标签: #centos安装parted