ceph-deploy bclinux aarch64 ceph 14.2.10
ssh-copy-id,部署机免密登录其他三台主机
所有机器硬盘配置参考如下,计划采用vdb作为ceph数据盘
下载ceph-deploy
pip install ceph-deploy
免密登录+设置主机名
hostnamectl --static set-hostname ceph-0 .. 3
配置hosts
172.17.163.105 ceph-0
172.17.112.206 ceph-1
172.17.227.100 ceph-2
172.17.67.157 ceph-3
scp /etc/hosts root@ceph-1:/etc/hostsscp /etc/hosts root@ceph-2:/etc/hostsscp /etc/hosts root@ceph-3:/etc/hosts
本机先安装各软件包
rpm -ivhU liboath/liboath-*
过滤掉安装不上的包
find rpmbuild/RPMS/ | grep \\.rpm | grep -v debug | grep -v k8s | grep -v mgr\-rook | grep -v mgr\-ssh | xargs -i echo "{} \\"
添加用户,安装软件
useradd ceph
yum install -y rpmbuild/RPMS/noarch/ceph-mgr-dashboard-14.2.10-0.oe1.bclinux.noarch.rpm \
rpmbuild/RPMS/noarch/ceph-mgr-diskprediction-cloud-14.2.10-0.oe1.bclinux.noarch.rpm \
rpmbuild/RPMS/noarch/ceph-grafana-dashboards-14.2.10-0.oe1.bclinux.noarch.rpm \
rpmbuild/RPMS/noarch/ceph-mgr-diskprediction-local-14.2.10-0.oe1.bclinux.noarch.rpm \
rpmbuild/RPMS/aarch64/librgw-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librados-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-base-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/rbd-mirror-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-ceph-argparse-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librbd-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-test-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/rados-objclass-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-rbd-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-mds-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-fuse-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-rbd-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-mgr-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librbd1-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-cephfs-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/libcephfs2-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-rgw-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/libradospp-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/rbd-nbd-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/libcephfs-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-mon-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-radosgw-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-ceph-argparse-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-cephfs-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-rados-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/rbd-fuse-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librgw2-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-rgw-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-common-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librados2-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-ceph-compat-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-osd-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-rados-14.2.10-0.oe1.bclinux.aarch64.rpm
安装成功日志(截图是二次reinstall)
注意:如未提前创建用户,警报
分发编译好的rpm包+el8的 liboath
ceph-0
rsync -avr -P liboath root@ceph-1:~/
rsync -avr -P liboath root@ceph-2:~/
rsync -avr -P liboath root@ceph-3:~/
rsync -avr -P rpmbuild/RPMS root@ceph-1:~/rpmbuild/
rsync -avr -P rpmbuild/RPMS root@ceph-2:~/rpmbuild/
rsync -avr -P rpmbuild/RPMS root@ceph-3:~/rpmbuild/
分别登录ceph-1、ceph-2、ceph-3执行(后续可以考虑ansible封装)
cd ~
rpm -ivhU liboath/liboath-*
useradd ceph
yum install -y rpmbuild/RPMS/noarch/ceph-mgr-dashboard-14.2.10-0.oe1.bclinux.noarch.rpm \
rpmbuild/RPMS/noarch/ceph-mgr-diskprediction-cloud-14.2.10-0.oe1.bclinux.noarch.rpm \
rpmbuild/RPMS/noarch/ceph-grafana-dashboards-14.2.10-0.oe1.bclinux.noarch.rpm \
rpmbuild/RPMS/noarch/ceph-mgr-diskprediction-local-14.2.10-0.oe1.bclinux.noarch.rpm \
rpmbuild/RPMS/aarch64/librgw-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librados-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-base-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/rbd-mirror-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-ceph-argparse-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librbd-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-test-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/rados-objclass-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-rbd-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-mds-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-fuse-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-rbd-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-mgr-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librbd1-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-cephfs-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/libcephfs2-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-rgw-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/libradospp-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/rbd-nbd-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/libcephfs-devel-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-mon-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-radosgw-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-ceph-argparse-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-cephfs-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-rados-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/rbd-fuse-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librgw2-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python3-rgw-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-common-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/librados2-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-ceph-compat-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/ceph-osd-14.2.10-0.oe1.bclinux.aarch64.rpm \
rpmbuild/RPMS/aarch64/python-rados-14.2.10-0.oe1.bclinux.aarch64.rpm
安装日志
时间同步ntpdate
ceph-0
yum install ntpdate
编辑/etc/ntp.conf
driftfile /var/lib/ntp/drift
restrict default nomodify notrap nopeer noepeer noquery
restrict source nomodify notrap noepeer noquery
restrict 127.0.0.1
restrict ::1
tos maxclock 5
includefile /etc/ntp/crypto/pw
keys /etc/ntp/keys
server asia.pool.ntp.org
启动ntpd
systemctl enable ntpd --now
ceph-1 ceph-2 ceph-3
yum install -y ntpdate
编辑/etc/ntp.conf
driftfile /var/lib/ntp/drift
restrict default nomodify notrap nopeer noepeer noquery
restrict source nomodify notrap noepeer noquery
restrict 127.0.0.1
restrict ::1
tos maxclock 5
includefile /etc/ntp/crypto/pw
keys /etc/ntp/keys
server ceph-0
启动ntpd
systemctl enable ntpd --now
部署mon节点,生成ceph.conf
cd /etc/ceph/
ceph-deploy new ceph-0 ceph-1 ceph-2 ceph-3
报错,如下
[ceph_deploy][ERROR ] UnsupportedPlatform: Platform is not supported: bclinux 21.10U3 LTS 21.10U3
修改/usr/lib/python2.7/site-packages/ceph_deploy/calamari.py,新增一个bclinux 差异如下
[root@ceph-0 ceph_deploy]# diff calamari.py calamari.py.bak -Npr
*** calamari.py 2023-11-10 16:56:49.445013228 +0800
--- calamari.py.bak 2023-11-10 16:56:14.793013228 +0800
*************** def distro_is_supported(distro_name):
*** 13,19 ****An enforcer of supported distros that can differ from what ceph-deploysupports."""
! supported = ['centos', 'redhat', 'ubuntu', 'debian', 'bclinux']if distro_name in supported:return Truereturn False
--- 13,19 ----An enforcer of supported distros that can differ from what ceph-deploysupports."""
! supported = ['centos', 'redhat', 'ubuntu', 'debian']if distro_name in supported:return Truereturn False
修改/usr/lib/python2.7/site-packages/ceph_deploy/hosts/__init__.py
[root@ceph-0 ceph_deploy]# diff -Npr hosts/__init__.py hosts/__init__.py.bak
*** hosts/__init__.py 2023-11-10 17:06:27.585013228 +0800
--- hosts/__init__.py.bak 2023-11-10 17:05:48.697013228 +0800
*************** def _get_distro(distro, fallback=None, u
*** 101,107 ****'fedora': fedora,'suse': suse,'virtuozzo': centos,
- 'bclinux': centos,'arch': arch}--- 101,106 ----
ceph-deploy new ceph-0 ceph-1 ceph-2 ceph-3
成功生成ceph.conf,过程日志
[root@ceph-0 ceph]# ceph-deploy new ceph-0 ceph-1 ceph-2 ceph-3
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy new ceph-0 ceph-1 ceph-2 ceph-3
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0xffffb246c280>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] ssh_copykey : True
[ceph_deploy.cli][INFO ] mon : ['ceph-0', 'ceph-1', 'ceph-2', 'ceph-3']
[ceph_deploy.cli][INFO ] func : <function new at 0xffffb236e9d0>
[ceph_deploy.cli][INFO ] public_network : None
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] cluster_network : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.cli][INFO ] fsid : None
[ceph_deploy.new][DEBUG ] Creating new cluster named ceph
[ceph_deploy.new][INFO ] making sure passwordless SSH succeeds
[ceph-0][DEBUG ] connected to host: ceph-0
[ceph-0][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-0][DEBUG ] detect machine type
[ceph-0][DEBUG ] find the location of an executable
[ceph-0][INFO ] Running command: /usr/sbin/ip link show
[ceph-0][INFO ] Running command: /usr/sbin/ip addr show
[ceph-0][DEBUG ] IP addresses found: [u'172.18.0.1', u'172.17.163.105']
[ceph_deploy.new][DEBUG ] Resolving host ceph-0
[ceph_deploy.new][DEBUG ] Monitor ceph-0 at 172.17.163.105
[ceph_deploy.new][INFO ] making sure passwordless SSH succeeds
[ceph-1][DEBUG ] connected to host: ceph-0
[ceph-1][INFO ] Running command: ssh -CT -o BatchMode=yes ceph-1
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-1][DEBUG ] connected to host: ceph-1
[ceph-1][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-1][DEBUG ] detect machine type
[ceph-1][DEBUG ] find the location of an executable
[ceph-1][INFO ] Running command: /usr/sbin/ip link show
[ceph-1][INFO ] Running command: /usr/sbin/ip addr show
[ceph-1][DEBUG ] IP addresses found: [u'172.17.112.206']
[ceph_deploy.new][DEBUG ] Resolving host ceph-1
[ceph_deploy.new][DEBUG ] Monitor ceph-1 at 172.17.112.206
[ceph_deploy.new][INFO ] making sure passwordless SSH succeeds
[ceph-2][DEBUG ] connected to host: ceph-0
[ceph-2][INFO ] Running command: ssh -CT -o BatchMode=yes ceph-2
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-2][DEBUG ] connected to host: ceph-2
[ceph-2][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-2][DEBUG ] detect machine type
[ceph-2][DEBUG ] find the location of an executable
[ceph-2][INFO ] Running command: /usr/sbin/ip link show
[ceph-2][INFO ] Running command: /usr/sbin/ip addr show
[ceph-2][DEBUG ] IP addresses found: [u'172.17.227.100']
[ceph_deploy.new][DEBUG ] Resolving host ceph-2
[ceph_deploy.new][DEBUG ] Monitor ceph-2 at 172.17.227.100
[ceph_deploy.new][INFO ] making sure passwordless SSH succeeds
[ceph-3][DEBUG ] connected to host: ceph-0
[ceph-3][INFO ] Running command: ssh -CT -o BatchMode=yes ceph-3
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-3][DEBUG ] connected to host: ceph-3
[ceph-3][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-3][DEBUG ] detect machine type
[ceph-3][DEBUG ] find the location of an executable
[ceph-3][INFO ] Running command: /usr/sbin/ip link show
[ceph-3][INFO ] Running command: /usr/sbin/ip addr show
[ceph-3][DEBUG ] IP addresses found: [u'172.17.67.157']
[ceph_deploy.new][DEBUG ] Resolving host ceph-3
[ceph_deploy.new][DEBUG ] Monitor ceph-3 at 172.17.67.157
[ceph_deploy.new][DEBUG ] Monitor initial members are ['ceph-0', 'ceph-1', 'ceph-2', 'ceph-3']
[ceph_deploy.new][DEBUG ] Monitor addrs are ['172.17.163.105', '172.17.112.206', '172.17.227.100', '172.17.67.157']
[ceph_deploy.new][DEBUG ] Creating a random mon key...
[ceph_deploy.new][DEBUG ] Writing monitor keyring to ceph.mon.keyring...
[ceph_deploy.new][DEBUG ] Writing initial config to ceph.conf...
自动生成的/etc/ceph/ceph.conf内容如下
[global]
fsid = ff72b496-d036-4f1b-b2ad-55358f3c16cb
mon_initial_members = ceph-0, ceph-1, ceph-2, ceph-3
mon_host = 172.17.163.105,172.17.112.206,172.17.227.100,172.17.67.157
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
由于只是测试环境,且未挂第二个网络,暂时不设置public_network参数
部署monitor
cd /etc/ceph
ceph-deploy mon create ceph-0 ceph-1 ceph-2 ceph-3
故障
[ceph-3][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-3.asok mon_status
[ceph-3][ERROR ] Traceback (most recent call last):
[ceph-3][ERROR ] File "/bin/ceph", line 151, in <module>
[ceph-3][ERROR ] from ceph_daemon import admin_socket, DaemonWatcher, Termsize
[ceph-3][ERROR ] File "/usr/lib/python2.7/site-packages/ceph_daemon.py", line 24, in <module>
[ceph-3][ERROR ] from prettytable import PrettyTable, HEADER
[ceph-3][ERROR ] ImportError: No module named prettytable
[ceph-3][WARNIN] monitor: mon.ceph-3, might not be running yet
[ceph-3][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-3.asok mon_status
[ceph-3][ERROR ] Traceback (most recent call last):
[ceph-3][ERROR ] File "/bin/ceph", line 151, in <module>
[ceph-3][ERROR ] from ceph_daemon import admin_socket, DaemonWatcher, Termsize
[ceph-3][ERROR ] File "/usr/lib/python2.7/site-packages/ceph_daemon.py", line 24, in <module>
[ceph-3][ERROR ] from prettytable import PrettyTable, HEADER
[ceph-3][ERROR ] ImportError: No module named prettytable
[ceph-3][WARNIN] monitor ceph-3 does not exist in monmap
[ceph-3][WARNIN] neither `public_addr` nor `public_network` keys are defined for monitors
[ceph-3][WARNIN] monitors may not be able to form quorum
No module named prettytable
pip install PrettyTable
下载离线包
分发到其他三台机器上离线安装
rsync -avr -P preetytable-python27 root@ceph-1:~/
rsync -avr -P preetytable-python27 root@ceph-2:~/
rsync -avr -P preetytable-python27 root@ceph-3:~/
再次部署monitor
cd /etc/ceph
ceph-deploy mon create ceph-0 ceph-1 ceph-2 ceph-3
日志记录
[root@ceph-0 ~]# cd /etc/ceph
[root@ceph-0 ceph]# ceph-deploy mon create ceph-0 ceph-1 ceph-2 ceph-3
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy mon create ceph-0 ceph-1 ceph-2 ceph-3
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] subcommand : create
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0xffff992fb320>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] mon : ['ceph-0', 'ceph-1', 'ceph-2', 'ceph-3']
[ceph_deploy.cli][INFO ] func : <function mon at 0xffff993967d0>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] keyrings : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts ceph-0 ceph-1 ceph-2 ceph-3
[ceph_deploy.mon][DEBUG ] detecting platform for host ceph-0 ...
[ceph-0][DEBUG ] connected to host: ceph-0
[ceph-0][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-0][DEBUG ] detect machine type
[ceph-0][DEBUG ] find the location of an executable
[ceph_deploy.mon][INFO ] distro info: bclinux 21.10U3 21.10U3 LTS
[ceph-0][DEBUG ] determining if provided host has same hostname in remote
[ceph-0][DEBUG ] get remote short hostname
[ceph-0][DEBUG ] deploying mon to ceph-0
[ceph-0][DEBUG ] get remote short hostname
[ceph-0][DEBUG ] remote hostname: ceph-0
[ceph-0][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-0][DEBUG ] create the mon path if it does not exist
[ceph-0][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-ceph-0/done
[ceph-0][DEBUG ] create a done file to avoid re-doing the mon deployment
[ceph-0][DEBUG ] create the init path if it does not exist
[ceph-0][INFO ] Running command: systemctl enable ceph.target
[ceph-0][INFO ] Running command: systemctl enable ceph-mon@ceph-0
[ceph-0][INFO ] Running command: systemctl start ceph-mon@ceph-0
[ceph-0][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-0.asok mon_status
[ceph-0][DEBUG ] ********************************************************************************
[ceph-0][DEBUG ] status for monitor: mon.ceph-0
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "election_epoch": 8,
[ceph-0][DEBUG ] "extra_probe_peers": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addrvec": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.67.157:3300",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v2"
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.67.157:6789",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v1"
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addrvec": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.112.206:3300",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v2"
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.112.206:6789",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v1"
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addrvec": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.227.100:3300",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v2"
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.227.100:6789",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v1"
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ],
[ceph-0][DEBUG ] "feature_map": {
[ceph-0][DEBUG ] "mon": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "features": "0x3ffddff8ffacffff",
[ceph-0][DEBUG ] "num": 1,
[ceph-0][DEBUG ] "release": "luminous"
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] "features": {
[ceph-0][DEBUG ] "quorum_con": "4611087854031667199",
[ceph-0][DEBUG ] "quorum_mon": [
[ceph-0][DEBUG ] "kraken",
[ceph-0][DEBUG ] "luminous",
[ceph-0][DEBUG ] "mimic",
[ceph-0][DEBUG ] "osdmap-prune",
[ceph-0][DEBUG ] "nautilus"
[ceph-0][DEBUG ] ],
[ceph-0][DEBUG ] "required_con": "2449958747315912708",
[ceph-0][DEBUG ] "required_mon": [
[ceph-0][DEBUG ] "kraken",
[ceph-0][DEBUG ] "luminous",
[ceph-0][DEBUG ] "mimic",
[ceph-0][DEBUG ] "osdmap-prune",
[ceph-0][DEBUG ] "nautilus"
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] "monmap": {
[ceph-0][DEBUG ] "created": "2023-11-11 09:54:05.372287",
[ceph-0][DEBUG ] "epoch": 1,
[ceph-0][DEBUG ] "features": {
[ceph-0][DEBUG ] "optional": [],
[ceph-0][DEBUG ] "persistent": [
[ceph-0][DEBUG ] "kraken",
[ceph-0][DEBUG ] "luminous",
[ceph-0][DEBUG ] "mimic",
[ceph-0][DEBUG ] "osdmap-prune",
[ceph-0][DEBUG ] "nautilus"
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] "fsid": "ff72b496-d036-4f1b-b2ad-55358f3c16cb",
[ceph-0][DEBUG ] "min_mon_release": 14,
[ceph-0][DEBUG ] "min_mon_release_name": "nautilus",
[ceph-0][DEBUG ] "modified": "2023-11-11 09:54:05.372287",
[ceph-0][DEBUG ] "mons": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.67.157:6789/0",
[ceph-0][DEBUG ] "name": "ceph-3",
[ceph-0][DEBUG ] "public_addr": "172.17.67.157:6789/0",
[ceph-0][DEBUG ] "public_addrs": {
[ceph-0][DEBUG ] "addrvec": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.67.157:3300",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v2"
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.67.157:6789",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v1"
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] "rank": 0
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.112.206:6789/0",
[ceph-0][DEBUG ] "name": "ceph-1",
[ceph-0][DEBUG ] "public_addr": "172.17.112.206:6789/0",
[ceph-0][DEBUG ] "public_addrs": {
[ceph-0][DEBUG ] "addrvec": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.112.206:3300",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v2"
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.112.206:6789",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v1"
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] "rank": 1
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.163.105:6789/0",
[ceph-0][DEBUG ] "name": "ceph-0",
[ceph-0][DEBUG ] "public_addr": "172.17.163.105:6789/0",
[ceph-0][DEBUG ] "public_addrs": {
[ceph-0][DEBUG ] "addrvec": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.163.105:3300",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v2"
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.163.105:6789",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v1"
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] "rank": 2
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.227.100:6789/0",
[ceph-0][DEBUG ] "name": "ceph-2",
[ceph-0][DEBUG ] "public_addr": "172.17.227.100:6789/0",
[ceph-0][DEBUG ] "public_addrs": {
[ceph-0][DEBUG ] "addrvec": [
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.227.100:3300",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v2"
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] {
[ceph-0][DEBUG ] "addr": "172.17.227.100:6789",
[ceph-0][DEBUG ] "nonce": 0,
[ceph-0][DEBUG ] "type": "v1"
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] "rank": 3
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ]
[ceph-0][DEBUG ] },
[ceph-0][DEBUG ] "name": "ceph-0",
[ceph-0][DEBUG ] "outside_quorum": [],
[ceph-0][DEBUG ] "quorum": [
[ceph-0][DEBUG ] 0,
[ceph-0][DEBUG ] 1,
[ceph-0][DEBUG ] 2,
[ceph-0][DEBUG ] 3
[ceph-0][DEBUG ] ],
[ceph-0][DEBUG ] "quorum_age": 917,
[ceph-0][DEBUG ] "rank": 2,
[ceph-0][DEBUG ] "state": "peon",
[ceph-0][DEBUG ] "sync_provider": []
[ceph-0][DEBUG ] }
[ceph-0][DEBUG ] ********************************************************************************
[ceph-0][INFO ] monitor: mon.ceph-0 is running
[ceph-0][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-0.asok mon_status
[ceph_deploy.mon][DEBUG ] detecting platform for host ceph-1 ...
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-1][DEBUG ] connected to host: ceph-1
[ceph-1][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-1][DEBUG ] detect machine type
[ceph-1][DEBUG ] find the location of an executable
[ceph_deploy.mon][INFO ] distro info: bclinux 21.10U3 21.10U3 LTS
[ceph-1][DEBUG ] determining if provided host has same hostname in remote
[ceph-1][DEBUG ] get remote short hostname
[ceph-1][DEBUG ] deploying mon to ceph-1
[ceph-1][DEBUG ] get remote short hostname
[ceph-1][DEBUG ] remote hostname: ceph-1
[ceph-1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-1][DEBUG ] create the mon path if it does not exist
[ceph-1][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-ceph-1/done
[ceph-1][DEBUG ] create a done file to avoid re-doing the mon deployment
[ceph-1][DEBUG ] create the init path if it does not exist
[ceph-1][INFO ] Running command: systemctl enable ceph.target
[ceph-1][INFO ] Running command: systemctl enable ceph-mon@ceph-1
[ceph-1][INFO ] Running command: systemctl start ceph-mon@ceph-1
[ceph-1][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-1.asok mon_status
[ceph-1][DEBUG ] ********************************************************************************
[ceph-1][DEBUG ] status for monitor: mon.ceph-1
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "election_epoch": 8,
[ceph-1][DEBUG ] "extra_probe_peers": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addrvec": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.67.157:3300",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v2"
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.67.157:6789",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v1"
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addrvec": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.163.105:3300",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v2"
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.163.105:6789",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v1"
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addrvec": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.227.100:3300",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v2"
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.227.100:6789",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v1"
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ],
[ceph-1][DEBUG ] "feature_map": {
[ceph-1][DEBUG ] "mon": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "features": "0x3ffddff8ffacffff",
[ceph-1][DEBUG ] "num": 1,
[ceph-1][DEBUG ] "release": "luminous"
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] "features": {
[ceph-1][DEBUG ] "quorum_con": "4611087854031667199",
[ceph-1][DEBUG ] "quorum_mon": [
[ceph-1][DEBUG ] "kraken",
[ceph-1][DEBUG ] "luminous",
[ceph-1][DEBUG ] "mimic",
[ceph-1][DEBUG ] "osdmap-prune",
[ceph-1][DEBUG ] "nautilus"
[ceph-1][DEBUG ] ],
[ceph-1][DEBUG ] "required_con": "2449958747315912708",
[ceph-1][DEBUG ] "required_mon": [
[ceph-1][DEBUG ] "kraken",
[ceph-1][DEBUG ] "luminous",
[ceph-1][DEBUG ] "mimic",
[ceph-1][DEBUG ] "osdmap-prune",
[ceph-1][DEBUG ] "nautilus"
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] "monmap": {
[ceph-1][DEBUG ] "created": "2023-11-11 09:54:05.372287",
[ceph-1][DEBUG ] "epoch": 1,
[ceph-1][DEBUG ] "features": {
[ceph-1][DEBUG ] "optional": [],
[ceph-1][DEBUG ] "persistent": [
[ceph-1][DEBUG ] "kraken",
[ceph-1][DEBUG ] "luminous",
[ceph-1][DEBUG ] "mimic",
[ceph-1][DEBUG ] "osdmap-prune",
[ceph-1][DEBUG ] "nautilus"
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] "fsid": "ff72b496-d036-4f1b-b2ad-55358f3c16cb",
[ceph-1][DEBUG ] "min_mon_release": 14,
[ceph-1][DEBUG ] "min_mon_release_name": "nautilus",
[ceph-1][DEBUG ] "modified": "2023-11-11 09:54:05.372287",
[ceph-1][DEBUG ] "mons": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.67.157:6789/0",
[ceph-1][DEBUG ] "name": "ceph-3",
[ceph-1][DEBUG ] "public_addr": "172.17.67.157:6789/0",
[ceph-1][DEBUG ] "public_addrs": {
[ceph-1][DEBUG ] "addrvec": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.67.157:3300",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v2"
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.67.157:6789",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v1"
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] "rank": 0
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.112.206:6789/0",
[ceph-1][DEBUG ] "name": "ceph-1",
[ceph-1][DEBUG ] "public_addr": "172.17.112.206:6789/0",
[ceph-1][DEBUG ] "public_addrs": {
[ceph-1][DEBUG ] "addrvec": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.112.206:3300",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v2"
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.112.206:6789",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v1"
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] "rank": 1
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.163.105:6789/0",
[ceph-1][DEBUG ] "name": "ceph-0",
[ceph-1][DEBUG ] "public_addr": "172.17.163.105:6789/0",
[ceph-1][DEBUG ] "public_addrs": {
[ceph-1][DEBUG ] "addrvec": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.163.105:3300",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v2"
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.163.105:6789",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v1"
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] "rank": 2
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.227.100:6789/0",
[ceph-1][DEBUG ] "name": "ceph-2",
[ceph-1][DEBUG ] "public_addr": "172.17.227.100:6789/0",
[ceph-1][DEBUG ] "public_addrs": {
[ceph-1][DEBUG ] "addrvec": [
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.227.100:3300",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v2"
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] {
[ceph-1][DEBUG ] "addr": "172.17.227.100:6789",
[ceph-1][DEBUG ] "nonce": 0,
[ceph-1][DEBUG ] "type": "v1"
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] "rank": 3
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ]
[ceph-1][DEBUG ] },
[ceph-1][DEBUG ] "name": "ceph-1",
[ceph-1][DEBUG ] "outside_quorum": [],
[ceph-1][DEBUG ] "quorum": [
[ceph-1][DEBUG ] 0,
[ceph-1][DEBUG ] 1,
[ceph-1][DEBUG ] 2,
[ceph-1][DEBUG ] 3
[ceph-1][DEBUG ] ],
[ceph-1][DEBUG ] "quorum_age": 921,
[ceph-1][DEBUG ] "rank": 1,
[ceph-1][DEBUG ] "state": "peon",
[ceph-1][DEBUG ] "sync_provider": []
[ceph-1][DEBUG ] }
[ceph-1][DEBUG ] ********************************************************************************
[ceph-1][INFO ] monitor: mon.ceph-1 is running
[ceph-1][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-1.asok mon_status
[ceph_deploy.mon][DEBUG ] detecting platform for host ceph-2 ...
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-2][DEBUG ] connected to host: ceph-2
[ceph-2][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-2][DEBUG ] detect machine type
[ceph-2][DEBUG ] find the location of an executable
[ceph_deploy.mon][INFO ] distro info: bclinux 21.10U3 21.10U3 LTS
[ceph-2][DEBUG ] determining if provided host has same hostname in remote
[ceph-2][DEBUG ] get remote short hostname
[ceph-2][DEBUG ] deploying mon to ceph-2
[ceph-2][DEBUG ] get remote short hostname
[ceph-2][DEBUG ] remote hostname: ceph-2
[ceph-2][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-2][DEBUG ] create the mon path if it does not exist
[ceph-2][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-ceph-2/done
[ceph-2][DEBUG ] create a done file to avoid re-doing the mon deployment
[ceph-2][DEBUG ] create the init path if it does not exist
[ceph-2][INFO ] Running command: systemctl enable ceph.target
[ceph-2][INFO ] Running command: systemctl enable ceph-mon@ceph-2
[ceph-2][INFO ] Running command: systemctl start ceph-mon@ceph-2
[ceph-2][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-2.asok mon_status
[ceph-2][DEBUG ] ********************************************************************************
[ceph-2][DEBUG ] status for monitor: mon.ceph-2
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "election_epoch": 8,
[ceph-2][DEBUG ] "extra_probe_peers": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addrvec": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.67.157:3300",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v2"
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.67.157:6789",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v1"
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addrvec": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.112.206:3300",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v2"
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.112.206:6789",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v1"
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addrvec": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.163.105:3300",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v2"
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.163.105:6789",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v1"
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ],
[ceph-2][DEBUG ] "feature_map": {
[ceph-2][DEBUG ] "mon": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "features": "0x3ffddff8ffacffff",
[ceph-2][DEBUG ] "num": 1,
[ceph-2][DEBUG ] "release": "luminous"
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] "features": {
[ceph-2][DEBUG ] "quorum_con": "4611087854031667199",
[ceph-2][DEBUG ] "quorum_mon": [
[ceph-2][DEBUG ] "kraken",
[ceph-2][DEBUG ] "luminous",
[ceph-2][DEBUG ] "mimic",
[ceph-2][DEBUG ] "osdmap-prune",
[ceph-2][DEBUG ] "nautilus"
[ceph-2][DEBUG ] ],
[ceph-2][DEBUG ] "required_con": "2449958747315912708",
[ceph-2][DEBUG ] "required_mon": [
[ceph-2][DEBUG ] "kraken",
[ceph-2][DEBUG ] "luminous",
[ceph-2][DEBUG ] "mimic",
[ceph-2][DEBUG ] "osdmap-prune",
[ceph-2][DEBUG ] "nautilus"
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] "monmap": {
[ceph-2][DEBUG ] "created": "2023-11-11 09:54:05.372287",
[ceph-2][DEBUG ] "epoch": 1,
[ceph-2][DEBUG ] "features": {
[ceph-2][DEBUG ] "optional": [],
[ceph-2][DEBUG ] "persistent": [
[ceph-2][DEBUG ] "kraken",
[ceph-2][DEBUG ] "luminous",
[ceph-2][DEBUG ] "mimic",
[ceph-2][DEBUG ] "osdmap-prune",
[ceph-2][DEBUG ] "nautilus"
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] "fsid": "ff72b496-d036-4f1b-b2ad-55358f3c16cb",
[ceph-2][DEBUG ] "min_mon_release": 14,
[ceph-2][DEBUG ] "min_mon_release_name": "nautilus",
[ceph-2][DEBUG ] "modified": "2023-11-11 09:54:05.372287",
[ceph-2][DEBUG ] "mons": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.67.157:6789/0",
[ceph-2][DEBUG ] "name": "ceph-3",
[ceph-2][DEBUG ] "public_addr": "172.17.67.157:6789/0",
[ceph-2][DEBUG ] "public_addrs": {
[ceph-2][DEBUG ] "addrvec": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.67.157:3300",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v2"
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.67.157:6789",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v1"
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] "rank": 0
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.112.206:6789/0",
[ceph-2][DEBUG ] "name": "ceph-1",
[ceph-2][DEBUG ] "public_addr": "172.17.112.206:6789/0",
[ceph-2][DEBUG ] "public_addrs": {
[ceph-2][DEBUG ] "addrvec": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.112.206:3300",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v2"
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.112.206:6789",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v1"
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] "rank": 1
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.163.105:6789/0",
[ceph-2][DEBUG ] "name": "ceph-0",
[ceph-2][DEBUG ] "public_addr": "172.17.163.105:6789/0",
[ceph-2][DEBUG ] "public_addrs": {
[ceph-2][DEBUG ] "addrvec": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.163.105:3300",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v2"
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.163.105:6789",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v1"
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] "rank": 2
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.227.100:6789/0",
[ceph-2][DEBUG ] "name": "ceph-2",
[ceph-2][DEBUG ] "public_addr": "172.17.227.100:6789/0",
[ceph-2][DEBUG ] "public_addrs": {
[ceph-2][DEBUG ] "addrvec": [
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.227.100:3300",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v2"
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] {
[ceph-2][DEBUG ] "addr": "172.17.227.100:6789",
[ceph-2][DEBUG ] "nonce": 0,
[ceph-2][DEBUG ] "type": "v1"
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] "rank": 3
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ]
[ceph-2][DEBUG ] },
[ceph-2][DEBUG ] "name": "ceph-2",
[ceph-2][DEBUG ] "outside_quorum": [],
[ceph-2][DEBUG ] "quorum": [
[ceph-2][DEBUG ] 0,
[ceph-2][DEBUG ] 1,
[ceph-2][DEBUG ] 2,
[ceph-2][DEBUG ] 3
[ceph-2][DEBUG ] ],
[ceph-2][DEBUG ] "quorum_age": 926,
[ceph-2][DEBUG ] "rank": 3,
[ceph-2][DEBUG ] "state": "peon",
[ceph-2][DEBUG ] "sync_provider": []
[ceph-2][DEBUG ] }
[ceph-2][DEBUG ] ********************************************************************************
[ceph-2][INFO ] monitor: mon.ceph-2 is running
[ceph-2][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-2.asok mon_status
[ceph_deploy.mon][DEBUG ] detecting platform for host ceph-3 ...
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-3][DEBUG ] connected to host: ceph-3
[ceph-3][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-3][DEBUG ] detect machine type
[ceph-3][DEBUG ] find the location of an executable
[ceph_deploy.mon][INFO ] distro info: bclinux 21.10U3 21.10U3 LTS
[ceph-3][DEBUG ] determining if provided host has same hostname in remote
[ceph-3][DEBUG ] get remote short hostname
[ceph-3][DEBUG ] deploying mon to ceph-3
[ceph-3][DEBUG ] get remote short hostname
[ceph-3][DEBUG ] remote hostname: ceph-3
[ceph-3][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-3][DEBUG ] create the mon path if it does not exist
[ceph-3][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-ceph-3/done
[ceph-3][DEBUG ] create a done file to avoid re-doing the mon deployment
[ceph-3][DEBUG ] create the init path if it does not exist
[ceph-3][INFO ] Running command: systemctl enable ceph.target
[ceph-3][INFO ] Running command: systemctl enable ceph-mon@ceph-3
[ceph-3][INFO ] Running command: systemctl start ceph-mon@ceph-3
[ceph-3][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-3.asok mon_status
[ceph-3][DEBUG ] ********************************************************************************
[ceph-3][DEBUG ] status for monitor: mon.ceph-3
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "election_epoch": 8,
[ceph-3][DEBUG ] "extra_probe_peers": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addrvec": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.112.206:3300",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v2"
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.112.206:6789",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v1"
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addrvec": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.163.105:3300",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v2"
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.163.105:6789",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v1"
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addrvec": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.227.100:3300",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v2"
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.227.100:6789",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v1"
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ],
[ceph-3][DEBUG ] "feature_map": {
[ceph-3][DEBUG ] "mon": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "features": "0x3ffddff8ffacffff",
[ceph-3][DEBUG ] "num": 1,
[ceph-3][DEBUG ] "release": "luminous"
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] "features": {
[ceph-3][DEBUG ] "quorum_con": "4611087854031667199",
[ceph-3][DEBUG ] "quorum_mon": [
[ceph-3][DEBUG ] "kraken",
[ceph-3][DEBUG ] "luminous",
[ceph-3][DEBUG ] "mimic",
[ceph-3][DEBUG ] "osdmap-prune",
[ceph-3][DEBUG ] "nautilus"
[ceph-3][DEBUG ] ],
[ceph-3][DEBUG ] "required_con": "2449958747315912708",
[ceph-3][DEBUG ] "required_mon": [
[ceph-3][DEBUG ] "kraken",
[ceph-3][DEBUG ] "luminous",
[ceph-3][DEBUG ] "mimic",
[ceph-3][DEBUG ] "osdmap-prune",
[ceph-3][DEBUG ] "nautilus"
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] "monmap": {
[ceph-3][DEBUG ] "created": "2023-11-11 09:54:05.372287",
[ceph-3][DEBUG ] "epoch": 1,
[ceph-3][DEBUG ] "features": {
[ceph-3][DEBUG ] "optional": [],
[ceph-3][DEBUG ] "persistent": [
[ceph-3][DEBUG ] "kraken",
[ceph-3][DEBUG ] "luminous",
[ceph-3][DEBUG ] "mimic",
[ceph-3][DEBUG ] "osdmap-prune",
[ceph-3][DEBUG ] "nautilus"
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] "fsid": "ff72b496-d036-4f1b-b2ad-55358f3c16cb",
[ceph-3][DEBUG ] "min_mon_release": 14,
[ceph-3][DEBUG ] "min_mon_release_name": "nautilus",
[ceph-3][DEBUG ] "modified": "2023-11-11 09:54:05.372287",
[ceph-3][DEBUG ] "mons": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.67.157:6789/0",
[ceph-3][DEBUG ] "name": "ceph-3",
[ceph-3][DEBUG ] "public_addr": "172.17.67.157:6789/0",
[ceph-3][DEBUG ] "public_addrs": {
[ceph-3][DEBUG ] "addrvec": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.67.157:3300",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v2"
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.67.157:6789",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v1"
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] "rank": 0
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.112.206:6789/0",
[ceph-3][DEBUG ] "name": "ceph-1",
[ceph-3][DEBUG ] "public_addr": "172.17.112.206:6789/0",
[ceph-3][DEBUG ] "public_addrs": {
[ceph-3][DEBUG ] "addrvec": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.112.206:3300",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v2"
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.112.206:6789",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v1"
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] "rank": 1
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.163.105:6789/0",
[ceph-3][DEBUG ] "name": "ceph-0",
[ceph-3][DEBUG ] "public_addr": "172.17.163.105:6789/0",
[ceph-3][DEBUG ] "public_addrs": {
[ceph-3][DEBUG ] "addrvec": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.163.105:3300",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v2"
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.163.105:6789",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v1"
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] "rank": 2
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.227.100:6789/0",
[ceph-3][DEBUG ] "name": "ceph-2",
[ceph-3][DEBUG ] "public_addr": "172.17.227.100:6789/0",
[ceph-3][DEBUG ] "public_addrs": {
[ceph-3][DEBUG ] "addrvec": [
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.227.100:3300",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v2"
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] {
[ceph-3][DEBUG ] "addr": "172.17.227.100:6789",
[ceph-3][DEBUG ] "nonce": 0,
[ceph-3][DEBUG ] "type": "v1"
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] "rank": 3
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ]
[ceph-3][DEBUG ] },
[ceph-3][DEBUG ] "name": "ceph-3",
[ceph-3][DEBUG ] "outside_quorum": [],
[ceph-3][DEBUG ] "quorum": [
[ceph-3][DEBUG ] 0,
[ceph-3][DEBUG ] 1,
[ceph-3][DEBUG ] 2,
[ceph-3][DEBUG ] 3
[ceph-3][DEBUG ] ],
[ceph-3][DEBUG ] "quorum_age": 931,
[ceph-3][DEBUG ] "rank": 0,
[ceph-3][DEBUG ] "state": "leader",
[ceph-3][DEBUG ] "sync_provider": []
[ceph-3][DEBUG ] }
[ceph-3][DEBUG ] ********************************************************************************
[ceph-3][INFO ] monitor: mon.ceph-3 is running
[ceph-3][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-3.asok mon_status
[errno 2] error connecting to the cluster

没有key?继续下个步骤再观察
收集秘钥
ceph-deploy gatherkeys ceph-0 ceph-1 ceph-2 ceph-3
日志
ceph -s 可以看到daemons服务了
部署admin节点
ceph-deploy admin ceph-0 ceph-1 ceph-2 ceph-3
[root@ceph-0 ceph]# ceph-deploy admin ceph-0 ceph-1 ceph-2 ceph-3
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy admin ceph-0 ceph-1 ceph-2 ceph-3
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0xffff91add0f0>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] client : ['ceph-0', 'ceph-1', 'ceph-2', 'ceph-3']
[ceph_deploy.cli][INFO ] func : <function admin at 0xffff91c777d0>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to ceph-0
[ceph-0][DEBUG ] connected to host: ceph-0
[ceph-0][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-0][DEBUG ] detect machine type
[ceph-0][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to ceph-1
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-1][DEBUG ] connected to host: ceph-1
[ceph-1][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-1][DEBUG ] detect machine type
[ceph-1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to ceph-2
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-2][DEBUG ] connected to host: ceph-2
[ceph-2][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-2][DEBUG ] detect machine type
[ceph-2][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to ceph-3
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-3][DEBUG ] connected to host: ceph-3
[ceph-3][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-3][DEBUG ] detect machine type
[ceph-3][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
ceph -s
部署OSD
ceph-deploy osd create ceph-0 --data /dev/vdb
ceph-deploy osd create ceph-1 --data /dev/vdb
ceph-deploy osd create ceph-2 --data /dev/vdb
ceph-deploy osd create ceph-3 --data /dev/vdb
日志ceph 0 1 2 3
[root@ceph-0 ceph]# ceph-deploy osd create ceph-0 --data /dev/vdb
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create ceph-0 --data /dev/vdb
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] bluestore : None
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0xffff9cc8cd20>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] fs_type : xfs
[ceph_deploy.cli][INFO ] block_wal : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] journal : None
[ceph_deploy.cli][INFO ] subcommand : create
[ceph_deploy.cli][INFO ] host : ceph-0
[ceph_deploy.cli][INFO ] filestore : None
[ceph_deploy.cli][INFO ] func : <function osd at 0xffff9cd1bed0>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] zap_disk : False
[ceph_deploy.cli][INFO ] data : /dev/vdb
[ceph_deploy.cli][INFO ] block_db : None
[ceph_deploy.cli][INFO ] dmcrypt : False
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] debug : False
[ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/vdb
[ceph-0][DEBUG ] connected to host: ceph-0
[ceph-0][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-0][DEBUG ] detect machine type
[ceph-0][DEBUG ] find the location of an executable
[ceph_deploy.osd][INFO ] Distro info: bclinux 21.10U3 21.10U3 LTS
[ceph_deploy.osd][DEBUG ] Deploying osd to ceph-0
[ceph-0][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-0][WARNIN] osd keyring does not exist yet, creating one
[ceph-0][DEBUG ] create a keyring file
[ceph-0][DEBUG ] find the location of an executable
[ceph-0][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/vdb
[ceph-0][WARNIN] Running command: /usr/bin/ceph-authtool --gen-print-key
[ceph-0][WARNIN] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new c1870346-8e19-4788-b1dd-19bd75d6ec2f
[ceph-0][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-837353d8-91ff-4418-bc8f-a655d94049d4 /dev/vdb
[ceph-0][WARNIN] stdout: Physical volume "/dev/vdb" successfully created.
[ceph-0][WARNIN] stdout: Volume group "ceph-837353d8-91ff-4418-bc8f-a655d94049d4" successfully created
[ceph-0][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 25599 -n osd-block-c1870346-8e19-4788-b1dd-19bd75d6ec2f ceph-837353d8-91ff-4418-bc8f-a655d94049d4
[ceph-0][WARNIN] stdout: Logical volume "osd-block-c1870346-8e19-4788-b1dd-19bd75d6ec2f" created.
[ceph-0][WARNIN] Running command: /usr/bin/ceph-authtool --gen-print-key
[ceph-0][WARNIN] Running command: /usr/bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0
[ceph-0][WARNIN] Running command: /usr/bin/chown -h ceph:ceph /dev/ceph-837353d8-91ff-4418-bc8f-a655d94049d4/osd-block-c1870346-8e19-4788-b1dd-19bd75d6ec2f
[ceph-0][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /dev/dm-3
[ceph-0][WARNIN] Running command: /usr/bin/ln -s /dev/ceph-837353d8-91ff-4418-bc8f-a655d94049d4/osd-block-c1870346-8e19-4788-b1dd-19bd75d6ec2f /var/lib/ceph/osd/ceph-0/block
[ceph-0][WARNIN] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-0/activate.monmap
[ceph-0][WARNIN] stderr: 2023-11-11 10:48:34.800 ffff843261e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory
[ceph-0][WARNIN] 2023-11-11 10:48:34.800 ffff843261e0 -1 AuthRegistry(0xffff7c081d58) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx
[ceph-0][WARNIN] stderr: got monmap epoch 1
[ceph-0][WARNIN] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-0/keyring --create-keyring --name osd.0 --add-key AQB/605l8419HxAAhIoXMxEJCV5J6qOB8AyHrw==
[ceph-0][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-0/keyring
[ceph-0][WARNIN] added entity osd.0 auth(key=AQB/605l8419HxAAhIoXMxEJCV5J6qOB8AyHrw==)
[ceph-0][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring
[ceph-0][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/
[ceph-0][WARNIN] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid c1870346-8e19-4788-b1dd-19bd75d6ec2f --setuser ceph --setgroup ceph
[ceph-0][WARNIN] --> ceph-volume lvm prepare successful for: /dev/vdb
[ceph-0][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0
[ceph-0][WARNIN] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-837353d8-91ff-4418-bc8f-a655d94049d4/osd-block-c1870346-8e19-4788-b1dd-19bd75d6ec2f --path /var/lib/ceph/osd/ceph-0 --no-mon-config
[ceph-0][WARNIN] Running command: /usr/bin/ln -snf /dev/ceph-837353d8-91ff-4418-bc8f-a655d94049d4/osd-block-c1870346-8e19-4788-b1dd-19bd75d6ec2f /var/lib/ceph/osd/ceph-0/block
[ceph-0][WARNIN] Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-0/block
[ceph-0][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /dev/dm-3
[ceph-0][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0
[ceph-0][WARNIN] Running command: /usr/bin/systemctl enable ceph-volume@lvm-0-c1870346-8e19-4788-b1dd-19bd75d6ec2f
[ceph-0][WARNIN] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-0-c1870346-8e19-4788-b1dd-19bd75d6ec2f.service → /usr/lib/systemd/system/ceph-volume@.service.
[ceph-0][WARNIN] Running command: /usr/bin/systemctl enable --runtime ceph-osd@0
[ceph-0][WARNIN] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@0.service → /usr/lib/systemd/system/ceph-osd@.service.
[ceph-0][WARNIN] Running command: /usr/bin/systemctl start ceph-osd@0
[ceph-0][WARNIN] --> ceph-volume lvm activate successful for osd ID: 0
[ceph-0][WARNIN] --> ceph-volume lvm create successful for: /dev/vdb
[ceph-0][INFO ] checking OSD status...
[ceph-0][DEBUG ] find the location of an executable
[ceph-0][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json
[ceph_deploy.osd][DEBUG ] Host ceph-0 is now ready for osd use.
[root@ceph-0 ceph]# ceph-deploy osd create ceph-1 --data /dev/vdb
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create ceph-1 --data /dev/vdb
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] bluestore : None
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0xffff87d9ed20>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] fs_type : xfs
[ceph_deploy.cli][INFO ] block_wal : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] journal : None
[ceph_deploy.cli][INFO ] subcommand : create
[ceph_deploy.cli][INFO ] host : ceph-1
[ceph_deploy.cli][INFO ] filestore : None
[ceph_deploy.cli][INFO ] func : <function osd at 0xffff87e2ded0>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] zap_disk : False
[ceph_deploy.cli][INFO ] data : /dev/vdb
[ceph_deploy.cli][INFO ] block_db : None
[ceph_deploy.cli][INFO ] dmcrypt : False
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] debug : False
[ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/vdb
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-1][DEBUG ] connected to host: ceph-1
[ceph-1][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-1][DEBUG ] detect machine type
[ceph-1][DEBUG ] find the location of an executable
[ceph_deploy.osd][INFO ] Distro info: bclinux 21.10U3 21.10U3 LTS
[ceph_deploy.osd][DEBUG ] Deploying osd to ceph-1
[ceph-1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-1][WARNIN] osd keyring does not exist yet, creating one
[ceph-1][DEBUG ] create a keyring file
[ceph-1][DEBUG ] find the location of an executable
[ceph-1][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/vdb
[ceph-1][WARNIN] Running command: /bin/ceph-authtool --gen-print-key
[ceph-1][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 4aa0152e-d817-4583-817b-81ada419624a
[ceph-1][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-89d26557-d392-4a46-8d3d-6904076cd4e0 /dev/vdb
[ceph-1][WARNIN] stdout: Physical volume "/dev/vdb" successfully created.
[ceph-1][WARNIN] stdout: Volume group "ceph-89d26557-d392-4a46-8d3d-6904076cd4e0" successfully created
[ceph-1][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 25599 -n osd-block-4aa0152e-d817-4583-817b-81ada419624a ceph-89d26557-d392-4a46-8d3d-6904076cd4e0
[ceph-1][WARNIN] stdout: Logical volume "osd-block-4aa0152e-d817-4583-817b-81ada419624a" created.
[ceph-1][WARNIN] Running command: /bin/ceph-authtool --gen-print-key
[ceph-1][WARNIN] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-1
[ceph-1][WARNIN] Running command: /bin/chown -h ceph:ceph /dev/ceph-89d26557-d392-4a46-8d3d-6904076cd4e0/osd-block-4aa0152e-d817-4583-817b-81ada419624a
[ceph-1][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3
[ceph-1][WARNIN] Running command: /bin/ln -s /dev/ceph-89d26557-d392-4a46-8d3d-6904076cd4e0/osd-block-4aa0152e-d817-4583-817b-81ada419624a /var/lib/ceph/osd/ceph-1/block
[ceph-1][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-1/activate.monmap
[ceph-1][WARNIN] stderr: 2023-11-11 10:49:41.805 ffff89d6d1e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory
[ceph-1][WARNIN] 2023-11-11 10:49:41.805 ffff89d6d1e0 -1 AuthRegistry(0xffff84081d58) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx
[ceph-1][WARNIN] stderr: got monmap epoch 1
[ceph-1][WARNIN] Running command: /bin/ceph-authtool /var/lib/ceph/osd/ceph-1/keyring --create-keyring --name osd.1 --add-key AQDC605lWArnLhAAEYhGC+H+Jy224yAIJhL0gA==
[ceph-1][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-1/keyring
[ceph-1][WARNIN] added entity osd.1 auth(key=AQDC605lWArnLhAAEYhGC+H+Jy224yAIJhL0gA==)
[ceph-1][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1/keyring
[ceph-1][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1/
[ceph-1][WARNIN] Running command: /bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 1 --monmap /var/lib/ceph/osd/ceph-1/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-1/ --osd-uuid 4aa0152e-d817-4583-817b-81ada419624a --setuser ceph --setgroup ceph
[ceph-1][WARNIN] --> ceph-volume lvm prepare successful for: /dev/vdb
[ceph-1][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1
[ceph-1][WARNIN] Running command: /bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-89d26557-d392-4a46-8d3d-6904076cd4e0/osd-block-4aa0152e-d817-4583-817b-81ada419624a --path /var/lib/ceph/osd/ceph-1 --no-mon-config
[ceph-1][WARNIN] Running command: /bin/ln -snf /dev/ceph-89d26557-d392-4a46-8d3d-6904076cd4e0/osd-block-4aa0152e-d817-4583-817b-81ada419624a /var/lib/ceph/osd/ceph-1/block
[ceph-1][WARNIN] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-1/block
[ceph-1][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3
[ceph-1][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1
[ceph-1][WARNIN] Running command: /bin/systemctl enable ceph-volume@lvm-1-4aa0152e-d817-4583-817b-81ada419624a
[ceph-1][WARNIN] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-1-4aa0152e-d817-4583-817b-81ada419624a.service → /usr/lib/systemd/system/ceph-volume@.service.
[ceph-1][WARNIN] Running command: /bin/systemctl enable --runtime ceph-osd@1
[ceph-1][WARNIN] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@1.service → /usr/lib/systemd/system/ceph-osd@.service.
[ceph-1][WARNIN] Running command: /bin/systemctl start ceph-osd@1
[ceph-1][WARNIN] --> ceph-volume lvm activate successful for osd ID: 1
[ceph-1][WARNIN] --> ceph-volume lvm create successful for: /dev/vdb
[ceph-1][INFO ] checking OSD status...
[ceph-1][DEBUG ] find the location of an executable
[ceph-1][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json
[ceph_deploy.osd][DEBUG ] Host ceph-1 is now ready for osd use.
[root@ceph-0 ceph]# ceph-deploy osd create ceph-2 --data /dev/vdb
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create ceph-2 --data /dev/vdb
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] bluestore : None
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0xffff9a808d20>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] fs_type : xfs
[ceph_deploy.cli][INFO ] block_wal : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] journal : None
[ceph_deploy.cli][INFO ] subcommand : create
[ceph_deploy.cli][INFO ] host : ceph-2
[ceph_deploy.cli][INFO ] filestore : None
[ceph_deploy.cli][INFO ] func : <function osd at 0xffff9a897ed0>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] zap_disk : False
[ceph_deploy.cli][INFO ] data : /dev/vdb
[ceph_deploy.cli][INFO ] block_db : None
[ceph_deploy.cli][INFO ] dmcrypt : False
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] debug : False
[ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/vdb
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-2][DEBUG ] connected to host: ceph-2
[ceph-2][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-2][DEBUG ] detect machine type
[ceph-2][DEBUG ] find the location of an executable
[ceph_deploy.osd][INFO ] Distro info: bclinux 21.10U3 21.10U3 LTS
[ceph_deploy.osd][DEBUG ] Deploying osd to ceph-2
[ceph-2][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-2][WARNIN] osd keyring does not exist yet, creating one
[ceph-2][DEBUG ] create a keyring file
[ceph-2][DEBUG ] find the location of an executable
[ceph-2][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/vdb
[ceph-2][WARNIN] Running command: /bin/ceph-authtool --gen-print-key
[ceph-2][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new fe7a2030-94ac-4bbb-af27-7950509b0960
[ceph-2][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-8d4ef242-6dc1-4161-9b6a-15a626b86c6f /dev/vdb
[ceph-2][WARNIN] stdout: Physical volume "/dev/vdb" successfully created.
[ceph-2][WARNIN] stdout: Volume group "ceph-8d4ef242-6dc1-4161-9b6a-15a626b86c6f" successfully created
[ceph-2][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 25599 -n osd-block-fe7a2030-94ac-4bbb-af27-7950509b0960 ceph-8d4ef242-6dc1-4161-9b6a-15a626b86c6f
[ceph-2][WARNIN] stdout: Logical volume "osd-block-fe7a2030-94ac-4bbb-af27-7950509b0960" created.
[ceph-2][WARNIN] Running command: /bin/ceph-authtool --gen-print-key
[ceph-2][WARNIN] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-2
[ceph-2][WARNIN] Running command: /bin/chown -h ceph:ceph /dev/ceph-8d4ef242-6dc1-4161-9b6a-15a626b86c6f/osd-block-fe7a2030-94ac-4bbb-af27-7950509b0960
[ceph-2][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3
[ceph-2][WARNIN] Running command: /bin/ln -s /dev/ceph-8d4ef242-6dc1-4161-9b6a-15a626b86c6f/osd-block-fe7a2030-94ac-4bbb-af27-7950509b0960 /var/lib/ceph/osd/ceph-2/block
[ceph-2][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-2/activate.monmap
[ceph-2][WARNIN] stderr: 2023-11-11 10:50:01.837 ffff947321e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory
[ceph-2][WARNIN] 2023-11-11 10:50:01.837 ffff947321e0 -1 AuthRegistry(0xffff8c081d58) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx
[ceph-2][WARNIN] stderr: got monmap epoch 1
[ceph-2][WARNIN] Running command: /bin/ceph-authtool /var/lib/ceph/osd/ceph-2/keyring --create-keyring --name osd.2 --add-key AQDW605lUIA0MhAAqOoCGrnDsVpfoIIKVtCXHg==
[ceph-2][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-2/keyring
[ceph-2][WARNIN] added entity osd.2 auth(key=AQDW605lUIA0MhAAqOoCGrnDsVpfoIIKVtCXHg==)
[ceph-2][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2/keyring
[ceph-2][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2/
[ceph-2][WARNIN] Running command: /bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 2 --monmap /var/lib/ceph/osd/ceph-2/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-2/ --osd-uuid fe7a2030-94ac-4bbb-af27-7950509b0960 --setuser ceph --setgroup ceph
[ceph-2][WARNIN] --> ceph-volume lvm prepare successful for: /dev/vdb
[ceph-2][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2
[ceph-2][WARNIN] Running command: /bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-8d4ef242-6dc1-4161-9b6a-15a626b86c6f/osd-block-fe7a2030-94ac-4bbb-af27-7950509b0960 --path /var/lib/ceph/osd/ceph-2 --no-mon-config
[ceph-2][WARNIN] Running command: /bin/ln -snf /dev/ceph-8d4ef242-6dc1-4161-9b6a-15a626b86c6f/osd-block-fe7a2030-94ac-4bbb-af27-7950509b0960 /var/lib/ceph/osd/ceph-2/block
[ceph-2][WARNIN] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-2/block
[ceph-2][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3
[ceph-2][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2
[ceph-2][WARNIN] Running command: /bin/systemctl enable ceph-volume@lvm-2-fe7a2030-94ac-4bbb-af27-7950509b0960
[ceph-2][WARNIN] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-2-fe7a2030-94ac-4bbb-af27-7950509b0960.service → /usr/lib/systemd/system/ceph-volume@.service.
[ceph-2][WARNIN] Running command: /bin/systemctl enable --runtime ceph-osd@2
[ceph-2][WARNIN] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@2.service → /usr/lib/systemd/system/ceph-osd@.service.
[ceph-2][WARNIN] Running command: /bin/systemctl start ceph-osd@2
[ceph-2][WARNIN] --> ceph-volume lvm activate successful for osd ID: 2
[ceph-2][WARNIN] --> ceph-volume lvm create successful for: /dev/vdb
[ceph-2][INFO ] checking OSD status...
[ceph-2][DEBUG ] find the location of an executable
[ceph-2][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json
[ceph_deploy.osd][DEBUG ] Host ceph-2 is now ready for osd use.
[root@ceph-0 ceph]# ceph-deploy osd create ceph-3 --data /dev/vdb
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create ceph-3 --data /dev/vdb
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] bluestore : None
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0xffff7f600d20>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] fs_type : xfs
[ceph_deploy.cli][INFO ] block_wal : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] journal : None
[ceph_deploy.cli][INFO ] subcommand : create
[ceph_deploy.cli][INFO ] host : ceph-3
[ceph_deploy.cli][INFO ] filestore : None
[ceph_deploy.cli][INFO ] func : <function osd at 0xffff7f68fed0>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] zap_disk : False
[ceph_deploy.cli][INFO ] data : /dev/vdb
[ceph_deploy.cli][INFO ] block_db : None
[ceph_deploy.cli][INFO ] dmcrypt : False
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] debug : False
[ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/vdb
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-3][DEBUG ] connected to host: ceph-3
[ceph-3][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-3][DEBUG ] detect machine type
[ceph-3][DEBUG ] find the location of an executable
[ceph_deploy.osd][INFO ] Distro info: bclinux 21.10U3 21.10U3 LTS
[ceph_deploy.osd][DEBUG ] Deploying osd to ceph-3
[ceph-3][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-3][WARNIN] osd keyring does not exist yet, creating one
[ceph-3][DEBUG ] create a keyring file
[ceph-3][DEBUG ] find the location of an executable
[ceph-3][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/vdb
[ceph-3][WARNIN] Running command: /bin/ceph-authtool --gen-print-key
[ceph-3][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 223dea89-7b5f-4584-b294-bbc0457cd250
[ceph-3][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-a75dd665-280f-4901-90db-d72aea971fd7 /dev/vdb
[ceph-3][WARNIN] stdout: Physical volume "/dev/vdb" successfully created.
[ceph-3][WARNIN] stdout: Volume group "ceph-a75dd665-280f-4901-90db-d72aea971fd7" successfully created
[ceph-3][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 25599 -n osd-block-223dea89-7b5f-4584-b294-bbc0457cd250 ceph-a75dd665-280f-4901-90db-d72aea971fd7
[ceph-3][WARNIN] stdout: Logical volume "osd-block-223dea89-7b5f-4584-b294-bbc0457cd250" created.
[ceph-3][WARNIN] Running command: /bin/ceph-authtool --gen-print-key
[ceph-3][WARNIN] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-3
[ceph-3][WARNIN] Running command: /bin/chown -h ceph:ceph /dev/ceph-a75dd665-280f-4901-90db-d72aea971fd7/osd-block-223dea89-7b5f-4584-b294-bbc0457cd250
[ceph-3][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3
[ceph-3][WARNIN] Running command: /bin/ln -s /dev/ceph-a75dd665-280f-4901-90db-d72aea971fd7/osd-block-223dea89-7b5f-4584-b294-bbc0457cd250 /var/lib/ceph/osd/ceph-3/block
[ceph-3][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-3/activate.monmap
[ceph-3][WARNIN] stderr: 2023-11-11 10:50:22.197 ffffa80151e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory
[ceph-3][WARNIN] 2023-11-11 10:50:22.197 ffffa80151e0 -1 AuthRegistry(0xffffa0081d58) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx
[ceph-3][WARNIN] stderr: got monmap epoch 1
[ceph-3][WARNIN] Running command: /bin/ceph-authtool /var/lib/ceph/osd/ceph-3/keyring --create-keyring --name osd.3 --add-key AQDr605lrGtEDRAAvHq/3Wbxx0jH8NgtcKN/aA==
[ceph-3][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-3/keyring
[ceph-3][WARNIN] added entity osd.3 auth(key=AQDr605lrGtEDRAAvHq/3Wbxx0jH8NgtcKN/aA==)
[ceph-3][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3/keyring
[ceph-3][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3/
[ceph-3][WARNIN] Running command: /bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 3 --monmap /var/lib/ceph/osd/ceph-3/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-3/ --osd-uuid 223dea89-7b5f-4584-b294-bbc0457cd250 --setuser ceph --setgroup ceph
[ceph-3][WARNIN] --> ceph-volume lvm prepare successful for: /dev/vdb
[ceph-3][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3
[ceph-3][WARNIN] Running command: /bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-a75dd665-280f-4901-90db-d72aea971fd7/osd-block-223dea89-7b5f-4584-b294-bbc0457cd250 --path /var/lib/ceph/osd/ceph-3 --no-mon-config
[ceph-3][WARNIN] Running command: /bin/ln -snf /dev/ceph-a75dd665-280f-4901-90db-d72aea971fd7/osd-block-223dea89-7b5f-4584-b294-bbc0457cd250 /var/lib/ceph/osd/ceph-3/block
[ceph-3][WARNIN] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-3/block
[ceph-3][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3
[ceph-3][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3
[ceph-3][WARNIN] Running command: /bin/systemctl enable ceph-volume@lvm-3-223dea89-7b5f-4584-b294-bbc0457cd250
[ceph-3][WARNIN] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-3-223dea89-7b5f-4584-b294-bbc0457cd250.service → /usr/lib/systemd/system/ceph-volume@.service.
[ceph-3][WARNIN] Running command: /bin/systemctl enable --runtime ceph-osd@3
[ceph-3][WARNIN] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@3.service → /usr/lib/systemd/system/ceph-osd@.service.
[ceph-3][WARNIN] Running command: /bin/systemctl start ceph-osd@3
[ceph-3][WARNIN] --> ceph-volume lvm activate successful for osd ID: 3
[ceph-3][WARNIN] --> ceph-volume lvm create successful for: /dev/vdb
[ceph-3][INFO ] checking OSD status...
[ceph-3][DEBUG ] find the location of an executable
[ceph-3][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json
[ceph_deploy.osd][DEBUG ] Host ceph-3 is now ready for osd use.
ceph -s 状态没变,还没有看到存储状态
部署mgr
ceph-deploy mgr create ceph-0 ceph-1 ceph-2 ceph-3
日志
[root@ceph-0 ceph]# ceph-deploy mgr create ceph-0 ceph-1 ceph-2 ceph-3
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy mgr create ceph-0 ceph-1 ceph-2 ceph-3
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] mgr : [('ceph-0', 'ceph-0'), ('ceph-1', 'ceph-1'), ('ceph-2', 'ceph-2'), ('ceph-3', 'ceph-3')]
[ceph_deploy.cli][INFO ] overwrite_conf : False
[ceph_deploy.cli][INFO ] subcommand : create
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0xffff94d07730>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] func : <function mgr at 0xffff94e71dd0>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.mgr][DEBUG ] Deploying mgr, cluster ceph hosts ceph-0:ceph-0 ceph-1:ceph-1 ceph-2:ceph-2 ceph-3:ceph-3
[ceph-0][DEBUG ] connected to host: ceph-0
[ceph-0][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-0][DEBUG ] detect machine type
[ceph_deploy.mgr][INFO ] Distro info: bclinux 21.10U3 21.10U3 LTS
[ceph_deploy.mgr][DEBUG ] remote host will use systemd
[ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to ceph-0
[ceph-0][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-0][WARNIN] mgr keyring does not exist yet, creating one
[ceph-0][DEBUG ] create a keyring file
[ceph-0][DEBUG ] create path recursively if it doesn't exist
[ceph-0][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.ceph-0 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-ceph-0/keyring
[ceph-0][INFO ] Running command: systemctl enable ceph-mgr@ceph-0
[ceph-0][WARNIN] Created symlink /etc/systemd/system/ceph-mgr.target.wants/ceph-mgr@ceph-0.service → /usr/lib/systemd/system/ceph-mgr@.service.
[ceph-0][INFO ] Running command: systemctl start ceph-mgr@ceph-0
[ceph-0][INFO ] Running command: systemctl enable ceph.target
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1613) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-1][DEBUG ] connected to host: ceph-1
[ceph-1][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-1][DEBUG ] detect machine type
[ceph_deploy.mgr][INFO ] Distro info: bclinux 21.10U3 21.10U3 LTS
[ceph_deploy.mgr][DEBUG ] remote host will use systemd
[ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to ceph-1
[ceph-1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-1][WARNIN] mgr keyring does not exist yet, creating one
[ceph-1][DEBUG ] create a keyring file
[ceph-1][DEBUG ] create path recursively if it doesn't exist
[ceph-1][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.ceph-1 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-ceph-1/keyring
[ceph-1][INFO ] Running command: systemctl enable ceph-mgr@ceph-1
[ceph-1][WARNIN] Created symlink /etc/systemd/system/ceph-mgr.target.wants/ceph-mgr@ceph-1.service → /usr/lib/systemd/system/ceph-mgr@.service.
[ceph-1][INFO ] Running command: systemctl start ceph-mgr@ceph-1
[ceph-1][INFO ] Running command: systemctl enable ceph.target
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1626) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-2][DEBUG ] connected to host: ceph-2
[ceph-2][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-2][DEBUG ] detect machine type
[ceph_deploy.mgr][INFO ] Distro info: bclinux 21.10U3 21.10U3 LTS
[ceph_deploy.mgr][DEBUG ] remote host will use systemd
[ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to ceph-2
[ceph-2][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-2][WARNIN] mgr keyring does not exist yet, creating one
[ceph-2][DEBUG ] create a keyring file
[ceph-2][DEBUG ] create path recursively if it doesn't exist
[ceph-2][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.ceph-2 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-ceph-2/keyring
[ceph-2][INFO ] Running command: systemctl enable ceph-mgr@ceph-2
[ceph-2][WARNIN] Created symlink /etc/systemd/system/ceph-mgr.target.wants/ceph-mgr@ceph-2.service → /usr/lib/systemd/system/ceph-mgr@.service.
[ceph-2][INFO ] Running command: systemctl start ceph-mgr@ceph-2
[ceph-2][INFO ] Running command: systemctl enable ceph.target
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
dhclient(1634) is already running - exiting. This version of ISC DHCP is based on the release available
on ftp.isc.org. Features have been added and other changes
have been made to the base software release in order to make
it work better with this distribution.Please report issues with this software via:
https://gitee.com/src-openeuler/dhcp/issuesexiting.
[ceph-3][DEBUG ] connected to host: ceph-3
[ceph-3][DEBUG ] detect platform information from remote host
21.10U3 LTS
bclinux
[ceph-3][DEBUG ] detect machine type
[ceph_deploy.mgr][INFO ] Distro info: bclinux 21.10U3 21.10U3 LTS
[ceph_deploy.mgr][DEBUG ] remote host will use systemd
[ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to ceph-3
[ceph-3][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-3][WARNIN] mgr keyring does not exist yet, creating one
[ceph-3][DEBUG ] create a keyring file
[ceph-3][DEBUG ] create path recursively if it doesn't exist
[ceph-3][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.ceph-3 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-ceph-3/keyring
[ceph-3][INFO ] Running command: systemctl enable ceph-mgr@ceph-3
[ceph-3][WARNIN] Created symlink /etc/systemd/system/ceph-mgr.target.wants/ceph-mgr@ceph-3.service → /usr/lib/systemd/system/ceph-mgr@.service.
[ceph-3][INFO ] Running command: systemctl start ceph-mgr@ceph-3
[ceph-3][INFO ] Running command: systemctl enable ceph.target
ceph -s 查看mgr
!osd经过了约15分钟,才显示数据空间情况
验证ceph块存储
创建存储池(后面两个参数还不清楚意思)
ceph osd pool create vdbench 250 250
指定类型为块存储
ceph osd pool application enable vdbench rbd
创建一个20G的镜像(未设置压缩)
rbd create image1 --size 20G --pool vdbench --image-format 2 --image-feature layering
映射到linux设备文件
rbd map vdbench/image1
参考以下日志,可以看到,已经生成了/dev/rdb0设备文件
参考文档:
ceph-deploy – 使用最少的基础架构部署 Ceph — ceph-deploy 2.1.0 文档
【精选】ceph-deploy部署指定版本ceph集群_mons are allowing insecure global_id reclaim_ggrong0213的博客-CSDN博客
Ceph使用---dashboard启用及Prometheus监控 - cyh00001 - 博客园 (cnblogs.com)
相关文章:

ceph-deploy bclinux aarch64 ceph 14.2.10
ssh-copy-id,部署机免密登录其他三台主机 所有机器硬盘配置参考如下,计划采用vdb作为ceph数据盘 下载ceph-deploy pip install ceph-deploy 免密登录设置主机名 hostnamectl --static set-hostname ceph-0 .. 3 配置hosts 172.17.163.105 ceph-0 172.…...
爬虫项目(13):使用lxml抓取相亲信息
文章目录 书籍推荐完整代码效果书籍推荐 如果你对Python网络爬虫感兴趣,强烈推荐你阅读《Python网络爬虫入门到实战》。这本书详细介绍了Python网络爬虫的基础知识和高级技巧,是每位爬虫开发者的必读之作。详细介绍见👉: 《Python网络爬虫入门到实战》 书籍介绍 完整代码…...
mysql-数据库三大范式是什么、mysql有哪些索引类型,分别有什么作用 、 事务的特性和隔离级别
1. 数据库三大范式是什么? 数据库三大范式是设计关系型数据库时的规范化原则,确保数据库结构的合理性和减少数据冗余。 这三大范式分别是: - **第一范式(1NF):** 数据表中的所有列都是不可分割的原子数据项…...

微信小程序案例3-2 计算器
文章目录 一、运行效果二、知识储备(一)data-*自定义属性(二)模块 三、实现步骤(一)准备工作1、创建项目2、设置导航栏 (二)实现页面结构1、编写页面整体结构2、编写结果区域的结构3…...

QT QSplitter
分裂器QSplitter类提供了一个分裂器部件。和QBoxLayout类似,可以完成布局管理器的功能,但是包含在它里面的部件,默认是可以随着分裂器的大小变化而变化的。 比如一个按钮放在布局管理器中,它的垂直方向默认是不会被拉伸的,但是放到分裂器中就可以被拉伸。还有一点不…...

银行支付凭证截图生成器在线,工商邮政农业招商建设,画板+透明标签+图片框
用易语言设计了一个非常牛X的截图生成器,娱乐使用哈,软件我在这里也不会分享,模版网上找的,百度图库搜到的,上面的LOGO用的是一个在线生成器,然后标签用的黑月透明标签,加一个通用对话框读取图片…...
微服务概述
微服务架构是一种软件设计和开发范式,旨在将大型应用程序分解为一组小而独立的服务单元,这些单元可以独立开发、测试、部署和扩展。每个服务都专注于一个明确定义的业务功能,并通过轻量级的通信机制进行交互。以下是微服务架构的一些关键方面…...

LabVIEW中NIPackageManager功能介绍
LabVIEW中PackageManager功能介绍 使用NIPackage Manager可安装、更新、修复和删除NI软件。 安装NI软件 使用PackageManager浏览和安装NI软件。 1. 在浏览产品选项卡上,单击产品类别以显示该类别中的可用产品。 2. 选择要安装的产品,然后单击…...
【C语言】sem_getvalue
sem_getvalue 是 POSIX 线程库中用于获取信号量当前值的一个函数。信号量(Semaphore)是用于编程中的同步工具,用于管理多个线程或进程对共享资源的并发访问。通常用于限制可以同时访问共享资源的线程数量。函数 sem_getvalue 的声明通常出现在…...
Linux的shell的$# | fi | 说明
$# | fi | 说明 在Linux的Shell脚本中,$# 是一个特殊变量,表示传递给脚本的参数个数。 例如,如果你运行一个脚本并传递了三个参数,那么在脚本内部使用 $# 将会得到 3。这对于确定脚本在执行时接收到了多少个参数是非常有用的。以…...
C //例 7.12 用选择法对数组中10个整数按由小到大排序。
C程序设计 (第四版) 谭浩强 例 7.12 例 7.12 用选择法对数组中10个整数按由小到大排序。 IDE工具:VS2010 Note: 使用不同的IDE工具可能有部分差异。 代码块 方法:使用指针、动态分配内存 #include <stdio.h> #include …...
Spring Bean循环依赖问题及解决
什么是循环依赖 类与类之间的依赖关系形成了闭环,就会导致循环依赖问题的产生。举例来说,假设存在两个服务类A和服务类B,如果A通过依赖注入的方式引用了B,且B通过依赖注入的方式引用了A,那么A和B之间就存在循环依赖。…...

Golang源码分析 | 程序引导过程
环境说明 CentOS Linux release 7.2 (Final) go version go1.16.3 linux/amd64 GNU gdb (GDB) Red Hat Enterprise Linux 7.6.1-80.el7使用gdb查看程序入口 编写一个简单的go程序 // main.go package mainfunc main() {print("Hello world") } 编译go …...

第三章:人工智能深度学习教程-基础神经网络(第四节-从头开始的具有前向和反向传播的深度神经网络 – Python)
本文旨在从头开始实现深度神经网络。我们将实现一个深度神经网络,其中包含一个具有四个单元的隐藏层和一个输出层。实施将从头开始,并实施以下步骤。算法: 1. 可视化输入数据 2. 确定权重和偏置矩阵的形状 3. 初始化矩阵、要使用的函数 4. 前…...

【入门Flink】- 08Flink时间语义和窗口概念
Flink-Windows 是将无限数据切割成有限的“数据块”进行处理,这就是所谓的“窗口”(Window)。 注意:Flink 中窗口并不是静态准备好的,而是动态创建——当有落在这个窗口区间范围的数据达到时,才创建对应的窗…...
【 OpenGauss源码学习 —— 列存储(CStore)(六)】
列存储(CStore)(六) 概述CStore::GetCUDataFromRemote 函数CStore::CheckConsistenceOfCUDescCtl 函数CStore::CheckConsistenceOfCUDesc 函数CStore::CheckConsistenceOfCUData 函数额外补充 声明:本文的部分内容参考…...

MUYUCMS v2.1:一款开源、轻量级的内容管理系统基于Thinkphp开发
MuYuCMS:一款基于Thinkphp开发的轻量级开源内容管理系统,为企业、个人站长提供快速建站解决方案。它具有以下的环境要求: 支持系统:Windows/Linux/Mac WEB服务器:Apache/Nginx/ISS PHP版本:php > 5.6 (…...

SDL2 显示文字
1.简介 SDL本身没有显示文字功能,它需要用扩展库SDL_ttf来显示文字。ttf是True Type Font的缩写,ttf是Windows下的缺省字体,它有美观,放大缩小不变形的优点,因此广泛应用很多场合。 使用ttf库的第一件事要从Windows的…...
c++ future 使用详解
c future 使用详解 std::future 头文件 #include <future>。 类模板,定义如下: template<class T> class future; template<class T> class future<T&>; template<> class future<void>;作用ÿ…...
好用的C C++ 日志宏 OutputDebugStringA 写到文件或界面
日志宏 #include <cstdio> #define OUTPUT_DEBUG_STRING(fmt, ...) do { \char szOutMsgFinal[10240] {0}; \std::snprintf(szOutMsgFinal, sizeof(szOutMsgFinal), "[%s|%d] " fmt "\n", __func__, __LINE__, ##__VA_ARGS__); \OutputDebugString…...
【网络】每天掌握一个Linux命令 - iftop
在Linux系统中,iftop是网络管理的得力助手,能实时监控网络流量、连接情况等,帮助排查网络异常。接下来从多方面详细介绍它。 目录 【网络】每天掌握一个Linux命令 - iftop工具概述安装方式核心功能基础用法进阶操作实战案例面试题场景生产场景…...
Admin.Net中的消息通信SignalR解释
定义集线器接口 IOnlineUserHub public interface IOnlineUserHub {/// 在线用户列表Task OnlineUserList(OnlineUserList context);/// 强制下线Task ForceOffline(object context);/// 发布站内消息Task PublicNotice(SysNotice context);/// 接收消息Task ReceiveMessage(…...

PPT|230页| 制造集团企业供应链端到端的数字化解决方案:从需求到结算的全链路业务闭环构建
制造业采购供应链管理是企业运营的核心环节,供应链协同管理在供应链上下游企业之间建立紧密的合作关系,通过信息共享、资源整合、业务协同等方式,实现供应链的全面管理和优化,提高供应链的效率和透明度,降低供应链的成…...

【CSS position 属性】static、relative、fixed、absolute 、sticky详细介绍,多层嵌套定位示例
文章目录 ★ position 的五种类型及基本用法 ★ 一、position 属性概述 二、position 的五种类型详解(初学者版) 1. static(默认值) 2. relative(相对定位) 3. absolute(绝对定位) 4. fixed(固定定位) 5. sticky(粘性定位) 三、定位元素的层级关系(z-i…...
C++.OpenGL (10/64)基础光照(Basic Lighting)
基础光照(Basic Lighting) 冯氏光照模型(Phong Lighting Model) #mermaid-svg-GLdskXwWINxNGHso {font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;fill:#333;}#mermaid-svg-GLdskXwWINxNGHso .error-icon{fill:#552222;}#mermaid-svg-GLd…...
【服务器压力测试】本地PC电脑作为服务器运行时出现卡顿和资源紧张(Windows/Linux)
要让本地PC电脑作为服务器运行时出现卡顿和资源紧张的情况,可以通过以下几种方式模拟或触发: 1. 增加CPU负载 运行大量计算密集型任务,例如: 使用多线程循环执行复杂计算(如数学运算、加密解密等)。运行图…...
拉力测试cuda pytorch 把 4070显卡拉满
import torch import timedef stress_test_gpu(matrix_size16384, duration300):"""对GPU进行压力测试,通过持续的矩阵乘法来最大化GPU利用率参数:matrix_size: 矩阵维度大小,增大可提高计算复杂度duration: 测试持续时间(秒&…...
【JavaSE】绘图与事件入门学习笔记
-Java绘图坐标体系 坐标体系-介绍 坐标原点位于左上角,以像素为单位。 在Java坐标系中,第一个是x坐标,表示当前位置为水平方向,距离坐标原点x个像素;第二个是y坐标,表示当前位置为垂直方向,距离坐标原点y个像素。 坐标体系-像素 …...
Caliper 配置文件解析:config.yaml
Caliper 是一个区块链性能基准测试工具,用于评估不同区块链平台的性能。下面我将详细解释你提供的 fisco-bcos.json 文件结构,并说明它与 config.yaml 文件的关系。 fisco-bcos.json 文件解析 这个文件是针对 FISCO-BCOS 区块链网络的 Caliper 配置文件,主要包含以下几个部…...
根据万维钢·精英日课6的内容,使用AI(2025)可以参考以下方法:
根据万维钢精英日课6的内容,使用AI(2025)可以参考以下方法: 四个洞见 模型已经比人聪明:以ChatGPT o3为代表的AI非常强大,能运用高级理论解释道理、引用最新学术论文,生成对顶尖科学家都有用的…...