Thanks for your reply Anastasios,
I was waiting for some answer.
My /etc/apt/sources.list.d/ceph.list content is:
deb https://download.ceph.com/debian-nautilus/ buster main
Even if I do “apt-get update”, the packages still the same.
The Ceph client (CephFS mount) is working well, but I can´t deploy new osds.
The error that I posted occurs when I do : “ceph-deploy osd create --data /dev/sdb node1”
I appreciate any help.
Rafael.
De: Anastasios Dados <tdados(a)hotmail.com>
Enviada em: segunda-feira, 29 de junho de 2020 20:01
Para: Rafael Quaglio <quaglio(a)bol.com.br>; ceph-users(a)ceph.io
Assunto: Re: [ceph-users] Debian install
Hello Rafael,
Can you check the apt sources list that exist from your ceph-deploy node? Maybe there you have put luminous debian packages version?
Regards,
Anastasios
On Mon, 2020-06-29 at 06:59 -0300, Rafael Quaglio wrote:
Hi,
We have already installed a new Debian (10.4) server and I need put it in a
Ceph cluster.
When I execute the command to install ceph on this node:
ceph-deploy install --release nautilus node1
It starts to install a version 12.x in my node...
(...)
[serifos][DEBUG ] After this operation, 183 MB of additional disk space will
be used.
[serifos][DEBUG ] Selecting previously unselected package python-cephfs.
(Reading database ... 30440 files and directories currently installed.)
[serifos][DEBUG ] Preparing to unpack
.../python-cephfs_12.2.11+dfsg1-2.1+b1_amd64.deb ...
[serifos][DEBUG ] Unpacking python-cephfs (12.2.11+dfsg1-2.1+b1) ...
[serifos][DEBUG ] Selecting previously unselected package ceph-common.
[serifos][DEBUG ] Preparing to unpack
.../ceph-common_12.2.11+dfsg1-2.1+b1_amd64.deb ...
[serifos][DEBUG ] Unpacking ceph-common (12.2.11+dfsg1-2.1+b1) ...
(...)
How do I upgrade this packages?
Even installed packages in this version, the installation
completes without erros.
The question is due to an error message that I'm recieving
when deploy a new osd.
ceph-deploy osd create --data /dev/sdb node1
At this point:
[ceph_deploy.osd][INFO ] Distro info: debian 10.4 buster
[ceph_deploy.osd][DEBUG ] Deploying osd to node1
[node1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[node1][DEBUG ] find the location of an executable
[node1][INFO ] Running command: sudo /usr/sbin/ceph-volume --cluster ceph
lvm create --bluestore --data /dev/sdb
[node1][WARNIN] --> RuntimeError: Unable to create a new OSD id
[node1][DEBUG ] Running command: /bin/ceph-authtool --gen-print-key
[node1][DEBUG ] Running command: /bin/ceph --cluster ceph --name
client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i -
osd new 76da6c51-8385-4ffc-9a8e-0dfc11e31feb
[node1][DEBUG ] stderr:
/build/ceph-qtARip/ceph-12.2.11+dfsg1/src/mon/MonMap.cc: In function 'void
MonMap::sanitize_mons(std::map<std::__cxx11::basic_string<char>,
entity_addr_t>&)' thread 7f2bc7fff700 time 2020-06-29 06:56:17.331350
[node1][DEBUG ] stderr:
/build/ceph-qtARip/ceph-12.2.11+dfsg1/src/mon/MonMap.cc: 77: FAILED
assert(mon_info[p.first].public_addr == p.second)
[node1][DEBUG ] stderr: ceph version 12.2.11
(26dc3775efc7bb286a1d6d66faee0ba30ea23eee) luminous (stable)
[node1][DEBUG ] stderr: 1: (ceph::__ceph_assert_fail(char const*, char
const*, int, char const*)+0xf5) [0x7f2bdaff5f75]
[node1][DEBUG ] stderr: 2:
(MonMap::sanitize_mons(std::map<std::__cxx11::basic_string<char,
std::char_traits<char>, std::allocator<char> >, entity_addr_t,
std::less<std::__cxx11::basic_string<char, std::char_traits<char>,
std::allocator<char> > >,
std::allocator<std::pair<std::__cxx11::basic_string<char,
std::char_traits<char>, std::allocator<char> > const, entity_addr_t> >
&)+0x568) [0x7f2bdb050038]
[node1][DEBUG ] stderr: 3:
(MonMap::decode(ceph::buffer::list::iterator&)+0x4da) [0x7f2bdb05500a]
[node1][DEBUG ] stderr: 4: (MonClient::handle_monmap(MMonMap*)+0x216)
[0x7f2bdb042a06]
[node1][DEBUG ] stderr: 5: (MonClient::ms_dispatch(Message*)+0x4ab)
[0x7f2bdb04729b]
[node1][DEBUG ] stderr: 6: (DispatchQueue::entry()+0xeba) [0x7f2bdb06bf5a]
[node1][DEBUG ] stderr: 7: (DispatchQueue::DispatchThread::entry()+0xd)
[0x7f2bdb1576fd]
[node1][DEBUG ] stderr: 8: (()+0x7fa3) [0x7f2be499dfa3]
[node1][DEBUG ] stderr: 9: (clone()+0x3f) [0x7f2be45234cf]
[node1][DEBUG ] stderr: NOTE: a copy of the executable, or `objdump -rdS
<executable>` is needed to interpret this.
[node1][ERROR ] RuntimeError: command returned non-zero exit status: 1
[ceph_deploy.osd][ERROR ] Failed to execute command: /usr/sbin/ceph-volume
--cluster ceph lvm create --bluestore --data /dev/sdb
[ceph_deploy][ERROR ] GenericError: Failed to create 1 OSDs
I think this error occurs because the wrong package that was
installed.
Thanks,
Rafael
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io <mailto:ceph-users@ceph.io>
To unsubscribe send an email to ceph-users-leave(a)ceph.io <mailto:ceph-users-leave@ceph.io>
Hi guys,
I'm using ceph fs , I met a problem of large object in my fs meta pool.
Anyone can tell me how to avoid the problem? thanks.
best regards,
kern
Hi,
We are experiencing a weird issue after upgrading our clusters from ceph
luminous to nautilus 14.2.9 - I am not even sure if this is ceph related
but this started to happen exactly after we upgraded, so, I am trying my
luck here.
We have one ceph rbd pool size 3 min size 2 from all bluestore osds (KRBD)
I will try to be clear enough.. though I cannot understand exactly whats
happening or whats causing the issue.
So, we have 1 virtual machine which uses a rbd image of 2TB -
virtio-scsi device.
Inside the VM we are trying to create ploop devices to be used for/by
containers(inside the VM on the 2TB rbd image QEMU DISK).
There is no way we can create ploop devices, it always crash, please
check the crash below:
https://pastebin.com/9khp9XS3 - sdb in the crash is the 2TB rbd image
which the VM uses.
There are no other read/write errors, we have health_ok, all OSDs are
fine, no errors on any of the phisical disks - this happens only when we
want to create ploop devices inside a VM and right after we upgraded our
cluster to nautilus 14.2.9.
I also did new images/other hosts.. same result. Did try a lot of
different versions of ploop packages, same result.
I would appreciate if someone else has encountered something similar and
if there is a workaround.
--
Best Regards,
------------------------------------------------------------------------
Daniel Stan
Senior System Administrator | NAV Communications (RO)
Office: +40 (21) 655-55-55 | E-Mail: daniel(a)nav.ro
Site: www.nav.ro <https://www.nav.ro> | Client: https://client.ro
Hello...it's the first time I need to use the lifecycle, and I created a
bucket and set it to expire in one day with s3cmd:
s3cmd expire --expiry-days=1 s3://bucket
The rgw_lifecycle_work_time is set to the default values(00:00-06:00). But
I noticed in the rgw logs a lot of messages like:
2020-06-16 00:00:00.311369 7fe2cac87700 0 RGWLC::process() failed to get
obj entry lc.8
2020-06-16 00:00:00.311623 7fe2c8c83700 0 RGWLC::process() failed to get
obj entry lc.16
2020-06-16 00:00:00.311862 7fe2c6c7f700 0 RGWLC::process() failed to get
obj entry lc.4
2020-06-16 00:00:00.319424 7fe2cac87700 0 RGWLC::process() failed to get
obj entry lc.10
2020-06-16 00:00:00.319647 7fe2c8c83700 0 RGWLC::process() failed to get
obj entry lc.18
2020-06-16 00:00:00.320682 7fe2c6c7f700 0 RGWLC::process() failed to get
obj entry lc.16
2020-06-16 00:00:00.327770 7fe2cac87700 0 RGWLC::process() failed to get
obj entry lc.6
2020-06-16 00:00:00.328941 7fe2c8c83700 0 RGWLC::process() failed to get
obj entry lc.17
2020-06-16 00:00:00.332463 7fe2c6c7f700 0 RGWLC::process() failed to get
obj entry lc.20
2020-06-16 00:00:00.336788 7fe2cac87700 0 RGWLC::process() failed to get
obj entry lc.1
2020-06-16 00:00:00.336924 7fe2c8c83700 0 RGWLC::process() failed to get
obj entry lc.24
2020-06-16 00:00:00.340915 7fe2c6c7f700 0 RGWLC::process() failed to get
obj entry lc.2
The object was deleted, but these messages keep appearing.
Is it safe to ignore them?
For the records, i'm using redhat luminous 12.2.12
Thanks, Marcelo.
Hello,
Is there a way to list all locks held by a client with the given IP address?
Also, I read somewhere that removing the lock with "rbd lock rm..."
automatically blacklists that client connection. Is that correct?
How do I blacklist a client with the given IP address?
Thanks,
Shridhar