Just checked cluster logs and they are full of:cephadm exited with an error code: 1,
stderr:Reconfig daemon osd.16 ... Traceback (most recent call last): File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 7931, in <module> main() File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 7919, in main r = ctx.func(ctx) File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 1717, in defaultimage return func(ctx) File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 4162, in command_deploy c = get_container(ctx, ctx.fsid, daemon_type, daemon_id, File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 2451, in get_container volume_mounts=get_container_mounts(ctx, fsid, daemon_type,
daemon_id), File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 2292, in get_container_mounts if HostFacts(ctx).selinux_enabled: File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 6451, in selinux_enabled return (self.kernel_security['type'] ==
'SELinux') and \ File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 6434, in kernel_security ret = _fetch_apparmor() File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 6415, in _fetch_apparmor item, mode = line.split(' ') ValueError: not enough
values to unpack (expected 2, got 1) Traceback (most recent call last): File
"/usr/share/ceph/mgr/cephadm/serve.py", line 1172, in _remote_connection yield
(conn, connr) File "/usr/share/ceph/mgr/cephadm/serve.py", line 1087, in
_run_cephadm code, '\n'.join(err))) orchestrator._interface.OrchestratorError:
cephadm exited with an error code: 1, stderr:Reconfig daemon osd.16 ... Traceback (most
recent call last): File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 7931, in <module> main() File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 7919, in main r = ctx.func(ctx) File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 1717, in _default_image return func(ctx) File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 4162, in command_deploy c = get_container(ctx, ctx.fsid, daemon_type, daemon_id, File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 2451, in get_container volume_mounts=get_container_mounts(ctx, fsid, daemon_type,
daemon_id), File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 2292, in get_container_mounts if HostFacts(ctx).selinux_enabled: File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 6451, in selinux_enabled return (self.kernel_security['type'] ==
'SELinux') and \ File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 6434, in kernel_security ret = _fetch_apparmor() File
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
line 6415, in _fetch_apparmor item, mode = line.split(' ') ValueError: not enough
values to unpack (expected 2, got 1)being repeated over and over again for each OSD.Again
listing "ValueError: not enough values to unpack (expected 2, got 1)"
On Mon May 03 2021 17:20:59 GMT+0800 (Singapore
Standard Time), Ashley Merrick <ashley(a)amerrick.co.uk> wrote:
Hello,Wondering if anyone had any feedback on some commands I could try to manually
update the current OSD that is down to 16.2.1 so I can at least get around this upgrade
bug and back to 100%?If there is any log's or if it seems a new bug and I should
create a bugzilla report do let me know.Thanks
On Fri Apr 30 2021 21:54:30 GMT+0800 (Singapore
Standard Time), Ashley Merrick <ashley(a)amerrick.co.uk> wrote:
Hello All,I was running 15.2.8 via cephadm on docker Ubuntu 20.04I just attempted to
upgrade to 16.2.1 via the automated method, it successfully upgraded the mon/mgr/mds and
some OSD's, however it then failed on an OSD and hasn't been able to pass even
after stopping and restarting the upgrade.It reported the following
""message": "Error: UPGRADEREDEPLOYDAEMON: Upgrading daemon osd.35 on
host sn-s01 failed.""If I run 'ceph health detail' I get lot's of
the following error : "ValueError: not enough values to unpack (expected 2, got
1)" throughout the detail reportUpon googling, it looks like I am hitting something
along the lines of
https://158.69.68.89/issues/48924 &
https://tracker.ceph.com/issues/49522What do I need to do to either get around this bug,
or a way I can manually upgrade the remaining ceph OSD's to 16.2.1, currently my
cluster is working but the last OSD it failed to upgrade is currently offline (I guess as
no image attached to it now as it failed to pull it), and I have a cluster with OSD's
from not 15.2.8 and 16.2.1Thanks
Sent via MXlogin
Sent via MXlogin
Sent via MXlogin