It seems like the default value for mgr/cephadm/container_image_base
has been changed from quay.io/ceph/ceph to
quay.io/ceph/ceph:v18 between Quincy and Reef. So a quick fix would be
to set it to the previous default:
ceph config set mgr mgr/cephadm/container_image_base quay.io/ceph/ceph
Apparently, this is the responsible change:
# Quincy
grep "DEFAULT_IMAGE\ =" /usr/share/ceph/mgr/cephadm/module.py
DEFAULT_IMAGE = 'quay.io/ceph/ceph'
# Reef
grep "DEFAULT_IMAGE\ =" /usr/share/ceph/mgr/cephadm/module.py
DEFAULT_IMAGE = 'quay.io/ceph/ceph:v18'
Although the description clearly states:
Hi,
either the cephadm version installed on the host should be updated
as well so it matches the cluster version or you can also use the
one that the orchestrator uses which stores its different versions
in this path (@Mykola thanks again for pointing that out), the
latest matches the current ceph version:
/var/lib/ceph/${fsid}/cephadm.*
If you set the executable bit you can use it as usual:
# pacific package version
$ rpm -qf /usr/sbin/cephadm
cephadm-16.2.11.65+g8b7e6fc0182-lp154.3872.1.noarch
$ chmod +x
/var/lib/ceph/201a2fbc-ce7b-44a3-9ed7-39427972083b/cephadm.7dcbd4aab60af3e83970c60d4a8a2cc6ea7b997ecc2f4de0a47eeacbb88dde46
$ python3
/var/lib/ceph/201a2fbc-ce7b-44a3-9ed7-39427972083b/cephadm.7dcbd4aab60af3e83970c60d4a8a2cc6ea7b997ecc2f4de0a47eeacbb88dde46
ls
[
{
"style": "cephadm:v1",
...
}
]
Also the command:
ceph orch upgrade start -ceph_version v18.2.0
That looks like a bug to me, it's reproducable:
$ ceph orch upgrade check --ceph-version 18.2.0
Error EINVAL: host ceph01 `cephadm pull` failed: cephadm exited with
an error code: 1, stderr: Pulling container image
quay.io/ceph/ceph:v18:v18.2.0...
Non-zero exit code 125 from /usr/bin/podman pull
quay.io/ceph/ceph:v18:v18.2.0 --authfile=/etc/ceph/podman-auth.json
/usr/bin/podman: stderr Error: invalid reference format
ERROR: Failed command: /usr/bin/podman pull
quay.io/ceph/ceph:v18:v18.2.0 --authfile=/etc/ceph/podman-auth.json
It works correctly with 17.2.6:
# ceph orch upgrade check --ceph-version 18.2.0
{
"needs_update": {
"crash.soc9-ceph": {
"current_id":
"2d45278716053f92517e447bc1a7b64945cc4ecbaff4fe57aa0f21632a0b9930",
"current_name":
"quay.io/ceph/ceph@sha256:1e442b0018e6dc7445c3afa7c307bc61a06189ebd90580a1bb8b3d0866c0d8ae",
"current_version": "17.2.6"
...
I haven't checked for existing tracker issues yet. I'd recommend to
check and create a bug report:
https://tracker.ceph.com/
Regards,
Eugen
Zitat von Martin Conway <martin.conway(a)anu.edu.au>au>:
> Hi
>
> I have been using Ceph for many years now, and recently upgraded to Reef.
>
> Seems I made the jump too quickly, as I have been hitting a few
> issues. I can't find any mention of them in the bug reports. I
> thought I would share them here in case it is something to do with
> my setup.
>
> On V18.2.0
>
> cephadm version
>
> Fails with the following output:
>
> Traceback (most recent call last):
> File "/usr/lib64/python3.6/runpy.py", line 193, in _run_module_as_main
> "__main__", mod_spec)
> File "/usr/lib64/python3.6/runpy.py", line 85, in _run_code
> exec(code, run_globals)
> File "/usr/sbin/cephadm/__main__.py", line 10096, in <module>
> File "/usr/sbin/cephadm/__main__.py", line 10084, in main
> File "/usr/sbin/cephadm/__main__.py", line 2240, in _infer_image
> File "/usr/sbin/cephadm/__main__.py", line 2338, in
infer_local_ceph_image
> File "/usr/sbin/cephadm/__main__.py", line 2301, in get_container_info
> File "/usr/sbin/cephadm/__main__.py", line 2301, in <listcomp>
> File "/usr/sbin/cephadm/__main__.py", line 222, in __getattr__
> AttributeError: 'CephadmContext' object has no attribute 'fsid'
>
> I don't know if it is related, but
>
> cephadm adopt --style legacy --name osd.X
>
> Tries to use a V15 image which then fails to start after being
> imported. The OSD in question has an SSD device from block.db if
> that is relevant.
>
> Using the latest head version of cephadm from github let me work
> around this issue, but the adopted OSDs were running
> 18.0.0-6603-g6c4ed58a and needed to be upgraded to 18.2.0.
>
Also the command:
ceph orch upgrade start -ceph_version v18.2.0
>
> Does not work, it fails to find the right image. From memory I
> think it tried to pull quay.io/ceph/ceph:v18:v18.2.0
>
> Ceph orch upgrade start quay.io/ceph/ceph:v18.2.0
>
> Does work as expected.
>
> Let me know if there is any other information that would be
> helpful, but I have since worked around these issues and have my
> ceph back in a happy state.
>
> Regards,
> Martin Conway
> IT and Digital Media Manager
> Research School of Physics
> Australian National University
> Canberra ACT 2601
>
> +61 2 6125 1599
>
https://physics.anu.edu.au<https://physics.anu.edu.au/>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users(a)ceph.io
> To unsubscribe send an email to ceph-users-leave(a)ceph.io