Hi,
just wondering if 'ceph-volume lvm zap --destroy /dev/sdc' would
help here. From your previous output you didn't specify the
--destroy flag.
Which cephadm version is installed on the host? Did you also upgrade
the OS when moving to Pacific? (Sorry if I missed that.
Zitat von Patrick Begou <Patrick.Begou(a)univ-grenoble-alpes.fr>fr>:
Le 02/10/2023 à 18:22, Patrick Bégou a écrit :
> Hi all,
>
> still stuck with this problem.
>
> I've deployed octopus and all my HDD have been setup as osd. Fine.
> I've upgraded to pacific and 2 osd have failed. They have been
> automatically removed and upgrade finishes. Cluster Health is
> finaly OK, no data loss.
>
> But now I cannot re-add these osd with pacific (I had previous
> troubles on these old HDDs, lost one osd in octopus and was able
> to reset and re-add it).
>
> I've tried manually to add the first osd on the node where it is
> located, following
>
https://docs.ceph.com/en/pacific/rados/operations/bluestore-migration/
> (not sure it's the best idea...) but it fails too. This node was
> the one used for deploying the cluster.
>
> [ceph: root@mostha1 /]# *ceph-volume lvm zap /dev/sdc*
> --> Zapping: /dev/sdc
> --> --destroy was not specified, but zapping a whole device will
> remove the partition table
> Running command: /usr/bin/dd if=/dev/zero of=/dev/sdc bs=1M
> count=10 conv=fsync
> stderr: 10+0 records in
> 10+0 records out
> 10485760 bytes (10 MB, 10 MiB) copied, 0.663425 s, 15.8 MB/s
> --> Zapping successful for: <Raw Device: /dev/sdc>
>
>
> [ceph: root@mostha1 /]# *ceph-volume lvm create --bluestore --data
> /dev/sdc*
> Running command: /usr/bin/ceph-authtool --gen-print-key
> Running command: /usr/bin/ceph --cluster ceph --name
> client.bootstrap-osd --keyring
> /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
> 9f1eb8ee-41e6-4350-ad73-1be21234ec7c
> stderr: 2023-10-02T16:09:29.855+0000 7fb4eb8c0700 -1 auth: unable
> to find a keyring on /var/lib/ceph/bootstrap-osd/ceph.keyring: (2)
> No such file or directory
> stderr: 2023-10-02T16:09:29.855+0000 7fb4eb8c0700 -1
> AuthRegistry(0x7fb4e405c4d8) no keyring found at
> /var/lib/ceph/bootstrap-osd/ceph.keyring, disabling cephx
> stderr: 2023-10-02T16:09:29.856+0000 7fb4eb8c0700 -1 auth: unable
> to find a keyring on /var/lib/ceph/bootstrap-osd/ceph.keyring: (2)
> No such file or directory
> stderr: 2023-10-02T16:09:29.856+0000 7fb4eb8c0700 -1
> AuthRegistry(0x7fb4e40601d0) no keyring found at
> /var/lib/ceph/bootstrap-osd/ceph.keyring, disabling cephx
> stderr: 2023-10-02T16:09:29.857+0000 7fb4eb8c0700 -1 auth: unable
> to find a keyring on /var/lib/ceph/bootstrap-osd/ceph.keyring: (2)
> No such file or directory
> stderr: 2023-10-02T16:09:29.857+0000 7fb4eb8c0700 -1
> AuthRegistry(0x7fb4eb8bee90) no keyring found at
> /var/lib/ceph/bootstrap-osd/ceph.keyring, disabling cephx
> stderr: 2023-10-02T16:09:29.858+0000 7fb4e965c700 -1
> monclient(hunting): handle_auth_bad_method server allowed_methods
> [2] but i only support [1]
> stderr: 2023-10-02T16:09:29.858+0000 7fb4e9e5d700 -1
> monclient(hunting): handle_auth_bad_method server allowed_methods
> [2] but i only support [1]
> stderr: 2023-10-02T16:09:29.858+0000 7fb4e8e5b700 -1
> monclient(hunting): handle_auth_bad_method server allowed_methods
> [2] but i only support [1]
> stderr: 2023-10-02T16:09:29.858+0000 7fb4eb8c0700 -1 monclient:
> authenticate NOTE: no keyring found; disabled cephx authentication
> stderr: [errno 13] RADOS permission denied (error connecting to
> the cluster)
> --> RuntimeError: Unable to create a new OSD id
>
> Any idea of what is wrong ?
>
> Thanks
>
> Patrick
> _______________________________________________
> ceph-users mailing list -- ceph-users(a)ceph.io
> To unsubscribe send an email to ceph-users-leave(a)ceph.io
I'm still trying to understand what can be wrong or how to debug
this situation where Ceph cannot see the devices.
The device :dev/sdc exists:
[root@mostha1 ~]# cephadm shell lsmcli ldl
Inferring fsid 250f9864-0142-11ee-8e5f-00266cf8869c
Using recent ceph image
quay.io/ceph/ceph@sha256:f30bf50755d7087f47c6223e6a921caf5b12e86401b3d49220230c84a8302a1e
Path | SCSI VPD 0x83 | Link Type | Serial Number | Health
Status
-------------------------------------------------------------------------
/dev/sda | 50024e92039e4f1c | PATA/SATA | S2B5J90ZA10142 | Good
/dev/sdb | 50014ee0ad5953c9 | PATA/SATA | WD-WMAYP0982329 | Good
/dev/sdc | 50024e920387fa2c | PATA/SATA | S2B5J90ZA02494 | Good
But I cannot do anything with it:
[root@mostha1 ~]# cephadm shell ceph orch device zap
mostha1.legi.grenoble-inp.fr /dev/sdc --force
Inferring fsid 250f9864-0142-11ee-8e5f-00266cf8869c
Using recent ceph image
quay.io/ceph/ceph@sha256:f30bf50755d7087f47c6223e6a921caf5b12e86401b3d49220230c84a8302a1e
Error EINVAL: Device path '/dev/sdc' not found on host
'mostha1.legi.grenoble-inp.fr'
Since I moved from octopus to Pacific.
Patrick
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io