Ceph version Quincy.
But now I am able to resolve the issue.
During mount i will not pass any monitor details, it will be
auto-discovered via SRV.
On Tue, Apr 11, 2023 at 6:09 PM Eugen Block <eblock(a)nde.ag> wrote:
What ceph version is this? Could it be this bug
[1]? Although the
error message is different, not sure if it could be the same issue,
and I don't have anything to test ipv6 with.
[1]
https://tracker.ceph.com/issues/47300
Zitat von Lokendra Rathour <lokendrarathour(a)gmail.com>om>:
Hi All,
Requesting any inputs around the issue raised.
Best Regards,
Lokendra
On Tue, 24 Jan, 2023, 7:32 pm Lokendra Rathour, <
lokendrarathour(a)gmail.com>
wrote:
> Hi Team,
>
>
>
> We have a ceph cluster with 3 storage nodes:
>
> 1. storagenode1 - abcd:abcd:abcd::21
>
> 2. storagenode2 - abcd:abcd:abcd::22
>
> 3. storagenode3 - abcd:abcd:abcd::23
>
>
>
> The requirement is to mount ceph using the domain name of MON node:
>
> Note: we resolved the domain name via DNS server.
>
>
> For this we are using the command:
>
> ```
>
> mount -t ceph [storagenode.storage.com]:6789:/ /backup -o
> name=admin,secret=AQCM+8hjqzuZEhAAcuQc+onNKReq7MV+ykFirg==
>
> ```
>
>
>
> We are getting the following logs in /var/log/messages:
>
> ```
>
> Jan 24 17:23:17 localhost kernel: libceph: resolve '
> storagenode.storage.com' (ret=-3): failed
>
> Jan 24 17:23:17 localhost kernel: libceph: parse_ips bad ip '
> storagenode.storage.com:6789'
>
> ```
>
>
>
> We also tried mounting ceph storage using IP of MON which is working
fine.
>
>
>
> Query:
>
>
> Could you please help us out with how we can mount ceph using FQDN.
>
>
>
> My /etc/ceph/ceph.conf is as follows:
>
> [global]
>
> ms bind ipv6 = true
>
> ms bind ipv4 = false
>
> mon initial members = storagenode1,storagenode2,storagenode3
>
> osd pool default crush rule = -1
>
> fsid = 7969b8a3-1df7-4eae-8ccf-2e5794de87fe
>
> mon host =
>
[v2:[abcd:abcd:abcd::21]:3300,v1:[abcd:abcd:abcd::21]:6789],[v2:[abcd:abcd:abcd::22]:3300,v1:[abcd:abcd:abcd::22]:6789],[v2:[abcd:abcd:abcd::23]:3300,v1:[abcd:abcd:abcd::23]:6789]
public network = abcd:abcd:abcd::/64
cluster network = eff0:eff0:eff0::/64
[osd]
osd memory target = 4294967296
[client.rgw.storagenode1.rgw0]
host = storagenode1
keyring = /var/lib/ceph/radosgw/ceph-rgw.storagenode1.rgw0/keyring
log file = /var/log/ceph/ceph-rgw-storagenode1.rgw0.log
rgw frontends = beast endpoint=[abcd:abcd:abcd::21]:8080
rgw thread pool size = 512
--
~ Lokendra
skype: lokendrarathour
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io