One might argue that that mount command should, and that it shouldn’t pass an FQDN to the
kernel
On Jan 24, 2023, at 23:42, Konstantin Shalygin
<k0ste(a)k0ste.ru> wrote:
Hi,
Do you think kernel should care about DNS resolution?
k
On 24 Jan 2023, at 19:07, kushagra.gupta(a)hsc.com
wrote:
Hi team,
We have a ceph cluster with 3 storage nodes:
1. storagenode1 - abcd:abcd:abcd::21
2. storagenode2 - abcd:abcd:abcd::22
3. storagenode3 - abcd:abcd:abcd::23
We have a dns server with ip abcd:abcd:abcd::31 which resolves the above ip's with a
single hostname.
The resolution is as follows:
```
$TTL 1D
@ IN SOA
storage.com root (
6 ; serial
1D ; refresh
1H ; retry
1W ; expire
3H ) ; minimum
IN NS master
master IN A 10.0.1.31
storagenode IN AAAA abcd:abcd:abcd::21
storagenode IN AAAA abcd:abcd:abcd::22
storagenode IN AAAA abcd:abcd:abcd::23
```
We want to mount the ceph storage on a node using this hostname.
For this we are using the command:
```
mount -t ceph [storagenode.storage.com]:6789:/ /backup -o
name=admin,secret=AQCM+8hjqzuZEhAAcuQc+onNKReq7MV+ykFirg==
```
We are getting the following logs in /var/log/messages:
```
Jan 24 17:23:17 localhost kernel: libceph: resolve 'storagenode.storage.com'
(ret=-3): failed
Jan 24 17:23:17 localhost kernel: libceph: parse_ips bad ip
'storagenode.storage.com:6789'
```
We also tried mounting ceph storage by removing the dns server and resolving the ip as
follows:
```
abcd:abcd:abcd::21 storagenode1
```
But we are getting similar results.
Also kindly note that we are able to perform the mount operation if we use ips instead of
domain name.
Could you please help us out with how we can mount ceph using FQDN.
Kindly let me know if any other imformation is required.
My ceph.conf configuration is as follows:
```
[global]
ms bind ipv6 = true
ms bind ipv4 = false
mon initial members = storagenode1,storagenode2,storagenode3
osd pool default crush rule = -1
fsid = 7969b8a3-1df7-4eae-8ccf-2e5794de87fe
mon host =
[v2:[abcd:abcd:abcd::21]:3300,v1:[abcd:abcd:abcd::21]:6789],[v2:[abcd:abcd:abcd::22]:3300,v1:[abcd:abcd:abcd::22]:6789],[v2:[abcd:abcd:abcd::23]:3300,v1:[abcd:abcd:abcd::23]:6789]
public network = abcd:abcd:abcd::/64
cluster network = eff0:eff0:eff0::/64
[osd]
osd memory target = 4294967296
[client.rgw.storagenode1.rgw0]
host = storagenode1
keyring = /var/lib/ceph/radosgw/ceph-rgw.storagenode1.rgw0/keyring
log file = /var/log/ceph/ceph-rgw-storagenode1.rgw0.log
rgw frontends = beast endpoint=[abcd:abcd:abcd::21]:8080
rgw thread pool size = 512
```
Thanks and Regards
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io