Hi MJ,
this should work. Note that when using cloned devices all traffic will go through the same
VLAN. In that case, I believe you an simply remove the cluster network definition and use
just one IP, there is no point having the second IP on the same VLAN. You will probably
have to do "noout,nodown" for the flip-over, which probably required a restart
of each OSD. I think, however, that a disappearing back network has no real consequences
as the heartbeats always go over both. There might be stuck replication traffic for a
while, but even this can be avoided with "osd pause".
Our configuration with 2 VLANS is this:
public network: ceph0.81: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 9000
cluster network: ceph0.82: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 9000
ceph0: flags=5187<UP,BROADCAST,RUNNING,MASTER,MULTICAST> mtu 9000
em1: flags=6211<UP,BROADCAST,RUNNING,SLAVE,MULTICAST> mtu 9000
em2: flags=6211<UP,BROADCAST,RUNNING,SLAVE,MULTICAST> mtu 9000
p1p1: flags=6211<UP,BROADCAST,RUNNING,SLAVE,MULTICAST> mtu 9000
p1p2: flags=6211<UP,BROADCAST,RUNNING,SLAVE,MULTICAST> mtu 9000
p2p1: flags=6211<UP,BROADCAST,RUNNING,SLAVE,MULTICAST> mtu 9000
p2p2: flags=6211<UP,BROADCAST,RUNNING,SLAVE,MULTICAST> mtu 9000
If you already have 2 VLANs with different IDs, then this flip-over is trivial. I did it
without service outage.
Best regards,
=================
Frank Schilder
AIT Risø Campus
Bygning 109, rum S14
________________________________________
From: mj <lists(a)merit.unu.edu>
Sent: 12 May 2020 13:12:47
To: ceph-users(a)ceph.io
Subject: [ceph-users] Re: Cluster network and public network
Hi,
On 11/05/2020 08:50, Wido den Hollander wrote:
Great to hear! I'm still behind this idea and all
the clusters I design
have a single (or LACP) network going to the host.
One IP address per node where all traffic goes over. That's Ceph, SSH,
(SNMP) Monitoring, etc.
Wido
We have an 'old-style' cluster, seperated LAN/cluster network. We would
like to move over to the 'new-style'.
Is it as easy as: define the NICs in a 2x10G LACP bond0, and add both
NICs to the bond0 config, and add configure like:
auto bond0
iface bond0 inet static
address 192.168.0.5
netmask 255.255.255.0
and add our cluster IP as a second IP, like
auto bond0:1
iface bond0:1 inet static
address 192.168.10.160
netmask 255.255.255.0
On all nodes, reboot, and everything will work?
Or are there ceph specifics to consider?
Thanks,
MJ
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io