Hi,
So you need to think about failure domains.
Failure domains will be set to host.
If you put all the DB's on one SSD and all the
WAL's on another SSD then a failure of either of those SSD's will result in a
failure of all the OSD's behind them. So in this case all 10 OSD's would have
failed.
Splitting it to 5 OSD's you have RocksDb and WAL on each SSD this then results in a
failure of an SSD only impacting 5 OSD's.
A failure of an SSD will take down all the OSD's that are behind that SSD.
That is what I wondered, thanks to confirm it.
That's one of the reasons I would always say you
need 1 nodes worth of spare capacity in the cluster to allow for automated re-builds to
happen.
As for your EC 7+5 I would have gone for some thing like 8+3 as then you have a spare
node active in the cluster and can still provide full protection in the event of a failure
of a node.
Make sense! On another cluster, I have an EC 7+5 pool for cephfs but there are 4 servers
per chassis. In case I lost one chassis, I still need
to access data. But for that cluster, you are right, 8+3 may be enough for redundancy.
Think about software updates that require a reboot of
a node. Any data written during that time will need recovering to bring it back to full
protection where as if you have a spare node then that data could be written and not
require a later recovery.
It is mostly a read-only cluster to distribute public datasets over S3 inside our network,
it is fine for me if write operations are not fully
protected during a couple of days. All writes operations are managed by us to update
datasets.
But as mentioned above, 8+3 may be a good compromise.
Best,
Yoann
On 03/09/2019, 10:29, "Yoann Moulin"
<yoann.moulin(a)epfl.ch> wrote:
Hello,
I am deploying a new Nautilus cluster and I would like to know what would be the best
OSD's scenario config in this case :
10x 6TB Disk OSDs (data)
2x 480G SSD previously used for journal and can be used for WAL and/or DB
Is it better to put all WAL on one SSD and all DBs on the other one? Or put WAL and
DB of the first 5 OSDs on the first SSD and the 5 others on
the second one.
A more general question, what is the impact on an OSD if we lose the WAL? The DB?
Both?
I plan to use EC 7+5 on 12 servers and I am OK if I lose one server temporarily. I
have spare servers and I can easily add another one in this
cluster.
To deploy this cluster, I use ceph-ansible (stable-4.0). I am not sure how to
configure the playbook to use SSD and disks with LVM.
https://github.com/ceph/ceph-ansible/blob/master/docs/source/osds/scenarios…
Is this good?
osd_objectstore: bluestore
lvm_volumes:
- data: data-lv1
data_vg: data-vg1
db: db-lv1
db_vg: db-vg1
wal: wal-lv1
wal_vg: wal-vg1
- data: data-lv2
data_vg: data-vg2
db: db-lv2
db_vg: db-vg2
wal: wal-lv2
wal_vg: wal-vg2
Is it possible to let the playbook configure LVM for each disk in a mixed case? It
looks like I must configure LVM before running the playbook
but I am not sure if I missed something.
Is wal_vg and db_vg can be identical (on VG per SSD shared with multiple OSDs)?
Thanks for your help.
Best regards,
--
Yoann Moulin
EPFL IC-IT
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
--
Yoann Moulin
EPFL IC-IT