Same happened to us two weeks ago using nautilus, although we added the rules and storage
classes.
----- Am 14. Sep 2020 um 16:02 schrieb Marc Roos M.Roos(a)f1-outsourcing.eu:
I did the same, 1 or 2 years ago, creating a
replicated_ruleset_hdd and
replicated_ruleset_ssd. Eventhough I did not have any ssd's on any of
the nodes at that time, adding this hdd type criteria made pg's migrate.
I thought it was strange that this happens on a hdd only cluster, so I
mentioned it here. I am not sure however if this is still an issue, but
better take this into account.
-----Original Message-----
To: ceph-users(a)ceph.io
Subject: [ceph-users] New pool with SSD OSDs
Hello!
We have a Ceph cluster with 30 HDD 4 TB in 6 hosts, only for RBD.
Now, we're receiving other 6 servers with 6 SSD 2 TB each and we want to
create a separate pool for RBD on SSD, and let unused and backup volumes
stays in HDD.
I have some questions:
As I am only using "replicated_rule". ¿If I add an SSD OSD to the
cluster, Ceph starts to migrate PGs to it?
If so, to prevent this, first I have to create rule like
# ceph osd crush rule create-replicated pool-hdd default host hdd
and then
#ceph osd pool set rbd crush_rule pool-hdd
?
Or, if Ceph does not mix automatically hdd and ssd, I create the SSD OSD
and then
# ceph osd crush rule create-replicated pool-ssd default host ssd
# ceph osd pool create pool-ssd 256 256 ssdpool
?
And then migrate images from one to another pool as needed.
Any thoughts are wellcome!
Thanks in advanced for your time.
Javier.-
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
--
Dipl.-Inf. André Gemünd, Leiter IT / Head of IT
Fraunhofer-Institute for Algorithms and Scientific Computing
andre.gemuend(a)scai.fraunhofer.de
Tel: +49 2241 14-2193
/C=DE/O=Fraunhofer/OU=SCAI/OU=People/CN=Andre Gemuend