Did you get an answer for this? My original thought when I read it was that
the osd would need to be recreated(as you noted).
-Brent
-----Original Message-----
From: Marc Roos <M.Roos(a)f1-outsourcing.eu>
Sent: Tuesday, March 3, 2020 10:58 AM
To: abhishek <abhishek(a)suse.com>om>; ceph-users <ceph-users(a)ceph.io>
Subject: [ceph-users] Re: v14.2.8 Nautilus released
This bluestore_min_alloc_size_ssd=4K, do I need to recreate these osd's?
Or does this magically change? What % performance increase can be expected?
-----Original Message-----
To: ceph-announce(a)ceph.io; ceph-users(a)ceph.io; dev(a)ceph.io;
ceph-devel(a)vger.kernel.org
Subject: [ceph-users] v14.2.8 Nautilus released
This is the eighth update to the Ceph Nautilus release series. This
release fixes issues across a range of subsystems. We recommend that all
users upgrade to this release. Please note the following important
changes in this release; as always the full changelog is posted at:
https://ceph.io/releases/v14-2-8-nautilus-released
Notable Changes
---------------
* The default value of `bluestore_min_alloc_size_ssd` has been changed
to 4K to improve performance across all workloads.
* The following OSD memory config options related to bluestore cache
autotuning can now
be configured during runtime:
- osd_memory_base (default: 768 MB)
- osd_memory_cache_min (default: 128 MB)
- osd_memory_expected_fragmentation (default: 0.15)
- osd_memory_target (default: 4 GB)
The above options can be set with::
ceph config set osd <option> <value>
* The MGR now accepts `profile rbd` and `profile rbd-read-only` user
caps.
These caps can be used to provide users access to MGR-based RBD
functionality
such as `rbd perf image iostat` an `rbd perf image iotop`.
* The configuration value `osd_calc_pg_upmaps_max_stddev` used for upmap
balancing has been removed. Instead use the mgr balancer config
`upmap_max_deviation` which now is an integer number of PGs of
deviation
from the target PGs per OSD. This can be set with a command like
`ceph config set mgr mgr/balancer/upmap_max_deviation 2`. The default
`upmap_max_deviation` is 1. There are situations where crush rules
would not allow a pool to ever have completely balanced PGs. For
example, if
crush requires 1 replica on each of 3 racks, but there are fewer OSDs
in 1 of
the racks. In those cases, the configuration value can be increased.
* RGW: a mismatch between the bucket notification documentation and the
actual
message format was fixed. This means that any endpoints receiving
bucket
notification, will now receive the same notifications inside a JSON
array
named 'Records'. Note that this does not affect pulling bucket
notification
from a subscription in a 'pubsub' zone, as these are already wrapped
inside
that array.
* CephFS: multiple active MDS forward scrub is now rejected. Scrub
currently
only is permitted on a file system with a single rank. Reduce the
ranks to one
via `ceph fs set <fs_name> max_mds 1`.
* Ceph now refuses to create a file system with a default EC data pool.
For
further explanation, see:
https://docs.ceph.com/docs/nautilus/cephfs/createfs/#creating-pools
* Ceph will now issue a health warning if a RADOS pool has a `pg_num`
value that is not a power of two. This can be fixed by adjusting
the pool to a nearby power of two::
ceph osd pool set <pool-name> pg_num <new-pg-num>
Alternatively, the warning can be silenced with::
ceph config set global mon_warn_on_pool_pg_num_not_power_of_two
false
Getting Ceph
------------
* Git at
git://github.com/ceph/ceph.git
* Tarball at
http://download.ceph.com/tarballs/ceph-14.2.8.tar.gz
* For packages, see
http://docs.ceph.com/docs/master/install/get-packages/
* Release git sha1: 2d095e947a02261ce61424021bb43bd3022d35cb
--
Abhishek Lekshmanan
SUSE Software Solutions Germany GmbH
GF: Felix Imendörffer HRB 21284 (AG Nürnberg)
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io To unsubscribe send an
email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io