Sorry I forgot to reply to this -- yes it's all good after you re-signed
and pushed.
Thanks!
dan
On Thu, Dec 12, 2019 at 9:57 PM David Galloway <dgallowa(a)redhat.com> wrote:
I just re-signed and pushed the 14.2.5 packages after
adding the
--no-database fix. Can you confirm ceph-debuginfo installs as expected
now please?
On 12/12/19 9:05 AM, Dan van der Ster wrote:
Thanks for this!
We're having trouble installing the ceph-debuginfo rpm, which shows a
negative file size (apparently because it now exceeds 2GB):
https://pastebin.com/5bnNCGHh
Does this change need to be applied to the release builds too?
https://tracker.ceph.com/issues/39387
-- Dan
On Tue, Dec 10, 2019 at 10:45 AM Abhishek Lekshmanan <abhishek(a)suse.com>
wrote:
> This is the fifth release of the Ceph
Nautilus release series. Among
the many
> notable changes, this release fixes a
critical BlueStore bug that was
introduced
> in 14.2.3. All Nautilus users are advised to
upgrade to this release.
>
> For the complete changelog entry, please visit the release blog at
>
https://ceph.io/releases/v14-2-5-nautilus-released/
>
> Notable Changes
> ---------------
>
> Critical fix:
>
> * This release fixes a `critical BlueStore bug <
https://tracker.ceph.com/issues/42223>`_
> introduced in 14.2.3 (and also present in
14.2.4) that can lead to
data
> corruption when a separate "WAL"
device is used.
>
> New health warnings:
>
> * Ceph will now issue health warnings if daemons have recently crashed.
Ceph
> has been collecting crash reports since the
initial Nautilus release,
but the
> health alerts are new. To view new crashes
(or all crashes, if you've
just
> upgraded)::
>
> ceph crash ls-new
>
> To acknowledge a particular crash (or all crashes) and silence the
health
warning::
>
> ceph crash archive <crash-id>
> ceph crash archive-all
>
> * Ceph will now issue a health warning if a RADOS pool has a ``pg_num``
> value that is not a power of two. This can be fixed by adjusting
> the pool to a nearby power of two::
>
> ceph osd pool set <pool-name> pg_num <new-pg-num>
>
> Alternatively, the warning can be silenced with::
>
> ceph config set global mon_warn_on_pool_pg_num_not_power_of_two
false
>
> * Ceph will issue a health warning if a RADOS pool's ``size`` is set to
1
> or, in other words, if the pool is
configured with no redundancy.
Ceph will
> stop issuing the warning if the pool size
is set to the minimum
> recommended value::
>
> ceph osd pool set <pool-name> size <num-replicas>
>
> The warning can be silenced with::
>
> ceph config set global mon_warn_on_pool_no_redundancy false
>
> * A health warning is now generated if the average osd heartbeat ping
> time exceeds a configurable threshold for any of the intervals
> computed. The OSD computes 1 minute, 5 minute and 15 minute
> intervals with average, minimum and maximum values. New configuration
> option `mon_warn_on_slow_ping_ratio` specifies a percentage of
> `osd_heartbeat_grace` to determine the threshold. A value of zero
> disables the warning. New configuration option
`mon_warn_on_slow_ping_time`
> specified in milliseconds over-rides the
computed value, causes a
warning
> when OSD heartbeat pings take longer than
the specified amount.
> A new admin command, `ceph daemon mgr.# dump_osd_network
[threshold]`, will
> list all connections with a ping time
longer than the specified
threshold or
> value determined by the config options, for
the average for any of
the 3 intervals.
> Another new admin command, `ceph daemon
osd.# dump_osd_network
[threshold]`,
> will do the same but only including
heartbeats initiated by the
specified OSD.
>
> Changes in the telemetry module:
>
> * The telemetry module now has a 'device' channel, enabled by default,
that
> will report anonymized hard disk and SSD
health metrics to
telemetry.ceph.com
> in order to build and improve device
failure prediction algorithms.
Because
> the content of telemetry reports has
changed, you will need to
re-opt-in
> with::
>
> ceph telemetry on
>
> You can view exactly what information will be reported first with::
>
> ceph telemetry show
> ceph telemetry show device # specifically show the device channel
>
> If you are not comfortable sharing device metrics, you can disable
that
> channel first before re-opting-in:
>
> ceph config set mgr mgr/telemetry/channel_crash false
> ceph telemetry on
>
> * The telemetry module now reports more information about CephFS file
systems,
> including:
>
> - how many MDS daemons (in total and per file system)
> - which features are (or have been) enabled
> - how many data pools
> - approximate file system age (year + month of creation)
> - how many files, bytes, and snapshots
> - how much metadata is being cached
>
> We have also added:
>
> - which Ceph release the monitors are running
> - whether msgr v1 or v2 addresses are used for the monitors
> - whether IPv4 or IPv6 addresses are used for the monitors
> - whether RADOS cache tiering is enabled (and which mode)
> - whether pools are replicated or erasure coded, and
> which erasure code profile plugin and parameters are in use
> - how many hosts are in the cluster, and how many hosts have each
type of
daemon
> - whether a separate OSD cluster network
is being used
> - how many RBD pools and images are in the cluster, and how many
pools
have RBD mirroring enabled
> - how many RGW daemons, zones, and
zonegroups are present; which
RGW frontends are in use
> - aggregate stats about the CRUSH map,
like which algorithms are
used, how
> big buckets are, how many rules are
defined, and what tunables
are in
> use
>
> If you had telemetry enabled, you will need to re-opt-in with::
>
> ceph telemetry on
>
> You can view exactly what information will be reported first with::
>
> ceph telemetry show # see everything
> ceph telemetry show basic # basic cluster info (including all of
the new
info)
>
> OSD:
>
> * A new OSD daemon command, 'dump_recovery_reservations', reveals the
> recovery locks held (in_progress) and waiting in priority queues.
>
> * Another new OSD daemon command, 'dump_scrub_reservations', reveals the
> scrub reservations that are held for local (primary) and remote
(replica)
PGs.
>
> RGW:
>
> * RGW now supports S3 Object Lock set of APIs allowing for a WORM model
for
> storing objects. 6 new APIs have been added
put/get bucket object
lock,
> put/get object retention, put/get object
legal hold.
>
> * RGW now supports List Objects V2
>
> Getting Ceph
> ------------
>
> * Git at
git://github.com/ceph/ceph.git
> * Tarball at
http://download.ceph.com/tarballs/ceph-14.2.5.tar.gz
> * For packages, see
http://docs.ceph.com/docs/master/install/get-packages/
> * Release git sha1:
ad5bd132e1492173c85fda2cc863152730b16a92
>
> --
> Abhishek Lekshmanan
> SUSE Software Solutions Germany GmbH