Hi Folks,
We are running ceph 14.2.16 and I like to reshard a bucket because I
have a large object warning!
so I did:
radosgw-admin bucket reshard --tenant="..." --bucket="..." --uid="..."
--num-shards=512
but I got receive an error:
ERROR: the bucket is currently undergoing resharding and cannot be
added to the reshard list at this time
`radosgw-admin reshard list` is empty so I assume I have to delete
some leftovers from the old resharding!? did someone has had this
before?
thanks for your input,
Ansgar
Hello,
I'm in the process of doubling the number of OSD nodes in my Nautilus
cluster - from 3 to 6. Based on answers receive from earlier posts to
this list, the new nodes have more NVMe that the old nodes. More to the
point, on the original nodes the amount of NVMe allocated to each OSD
was about 120GB, so the RocksDB was limited to 30GB. However, for my
workload 300GB is probably recommended.
As I prepare to lay out the NVMe on these new nodes, I'm still trying to
understand how to size the DB and WAL for my OSDs and whether Journal is
even needed.
According to https://docs.ceph.com/en/nautilus/ceph-volume/lvm/prepare/
<https://docs.ceph.com/en/nautilus/ceph-volume/lvm/prepare/>
> Bluestore supports the following configurations:
>
> * A block device, a block.wal, and a block.db device
> * A block device and a block.wal device
> * A block device and a block.db device
> * A single block device
>
First question: On my first nodes I managed to get a DB, but no WAL.
My current perception is that WAL and DB occupy separate
physical/logical partitions. By specifying a WAL size and a DB size,
ceph-volume will create the corresponding logical volumes on the NVMe.
Is this correct? It is also possible to lay these out as basic logical
partitions?
Second question: How do I decide whether I need WAL, DB, or both?
Third question: Once I answer the above WAL/DB question, what are the
guidelines for sizing them?
Thanks.
-Dave
--
Dave Hall
Binghamton University
1. is backfilling/remapping so smart that it will do what ever it can? Or are there situations like a pg a. is scheduled to be moved, but can not be moved because of min_size. Now another pg b. cannot be moved because pg a. allocated osd space and the backfill ratio will be met.
Yet if the order was reversed if the pg b. would be moved and pg a. would not be moved because it is stuck anyway and the backfill ratio will be met.
2. If a down host comes up again and it's osd are started. Is data still being copied, or does ceph see that checksums(?) are the same and just sets a pointer(?) back to the old location?
> -----Original Message-----
> Sent: 09 March 2021 23:59
> To: ceph-users(a)ceph.io
> Subject: [ceph-users] node down pg with backfill_wait waiting for
> incomplete?
>
>
> I have a node down and pg's are remapping/backfilling. I have also a lot
> of pg's in backfill_wait.
>
> I was wondering if there is a specific order that this is being
> executed. Eg I have a large'garbage' pool ec21 that is stuck. I could
> resolve that by changing the min size. However I rather have this not
> remapped, and wait for this node to be back on line.
>
> Question is: is other remapping/backfilling waiting for this stuck to be
> fixed, or is backfilling/remapping so smart that it will do what ever it
> can?
>
>
>
>
>
>
> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -.
> F1 Outsourcing Development Sp. z o.o.
> Poland
>
> t: +48 (0)12 4466 845
> f: +48 (0)12 4466 843
> e: marc(a)f1-outsourcing.eu
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users(a)ceph.io
> To unsubscribe send an email to ceph-users-leave(a)ceph.io
Dear Cepher,
For some reasons, i had a cluster with several 20 TB pools and 100TB ones, which were previously linked with iSCSI for virtual machines. When deleting those big rbd images, it turns out to be extremely slow, taking hours if not days. The Ceph cluster is running on Luminous 12.2.13 with bluestore
How to speed up removing big rbd pools?
thanks,
samuel
huxiaoyu(a)horebdata.cn
Hi,
I am having a weird phenomenon, which I am having trouble to debug. We
have 16 OSDs per host, so when I reboot one node, 16 OSDs will be
missing for a short time. Since our minimum CRUSH failure domain is
host, this should not cause any problems. Unfortunately, I always have
handful (1-5) PGs that become inactive nonetheless and are stuck in the
state undersized+degraded+peered until the host and its OSDs are back
up. The other 2000+ PGs that are also on these OSDs do not have this
problem. In total, we have between 110 and 150 PGs per OSD with a
configured maximum of 250, which should give us enough headspace.
The affected pools always seem to be RBD pools or at least I haven't
seen it on our much larger RGW pool yet. The pool's CRUSH rule looks
like this:
rule rbd-data {
id 8
type replicated
min_size 2
max_size 10
step take default
step chooseleaf firstn 0 type host
step emit
}
ceph pg dump_stuck inactive gives me this:
PG_STAT STATE UP UP_PRIMARY ACTING
ACTING_PRIMARY
115.3 undersized+degraded+peered [194,267] 194
[194,267] 194
115.13 undersized+degraded+peered [151,1122] 151
[151,1122] 151
116.12 undersized+degraded+peered [288,726] 288
[288,726] 288
and when I query one of the inactive PGs, I see (among other things):
"up": [
288,
726
],
"acting": [
288,
726
],
"acting_recovery_backfill": [
"288",
"726"
],
"recovery_state": [
{
"name": "Started/Primary/Active",
"enter_time": "2021-03-10T16:23:09.301174+0100",
"might_have_unfound": [],
"recovery_progress": {
"backfill_targets": [],
"waiting_on_backfill": [],
"last_backfill_started": "MIN",
"backfill_info": {
"begin": "MIN",
"end": "MIN",
"objects": []
},
"peer_backfill_info": [],
"backfills_in_flight": [],
"recovering": [],
"pg_backend": {
"pull_from_peer": [],
"pushing": []
}
}
},
{
"name": "Started",
"enter_time": "2021-03-10T16:23:08.297622+0100"
}
],
So you can see that two out of three OSDs on other hosts are indeed up
and active and the . I also see the ceph-osd daemons running on those
hosts, so the data is definitely there and the PG should be available.
Do you have any idea why these PGs may be becoming inactive nonetheless?
I am suspecting some kind of concurrency limit, but I wouldn't know
which one that could be.
Thanks
Janek
Ok I fix it works
-----Message d'origine-----
De : St-Germain, Sylvain (SSC/SPC) <sylvain.st-germain(a)canada.ca>
Envoyé : 9 mars 2021 17:41
À : St-Germain, Sylvain (SSC/SPC) <sylvain.st-germain(a)canada.ca>; ceph-users(a)ceph.io
Objet : RE: Rados gateway basic pools missing
Ok in the interface when I create a bucket the index in created automatically
1 device_health_metrics
2 cephfs_data
3 cephfs_metadata
4 .rgw.root
5 default.rgw.log
6 default.rgw.control
7 default.rgw.meta
8 default.rgw.buckets.index
* I think I just could not make an insertion using s3cmd
List command - connection problem
# s3cmd la
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
An unexpected error has occurred.
Please try reproducing the error using
the latest s3cmd code from the git master
branch found at:
https://github.com/s3tools/s3cmd
and have a look at the known issues list:
https://github.com/s3tools/s3cmd/wiki/Common-known-issues-and-their-solutio…
If the error persists, please report the
following lines (removing any private
info as necessary) to:
s3tools-bugs(a)lists.sourceforge.net
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
Invoked as: /usr/bin/s3cmd la
Problem: <class 'ConnectionRefusedError: [Errno 111] Connection refused
S3cmd: 2.0.2
python: 3.8.5 (default, Jan 27 2021, 15:41:15)
[GCC 9.3.0]
environment LANG=en_CA.UTF-8
Traceback (most recent call last):
File "/usr/bin/s3cmd", line 3092, in <module>
rc = main()
File "/usr/bin/s3cmd", line 3001, in main
rc = cmd_func(args)
File "/usr/bin/s3cmd", line 164, in cmd_all_buckets_list_all_content
response = s3.list_all_buckets()
File "/usr/lib/python3/dist-packages/S3/S3.py", line 302, in list_all_buckets
response = self.send_request(request)
File "/usr/lib/python3/dist-packages/S3/S3.py", line 1258, in send_request
conn = ConnMan.get(self.get_hostname(resource['bucket']))
File "/usr/lib/python3/dist-packages/S3/ConnMan.py", line 253, in get
conn.c.connect()
File "/usr/lib/python3.8/http/client.py", line 921, in connect
self.sock = self._create_connection(
File "/usr/lib/python3.8/socket.py", line 808, in create_connection
raise err
File "/usr/lib/python3.8/socket.py", line 796, in create_connection
sock.connect(sa)
ConnectionRefusedError: [Errno 111] Connection refused
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
An unexpected error has occurred.
Please try reproducing the error using
the latest s3cmd code from the git master
branch found at:
https://github.com/s3tools/s3cmd
and have a look at the known issues list:
https://github.com/s3tools/s3cmd/wiki/Common-known-issues-and-their-solutio…
If the error persists, please report the
above lines (removing any private
info as necessary) to:
s3tools-bugs(a)lists.sourceforge.net
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
-----Message d'origine-----
De : St-Germain, Sylvain (SSC/SPC) <sylvain.st-germain(a)canada.ca> Envoyé : 9 mars 2021 17:19 À : ceph-users(a)ceph.io Objet : [ceph-users] Rados gateway basic pools missing
Hi everyone,
I just rebuild a (test) cluster using :
OS : Ubuntu 20.04.2 LTS
CEPH : ceph version 15.2.9 (357616cbf726abb779ca75a551e8d02568e15b17) octopus (stable)
3 nodes : monitor/storage
1. The cluster looks good :
# ceph -s
cluster:
id: 9a89aa5a-1702-4f87-a99c-f94c9f2cdabd
health: HEALTH_OK
services:
mon: 3 daemons, quorum dao-wkr-04,dao-wkr-05,dao-wkr-06 (age 7m)
mgr: dao-wkr-05(active, since 8m), standbys: dao-wkr-04, dao-wkr-06
mds: cephfs:1 {0=dao-wkr-04=up:active} 2 up:standby
osd: 9 osds: 9 up (since 7m), 9 in (since 4h)
rgw: 3 daemons active (dao-wkr-04.rgw0, dao-wkr-05.rgw0, dao-wkr-06.rgw0)
task status:
data:
pools: 7 pools, 121 pgs
objects: 234 objects, 16 KiB
usage: 9.0 GiB used, 2.0 TiB / 2.0 TiB avail
pgs: 121 active+clean
2. except that the main pools for the radosgw are not there
# sudo ceph osd lspools
1 device_health_metrics
2 cephfs_data
3 cephfs_metadata
4 .rgw.root
5 default.rgw.log
6 default.rgw.control
7 default.rgw.meta
Missing : default.rgw.buckets.index & default.rgw.buckets.data
What do you think ?
Thx !
Sylvain
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io To unsubscribe send an email to ceph-users-leave(a)ceph.io
Hi,
I am in the process of resharding large buckets and to find them I ran
radosgw-admin bucket limit check | grep '"fill_status": "OVER' -B5
and I see that there are two buckets with negative num_objects
"bucket": "ncprod",
"tenant": "",
"num_objects": -482,
"num_shards": 0,
"objects_per_shard": -482,
"fill_status": "OVER 100.000000%"
--
"bucket": "fileshare-s3",
"tenant": "",
"num_objects": -137,
"num_shards": 0,
"objects_per_shard": -137,
"fill_status": "OVER 100.000000%"
Is this an error?
--
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
groüen Saal.