On 10/27/19 6:01 AM, Frank R wrote:
> I hate to be a pain but I have one more question.
>
> After I run
>
> radosgw-admin reshard stale-instances rm
>
> if I run
>
> radosgw-admin reshard stale-instances list
>
> some new entries appear for a bucket that no longer exists. Is there a
> way to cancel the operation on the old bucket?
>
`radosgw-admin reshard stale-instances rm` should fix all your issues,
if this not you should debug this.
k
This seems to happen mostly when listing folders containing 10k+ folders.
The dirlisting hangs indefinitely or until i restart the active MDS and
then the hanging "ls" command will finish running.
Every time restarting the active MDS fixes the problem for a while.
Hi Everyone,
So, I'm in the process of trying to migrate our rgw.buckets.data pool from
a replicated rule pool to an erasure coded pool. I've gotten the EC pool
set up, good EC profile and crush ruleset, pool created successfully, but
when I go to "rados cppool xxx.rgw.buckets.data xxx.rgw.buckets.data.new",
I get this error after it transfers 4GB of data:
error copying object: (2) No such file or directory
error copying pool xxx.rgw.buckets.data => xxx.rgw.buckets.data.new: (2) No
such file or directory
Is "rados cppool" still the blessed way to do the migration, or has
something better/not deprecated been developed that I can use?
Thanks,
Mac
Hi all,
Does anyone have a good config for lower memory radosgw machines?
We have 16GB VMs and our radosgw's go OOM when we have lots of
parallel clients (e.g. I see around 500 objecter_ops via the rgw
asok).
Maybe lowering rgw_thread_pool_size from 512 would help?
(This is running latest luminous).
Thanks, Dan
Hi all,
I want to log IP client to rados gateway log to check information about
loadbalancing and other things. I using LB before rados gateway nodes, what
need to be done for configuration in rados gateway?
thank you very much.
Br,
----------------------------------------------
Dương Tuấn Dũng
Email: dungdt.aicgroup(a)gmail.com
ĐT: 0986153686
I have 104 pg stays in unknown states for a long time
[root@node-1 /]# ceph -s
cluster:
id: 653c6c1a-607e-4a62-bb92-dfe2f0d7afb6
health: HEALTH_ERR
1 osds down
Reduced data availability: 104 pgs inactive
24 slow requests are blocked > 32 sec. Implicated osds 0,1,2,8,9,10
14 stuck requests are blocked > 4096 sec. Implicated osds 5,6
services:
mon: 3 daemons, quorum node-1,node-2,node-3
mgr: node-1(active), standbys: node-2, node-3
osd: 12 osds: 11 up, 12 in
flags nodeep-scrub
rbd-mirror: 1 daemon active
data:
pools: 7 pools, 360 pgs
objects: 1.80k objects, 3.91GiB
usage: 17.6GiB used, 7.96TiB / 7.98TiB avail
pgs: 28.889% pgs unknown
256 active+clean
104 unknown
io:
client: 1.56MiB/s wr, 0op/s rd, 83op/s wr
[root@node-1 /]# ceph health detail
HEALTH_ERR Reduced data availability: 104 pgs inactive; 30 slow
requests are blocked > 32 sec. Implicated osds 0,1,2,4,8,9,10; 14
stuck requests are blocked > 4096 sec. Implicated osds 5,6
PG_AVAILABILITY Reduced data availability: 104 pgs inactive
pg 1.0 is stuck inactive for 2857.069686, current state unknown,
last acting []
pg 1.1 is stuck inactive for 2857.069686, current state unknown,
last acting []
pg 1.2 is stuck inactive for 2857.069686, current state unknown,
last acting []
pg 1.3 is stuck inactive for 2857.069686, current state unknown,
last acting []
pg 1.4 is stuck inactive for 2857.069686, current state unknown,
last acting []
pg 1.5 is stuck inactive for 2857.069686, current state unknown,
last acting []
pg 1.6 is stuck inactive for 2857.069686, current state unknown,
last acting []
pg 1.7 is stuck inactive for 2857.069686, current state unknown,
last acting []
pg 2.0 is stuck inactive for 2857.069686, current state unknown,
last acting []
......
[root@node-1 /]# ceph pg dump_stuck inactive
ok
PG_STAT STATE UP UP_PRIMARY ACTING ACTING_PRIMARY
3.1d unknown [] -1 [] -1
3.1c unknown [] -1 [] -1
3.1b unknown [] -1 [] -1
3.1a unknown [] -1 [] -1
3.19 unknown [] -1 [] -1
......
my pool size = 3
[root@node-1 /]# ceph pg 3.1d query
Error ENOENT: i don't have pgid 3.1d
Hi,
in my unhealthy cluster I cannot run several ceph osd command because
they hang, e.g.
ceph osd df
ceph osd pg dump
Also, ceph balancer status hangs.
How can I fix this issue?
THX
I'm wondering if it's possible to enable compression on existing RGW buckets? The cluster is running Luminous 12.2.12 with FileStore as the backend (no BlueStore compression then).
We have a cluster that recently started to rapidly fill up with compressible content (qcow2 images) and I would like to enable compression for new uploads to slow the growth. The documentation seems to imply that changing zone placement rules can only be done at creation time. Is there something I'm missing that would allow me to enable compression on a per-bucket or even a per-user basis after a cluster has been used for quite a while?
Thanks,
Bryan