Hello,
i played around with some log level i can´t remember and my logs are
now getting bigger than my DVD-Movie collection.
E.g.: journalctl -b -u
ceph-5436dd5d-83d4-4dc8-a93b-60ab5db145df(a)mon.ceph03.service >
out.file is 1,1GB big.
I did already try:
ceph tell mon.ceph03 config set debug_mon 0/10
ceph tell mon.ceph03 config set debug_osd 0/10
ceph tell mon.ceph03 config set debug_mgr 0/10
ceph tell mon.ceph03 config set "mon_health_to_clog" false
ceph tell mon.ceph03 config set "mon_health_log_update_period" 30
ceph tell mon.ceph03 config set "debug_mgr" "0/0"
which made it better, but i really cant remember it all and would like
to have the default values.
Is there a way to reset those Log Values?
Cheers,
Michael
Hi Friends,
We have some inconsistent storage space usage reporting. We used only 46TB
with single copy but the space used on the pool is close to 128TB.
Any idea where's the extra space is utilized and how to reclaim it?
Ceph Version : 12.2.11 with XFS OSDs. We are planning to upgrade soon.
# ceph df detail
GLOBAL:
SIZE AVAIL RAW USED %RAW USED OBJECTS
363TiB 131TiB 231TiB 63.83 43.80M
POOLS:
NAME ID QUOTA OBJECTS QUOTA BYTES USED %USED
MAX AVAIL OBJECTS DIRTY READ WRITE RAW USED
fcp 15 N/A N/A 23.6TiB 42.69
31.7TiB 3053801 3.05M 6.10GiB 12.6GiB 47.3TiB
nfs 16 N/A N/A 128TiB 66.91
63.4TiB 33916181 33.92M 3.93GiB 4.73GiB 128TiB
# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/nbd0 200T 46T 155T 23% /vol/dir_research
#ceph osd pool get nfs all
size: 1
min_size: 1
crash_replay_interval: 0
pg_num: 128
pgp_num: 128
crush_rule: replicated_ruleset
hashpspool: true
nodelete: false
nopgchange: false
nosizechange: false
write_fadvise_dontneed: false
noscrub: false
nodeep-scrub: false
use_gmt_hitset: 1
auid: 0
fast_read: 0
Appreciate your help.
Thanks,
-Vikas
Hello List,
sometimes some OSD get taken our for some reason ( i am still looking
for the reason, and i guess its due to some overload), however, when i
try to restart them i get:
Nov 02 08:05:26 ceph05 bash[9811]: Error: No such container:
ceph-5436dd5d-83d4-4dc8-a93b-60ab5db145df-osd.47
Nov 02 08:05:29 ceph05 bash[9811]: /usr/bin/docker: Error response
from daemon: endpoint with name
ceph-5436dd5d-83d4-4dc8-a93b-60ab5db145df-osd.47 already exists in
network host.
Nov 02 08:05:29 ceph05 systemd[1]:
ceph-5436dd5d-83d4-4dc8-a93b-60ab5db145df(a)osd.47.service: Main process
exited, code=exited, status=125/n/a
Nov 02 08:05:34 ceph05 systemd[1]:
ceph-5436dd5d-83d4-4dc8-a93b-60ab5db145df(a)osd.47.service: Failed with
result 'exit-code'.
Nov 02 08:05:44 ceph05 systemd[1]:
ceph-5436dd5d-83d4-4dc8-a93b-60ab5db145df(a)osd.47.service: Service
RestartSec=10s expired, scheduling restart.
Nov 02 08:05:44 ceph05 systemd[1]:
ceph-5436dd5d-83d4-4dc8-a93b-60ab5db145df(a)osd.47.service: Scheduled
restart job, restart counter is at 5.
Nov 02 08:05:44 ceph05 systemd[1]: Stopped Ceph osd.47 for
5436dd5d-83d4-4dc8-a93b-60ab5db145df.
Nov 02 08:05:44 ceph05 systemd[1]:
ceph-5436dd5d-83d4-4dc8-a93b-60ab5db145df(a)osd.47.service: Start
request repeated too quickly.
Nov 02 08:05:44 ceph05 systemd[1]:
ceph-5436dd5d-83d4-4dc8-a93b-60ab5db145df(a)osd.47.service: Failed with
result 'exit-code'.
Nov 02 08:05:44 ceph05 systemd[1]: Failed to start Ceph osd.47 for
5436dd5d-83d4-4dc8-a93b-60ab5db145df.
I need to reboot the full host to get the OSD back in again. As far i
can see this is some docker problem?
root@ceph05:~# docker ps | grep osd.47 => no hit
root@ceph05:~# docker network prune => does not solve the problem
Any hint on that?
Thanks,
Michael
Hi Frank
Found the issue and fixed. It was a one copy of 0 byte object. Removed it. Deep scrub the PG fixed the issue.
# find /var/lib/ceph/osd/ -type f -name "1000023675e*"
/var/lib/ceph/osd/ceph-2/current/3.b_head/DIR_B/DIR_A/DIR_E/1000023675e.00000000__head_AE97EEAB__3
# ls -l /var/lib/ceph/osd/ceph-2/current/3.b_head/DIR_B/DIR_A/DIR_E/1000023675e.00000000__head_AE97EEAB__3-rw-r--r-- 1 ceph ceph 0 Oct 31 19:18 /var/lib/ceph/osd/ceph-2/current/3.b_head/DIR_B/DIR_A/DIR_E/1000023675e.00000000__head_AE97EEAB__3
Once again, many thanks for your help.
Best regards
Sagara
Hi all,
We still have legacy caps on our nautilus rbd cluster. I just wanted
to check if this is totally safe (and to post here ftr because I don't
think this has ever been documented)
Here are the current caps:
[client.images]
key = xxx
caps mgr = "allow r"
caps mon = "allow r, allow command \"osd blacklist\""
caps osd = "allow class-read object_prefix rbd_children, allow rwx pool=images"
[client.volumes]
key = xxx
caps mgr = "allow r"
caps mon = "allow r, allow command \"osd blacklist\""
caps osd = "allow class-read object_prefix rbd_children, allow rwx
pool=volumes, allow rx pool=images, allow rwx pool=cinder-critical"
Now that we upgraded to nautilus we would do:
# ceph auth caps client.images mon 'profile rbd' osd 'profile rbd
pool=images' mgr 'profile rbd pool=images'
# ceph auth caps client.volumes mon 'profile rbd' osd 'profile rbd
pool=volumes, profile rbd-read-only pool=images, profile rbd
pool=cinder-critical' mgr 'profile rbd pool=volumes, profile rbd
pool=cinder-critical'
Does that look correct? Does this apply without impacting any client IOs ?
Thanks!
Dan
Dear Ceph folks,
I am faced with a special use Ceph scenario, where no SSD is available. Three node Ceph cluster, each with 14 HDDs, no SSD. In this case, should i create a separate DB/WAL partition for each OSD on a HDD? and if yes, how big the partition?
Which is the best in terms of performance, with a separate DB/WAL partition, or without?
Thanks a lot,
samuel
huxiaoyu(a)horebdata.cn
Hi all
I have a Ceph cluster (Nautilus 14.2.11) with 3 Ceph nodes.
A crash happened and all 3 Ceph nodes went down.
One (1) PG turned "active+clean+inconsistent", I tried to repair it. After the repair, now shows "active+clean+inconsistent+failed_repair" for the PG in the question and cannot bring the cluster to "active+clean".
How do I rescue the cluster? Is this a false positive?
Here are the detail:
All three Ceph nodes run ceph-mon, ceph-mgr, ceph-osd and ceph-mds.
1. ceph -s
health: HEALTH_ERR 3 scrub errors Possible data damage: 1 pg inconsistent
pgs: 191 active+clean 1 active+clean+inconsistent
2. ceph health detailHEALTH_ERR 3 scrub errors; Possible data damage: 1 pg inconsistentOSD_SCRUB_ERRORS 3 scrub errorsPG_DAMAGED Possible data damage: 1 pg inconsistent pg 3.b is active+clean+inconsistent, acting [0,1,2]
3. rados list-inconsistent-pg rbd[]
4. ceph pg deep-scrub 3.b
5. ceph pg repair 3.b
6. ceph health detailHEALTH_ERR 3 scrub errors; Possible data damage: 1 pg inconsistentOSD_SCRUB_ERRORS 3 scrub errorsPG_DAMAGED Possible data damage: 1 pg inconsistent pg 3.b is active+clean+inconsistent+failed_repair, acting [0,1,2]
7. rados list-inconsistent-obj 3.b --format=json-pretty{ "epoch": 4769, "inconsistents": []}
8. ceph pg 3.b list_unfound { "num_missing": 0, "num_unfound": 0, "objects": [], "more": false}
Appreciate your help.
ThanksSagara
Hi,
AWIK, the read latency primarily depends on HW latency,
not much can be tuned in SW. Is that right?
I ran a fio random read with iodepth 1 within a VM backed by
Ceph with HDD OSD and here is what I got.
=================
read: IOPS=282, BW=1130KiB/s (1157kB/s)(33.1MiB/30001msec)
slat (usec): min=4, max=181, avg=14.04, stdev=10.16
clat (usec): min=178, max=393831, avg=3521.86, stdev=5771.35
lat (usec): min=188, max=393858, avg=3536.38, stdev=5771.51
=================
I checked HDD average latency is 2.9 ms. Looks like the test
result makes perfect sense, isn't it?
If I want to get shorter latency (more IOPS), I will have to go
for better disk, eg. SSD. Right?
Thanks!
Tony