Hi,
any chance configuration of dev(a)ceph.io can be changed in a way that the
mailing list is not been spammed?
A quick search of mails in September showed my a list of addresses where
most of the have sent spam mails several times. It might make sense to
unsubscribe and block them. Hand-selecting all subscribers from domain
@126.com in future might also be a good idea.
chezhiting70437(a)126.com
feijuwei714(a)126.com
MarilynnGardemalGWJ(a)yahoo.com
kuanshaodui658(a)126.com
lianglu769(a)126.com
liangzhun70(a)126.com
liaojie8206855(a)126.com
rongsha66178(a)126.com
shaozhuo1368(a)126.com
sizhaogou8020114(a)126.com
tulian17573320(a)126.com
yizi5159030833(a)126.com
zhanhuang669(a)126.com
zhongji818027255(a)126.com
wadeyuy(a)foxmail.com
harry879y(a)foxmail.com
info(a)vixech.info
xinyuezhang(a)mail.tsinghua.edu.cn
Thanks a lot.
Best wishes
Holger
--
SUSE Software Solutions Germany GmbH
Maxfeldstr. 5, 90409 Nuernberg, Germany
Geschäftsführer: Felix Imendörffer, HRB 247165 (AG München)
Hi,
We implemented a new cache tier mode - local mode. In this mode, an
osd is configured to manage two data devices, one is fast device, one
is slow device. Hot objects are promoted from slow device to fast
device, and demoted from fast device to slow device when they become
cold.
The introduction of tier local mode in detail is
https://tracker.ceph.com/issues/42286
tier local mode: https://github.com/yanghonggang/ceph/commits/wip-tier-new
This work is based on ceph v12.2.5. I'm glad to port it to master
branch if needed.
Any advice and suggestions will be greatly appreciated.
thx,
Yang Honggang
While going through the MDS code i could see Base/System Inode but didn't
understand what it was( how it's special? ). Can someone help me understand
it?
Regards,
Sidharth Anupkrishnan
https://tracker.ceph.com/issues/41834
This is broadly affecting Ceph QA. Hoping this mail will get the
notice of the person whose changes maybe broke it.
--
Patrick Donnelly, Ph.D.
He / Him / His
Senior Software Engineer
Red Hat Sunnyvale, CA
GPG: 19F28A586F808C2402351B93C3301A3E258DD79D
On raspberry pi 4 (arm v7 architecture), I am able to do ./install_deps as well as ./do_cmake but when I make, I get the following compile error
/home/sharath/os/ceph/src/include/buffer.h:740:12: error: 'void ceph::buffer::v14_2_0::list::iterator_impl<is_const>::advance(size_t)' cannot be overloaded with 'void ceph::buffer::v14_2_0::list::iterator_impl<is_const>::advance(unsigned int)'
void advance(size_t o) { advance(static_cast<unsigned>(o)); }
^~~~~~~
my gcc version is
sharath@raspberrypi:~/os/ceph/build $ gcc --version
gcc (Raspbian 8.3.0-6+rpi1) 8.3.0
excerpt from ./do_cmake
+ cmake -DCMAKE_BUILD_TYPE=Debug -DCMAKE_BUILD_TYPE=RelWithDebInfo ..
-- The CXX compiler identification is GNU 8.3.0
-- The C compiler identification is GNU 8.3.0
-- The ASM compiler identification is GNU
-- Found assembler: /usr/bin/cc
-- Check for working CXX compiler: /usr/bin/c++
-- Check for working CXX compiler: /usr/bin/c++ -- works
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Check for working C compiler: /usr/bin/cc
-- Check for working C compiler: /usr/bin/cc -- works
Please help
---------- Forwarded message ---------
发件人: Xiangyang Yu <penglaiyxy(a)gmail.com>
Date: 2019年10月15日周二 上午11:32
Subject: OSD reconnected across map epochs, inconsistent pg logs created
To: ceph-devel <ceph-devel(a)vger.kernel.org>
Cc: Gregory Farnum <gfarnum(a)redhat.com>
Hi cephers,
I met a rare case and got inconsistent objects.
As described below:
Assume pg 1.1a maps to osds[1,5,9], osd1 is the primary osd.
Time 1: osd1 osd5 osd9 was online and could send message to each other.
Time 2: old5, osd9 received an new osdmap that showed osd.1 was down
,and at the same time, osd1’s public network was down
manually(physical down),but osd.0’s cluster network is still online.
Time 3:
Because of receiving an new osdmap that showed osd1 was down, osd5 and
osd9 shutdowned their connections towards osd1 up (through mark_down()
). so there were no existing connections for osd1.
As for osd.1, connections between osd.5/osd.9 encountered a
failure(disconnected by osd.5/osd.9 explicitly) and were going to
enter STANDBY state . As a consequence, these connections were still
existing( their cs_seq > 0).
After a short while, osd1 generated two scrub operations(enable
deep-scrub) about updating some objects version
info(scrub_snapshot_metadata()), and was going to reestablish
connections among osd5 and osd9. When osd1 was sending the first
operation op1(by send_message()), the cluster messenger would
reconnect the osd5/osd9 and then placing the op1 in out_q。During the
connection was enter STATE_OPEN, there was a RESETSESSION between osd1
and osd5/osd9, which lead osd1 to discard the msg in out_q (by
was_session_reset()). After the connection was established, osd1 sent
the second operation op2 to osd5/osd9. Have this attention : OSD.5 and
OSD.9 was updating the osdmap, but have not committed the update. So
op2 did not discard because of mismatch epoch and eventually applied
to osd.5 and osd.9.
Eventually, there two pg log were recorded on osd1(op1,op2), but only
one pg log(op2) on osd5/osd9.
Time4: when osd1 public network recovered soon, during pg peering, the
primary osd(osd1) could not find any difference about pg log among
osd5 and osd9. When pg 1.1a deep-scrubed over, there would trigger an
inconsistent error about object version info(the version info op1
associatived).
This is a rarely situation we meet with. In some case, I think this
would cause the msgs out of order . If I misdiagnosed it,please tell
me.
I have talked with gerg about the problem before.
This is my PR:
https://github.com/ceph/ceph/pull/30609
(I insist on my pull request, because in my opinion, there is no
difference between peer and losslyness connections, any endpoint can
connect to another endpoint. If the osd.5 and osd.9 start a connection
,then the op1 in osd.1 can not discard and would walk through a
replace session flow. )
The event is recorded in tracker:
https://tracker.ceph.com/issues/42058
Anyway, I want to receive any advice on how to resolve the problem
properly, thanks.