I updated teuthology yesterday and since then have seen a log of the
following errors
...src/teuthology/virtualenv/local/lib/python2.7/site-packages/paramiko/ecdsakey.py:164:
CryptographyDeprecationWarning: Support for unsafe construction of
public numbers from encoded data will be removed in a future version.
Please use EllipticCurvePublicKey.from_encoded_point
self.ecdsa_curve.curve_class(), pointinfo
2019-07-31 01:45:18,976.976 ERROR:paramiko.transport:Exception: Error
reading SSH protocol banner
2019-07-31 01:45:18,976.976 ERROR:paramiko.transport:Traceback (most
recent call last):
2019-07-31 01:45:18,976.976 ERROR:paramiko.transport: File
"/home/bhubbard/src/teuthology/virtualenv/local/lib/python2.7/site-packages/paramiko/transport.py",
line 1966, in run
2019-07-31 01:45:18,976.976 ERROR:paramiko.transport: self._check_banner()
2019-07-31 01:45:18,977.977 ERROR:paramiko.transport: File
"/home/bhubbard/src/teuthology/virtualenv/local/lib/python2.7/site-packages/paramiko/transport.py",
line 2143, in _check_banner
2019-07-31 01:45:18,977.977 ERROR:paramiko.transport: "Error
reading SSH protocol banner" + str(e)
2019-07-31 01:45:18,977.977 ERROR:paramiko.transport:SSHException:
Error reading SSH protocol banner
Sometimes these are fatal and sometimes not. Wondering if anyone else
has seen them?
--
Cheers,
Brad
David G,
I've been looking over logs and ceph pg dump pgs on the LRC and things
look good to me. If you see anything not working file a tracker or if
you have any questions please contact me.
There is one thing that you should be aware of. There are still
filestore objectstores for some of the OSDs. The auto_repair feature is
not supported for filestore, so when they deep-scrub they won't repair.
With auto_repair enabled in this mixed cluster the LRC will auto_repair
if the primary OSD for a PG is bluestore even if some replicas are
filestore. So I would convert the the remaining filestore OSDs to
bluestore. If you are paranoid you should disable auto_repair until the
conversion is completed.
David Z
On 7/2/19 3:11 PM, David Zafman wrote:
>
> I don't see that now in ceph status. A pg's deep scrub would have to
> be over 5 days overdue for that warning to occur.
>
> David
>
> On 7/2/19 2:29 PM, David Galloway wrote:
>> This build is installed now.
>>
>> It looks like "1 pgs not scrubbed in time" is back.
>>
>> On 6/28/19 12:27 PM, David Zafman wrote:
>>> David,
>>>
>>> I have a new scrub handling code built for Nautilus. Could we
>>> install this on the LRC to see how well it works in a more realistic
>>> environment?
>>>
>>> https://shaman.ceph.com/builds/ceph/wip-zafman-testing-nautilus/31ff31f2c8d…
>>>
>>>
>>>
>>> Thanks
>>>
>>> David Zafman
>>>