[prev in list] [next in list] [prev in thread] [next in thread]
List: ceph-users
Subject: [ceph-users] Re: High ceph_osd_commit_latency_ms on Toshiba MG07ACA14TE HDDs
From: Mark Nelson <mnelson () redhat ! com>
Date: 2021-11-04 10:47:50
Message-ID: 156922e2-9d9c-8cda-f329-c27dc5c84979 () redhat ! com
[Download RAW message or body]
Hi Dan,
I can't speak for those specific Toshiba drives, but we have absolutely
seen very strange behavior (sometimes with cache enabled and sometimes
not) with different drives and firmwares over the years from various
manufacturers. There was one especially bad case from back in the
Inktank days, but my memory is a bit fuzzy. I think we were seeing
weird periodic commit latency spikes that grew worse over time. That
one might have been cache related. I believe we ended up doing a lot of
tests with blktrace and iowatcher to show the manufacturer what we were
seeing, but I don't recall if anything ever got fixed.
Mark
On 11/4/21 5:33 AM, Dan van der Ster wrote:
> Hello Benoît, (and others in this great thread),
>
> Apologies for replying to this ancient thread.
>
> We have been debugging similar issues during an ongoing migration to
> new servers with TOSHIBA MG07ACA14TE hdds.
>
> We see a similar commit_latency_ms issue on the new drives (~60ms in
> our env vs ~20ms for some old 6TB Seagates).
> However, disabling the write cache (hdparm -W 0) made absolutely no
> difference for us.
>
> So we're wondering:
> * Are we running the same firmware as you? (We have 0104). I wonder if
> Toshiba has changed the implementation of the cache in the meantime...
> * Is anyone aware of some HBA or other setting in the middle that
> might be masking this setting from reaching the drive?
>
> Best Regards,
>
> Dan
>
>
>
> On Wed, Jun 24, 2020 at 9:44 AM Benoît Knecht <bknecht@protonmail.ch> wrote:
> > Hi,
> >
> > We have a Nautilus (14.2.9) Ceph cluster with two types of HDDs:
> >
> > - TOSHIBA MG07ACA14TE [1]
> > - HGST HUH721212ALE604 [2]
> >
> > They're all bluestore OSDs with no separate DB+WAL and part of the same pool.
> >
> > We noticed that while the HGST OSDs have a commit latency of about 15ms, the \
> > Toshiba OSDs hover around 150ms (these values come from the \
> > `ceph_osd_commit_latency_ms` metric in Prometheus).
> > On paper, it seems like those drives have very similar specs, so it's not clear \
> > to me why we're seeing such a large difference when it comes to commit latency.
> > Has anyone had any experience with those Toshiba drives? Or looking at the specs, \
> > do you spot anything suspicious?
> > And if you're running a Ceph cluster with various disk brands/models, have you \
> > ever noticed some of them standing out when looking at \
> > `ceph_osd_commit_latency_ms`?
> > Thanks in advance for your feedback.
> >
> > Cheers,
> >
> > --
> > Ben
> >
> > [1]: https://toshiba.semicon-storage.com/content/dam/toshiba-ss/asia-pacific/docs/product/storage/product-manual/eHDD-MG07ACA-Product-Manual.pdf
> > [2]: https://documents.westerndigital.com/content/dam/doc-library/en_us/assets/pu \
> > blic/western-digital/product/data-center-drives/ultrastar-dc-hc500-series/data-sheet-ultrastar-dc-hc520.pdf
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-leave@ceph.io
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-leave@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-leave@ceph.io
[prev in list] [next in list] [prev in thread] [next in thread]
Configure |
About |
News |
Add a list |
Sponsored by KoreLogic