cephmetrics icon indicating copy to clipboard operation
cephmetrics copied to clipboard

Per-OSD dashboard should include read and write latency

Open dillaman opened this issue 7 years ago • 3 comments

Something like the following:

irate(ceph_osd_op_r_latency_sum[1m]) / on (ceph_daemon) irate(ceph_osd_op_r_latency_count[1m]) * 1000

irate(ceph_osd_op_w_latency_sum[1m]) / on (ceph_daemon) irate(ceph_osd_op_w_latency_count[1m]) * 1000

It might also be nice to have a global latency view on the homepage (perhaps merged w/ the apply vs commit latency graph?).

screenshot from 2018-07-12 12-22-13

dillaman avatar Jul 12 '18 12:07 dillaman

per osd read/write is a good idea. Need to think about the best place to put it though. The ceph osd information dashboard picks up all osd's so in a bigger cluster that'll be a mess. Maybe a percentile view would work under OSD Summary, and add the per osd r/w down into the filestore and bluestore sections that makes use of the osd id pulldown?

pcuzner avatar Jul 12 '18 23:07 pcuzner

Yeah — even if it’s just a top-k high latency graph w/ the ability to drill down into OSDs by host or name if needed perhaps.

dillaman avatar Jul 12 '18 23:07 dillaman

It would be nice to have views of both per OSD r/w and top-k high latency graph.

I've quick implemented the per OSD view on the ceph-overview for PoC. It seems very insightful information.

per-osd-op-latency-poc

sogabe avatar Aug 01 '18 07:08 sogabe