Skip to content

Commit

Permalink
fix mimir writes resources disk usage dashboard's graphs (#502)
Browse files Browse the repository at this point in the history
* fix mimir writes resources disk usage dashboard's graphs

* move dashboard to adequate directory

* Reformat file

* merge with main

---------

Co-authored-by: QuentinBisson <[email protected]>
  • Loading branch information
QuantumEnigmaa and QuentinBisson authored Apr 16, 2024
1 parent e1525ec commit 0034d01
Show file tree
Hide file tree
Showing 2 changed files with 8 additions and 4 deletions.
4 changes: 4 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,10 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0

## [Unreleased]

### Fixed

- Fix `Mimir / writes resources` disk usage related graphs.

## [3.11.1] - 2024-04-16

### Changed
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -964,7 +964,7 @@
"span": 4,
"targets": [
{
"expr": "sum by(instance, pod, device) (\n rate(\n node_disk_written_bytes_total[$__rate_interval]\n )\n)\n+\nignoring(pod) group_right() (\n label_replace(\n count by(\n instance,\n pod,\n device\n )\n (\n container_fs_writes_bytes_total{\n cluster=~\"$cluster\", namespace=~\"$namespace\",\n container=~\"ingester\",\n device!~\".*sda.*\"\n }\n ),\n \"device\",\n \"$1\",\n \"device\",\n \"/dev/(.*)\"\n ) * 0\n)\n\n",
"expr": "sum by(node, device) (\n rate(\n node_disk_written_bytes_total[$__rate_interval]\n )\n)\n+\non(node, device) group_right() (\n label_replace(\n count by(\n node,\n device,\n pod,\n )\n (\n container_fs_writes_bytes_total{\n cluster_id=~\"$cluster\", namespace=~\"$namespace\",\n container=~\"ingester\",\n device!~\".*sda.*\"\n }\n ),\n \"device\",\n \"$1\",\n \"device\",\n \"/dev/(.*)\"\n ) * 0\n)\n\n",
"format": "time_series",
"legendFormat": "{{pod}} - {{device}}",
"legendLink": null
Expand Down Expand Up @@ -1012,7 +1012,7 @@
"span": 4,
"targets": [
{
"expr": "sum by(instance, pod, device) (\n rate(\n node_disk_read_bytes_total[$__rate_interval]\n )\n) + ignoring(pod) group_right() (\n label_replace(\n count by(\n instance,\n pod,\n device\n )\n (\n container_fs_writes_bytes_total{\n cluster=~\"$cluster\", namespace=~\"$namespace\",\n container=~\"ingester\",\n device!~\".*sda.*\"\n }\n ),\n \"device\",\n \"$1\",\n \"device\",\n \"/dev/(.*)\"\n ) * 0\n)\n\n",
"expr": "sum by(node, device) (\n rate(\n node_disk_read_bytes_total[$__rate_interval]\n )\n) + on(node, device) group_right() (\n label_replace(\n count by(\n node,\n pod,\n device\n )\n (\n container_fs_writes_bytes_total{\n cluster_id=~\"$cluster\", namespace=~\"$namespace\",\n container=~\"ingester\",\n device!~\".*sda.*\"\n }\n ),\n \"device\",\n \"$1\",\n \"device\",\n \"/dev/(.*)\"\n ) * 0\n)\n\n",
"format": "time_series",
"legendFormat": "{{pod}} - {{device}}",
"legendLink": null
Expand Down Expand Up @@ -1063,7 +1063,7 @@
"span": 4,
"targets": [
{
"expr": "max by(persistentvolumeclaim) (\n kubelet_volume_stats_used_bytes{cluster=~\"$cluster\", namespace=~\"$namespace\"} /\n kubelet_volume_stats_capacity_bytes{cluster=~\"$cluster\", namespace=~\"$namespace\"}\n)\nand\ncount by(persistentvolumeclaim) (\n kube_persistentvolumeclaim_labels{\n cluster=~\"$cluster\", namespace=~\"$namespace\",\n label_name=~\"(ingester).*\"\n }\n)\n",
"expr": "max by(persistentvolumeclaim) (\n kubelet_volume_stats_used_bytes{cluster_id=~\"$cluster\", namespace=~\"$namespace\", persistentvolumeclaim=~\"storage-mimir-ingester-.*\"} /\n kubelet_volume_stats_capacity_bytes{cluster_id=~\"$cluster\", namespace=~\"$namespace\", persistentvolumeclaim=~\"storage-mimir-ingester-.*\"}\n)\n",
"format": "time_series",
"legendFormat": "{{persistentvolumeclaim}}",
"legendLink": null
Expand Down Expand Up @@ -1115,7 +1115,7 @@
"multi": false,
"name": "cluster",
"options": [],
"query": "label_values(cortex_build_info, cluster)",
"query": "label_values(mimir_build_info, cluster_id)",
"refresh": 1,
"regex": "",
"sort": 1,
Expand Down

0 comments on commit 0034d01

Please sign in to comment.