-
Notifications
You must be signed in to change notification settings - Fork 200
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
fix grafana dashboard and clarify dashboard usage more clearly; Fix t…
…he image display problem in document and document format Signed-off-by: jiangsanyin <[email protected]>
- Loading branch information
1 parent
b0873aa
commit 1fe9420
Showing
4 changed files
with
2,196 additions
and
1,208 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,54 +1,206 @@ | ||
## Grafana Dashboard | ||
# Abstract | ||
|
||
- You can load this dashboard json file [gpu-dashboard.json](./gpu-dashboard.json) | ||
For the sake of simplicity, this article provides only one possible way to ultimately use prometheus to capture monitoring metrics as a data source and grafana to present monitoring information. | ||
|
||
- This dashboard also includes some NVIDIA DCGM metrics: | ||
Many users feedback from creating issues that they do not know how to install and configure related components, resulting in failure to use related dashboard normally. The installation and configuration steps are described as follows, Hope you use it smoothly! Any feedback is welcome. | ||
|
||
[dcgm-exporter](https://github.com/NVIDIA/dcgm-exporter) deploy:`kubectl create -f https://raw.githubusercontent.com/NVIDIA/dcgm-exporter/master/dcgm-exporter.yaml` | ||
This article assumes that Kubernetes cluster and HAMi has been deployed successfully. The following components are installed in a kubernetes cluster. The components or software versions are as follows: | ||
|
||
- use this prometheus custom metric configure: | ||
| components or software name | version | remark | | ||
| --------------------------- | ------------------- | ---------------- | | ||
| kubernetes cluster | v1.23.10 | in AMD64 servers | | ||
| kube-prometheus stack | branch release-0.11 | | | ||
| dcgm-exporter | tag 3.2.5-3.1.7 | | | ||
|
||
```yaml | ||
- job_name: 'kubernetes-vgpu-exporter' | ||
kubernetes_sd_configs: | ||
- role: endpoints | ||
relabel_configs: | ||
- source_labels: [__meta_kubernetes_endpoints_name] | ||
regex: vgpu-device-plugin-monitor | ||
# Deploy and configure kube-prometheus stack | ||
|
||
## Deploy kube-prometheus stack | ||
|
||
**Note:**See the version compatibility matrix for kubernetes and kube-prometheus stack in:https://github.com/prometheus-operator/kube-prometheus?tab=readme-ov-file#compatibility | ||
|
||
```shell | ||
#Clone kube-prometheus code repository(using release-0.11 here) | ||
git clone -b release-0.11 https://github.com/prometheus-operator/kube-prometheus.git | ||
cd kube-prometheus | ||
|
||
#Change type of grafana service into NodePort by Adding "type: NodePort" under spec section | ||
vi manifests/grafana-service.yaml | ||
... | ||
spec: | ||
type: NodePort | ||
... | ||
#Similarly, change the prometheus and alertmanager service types to NodePort. Their configuration files are prometheus-service.yaml and alertmanager-service.yaml, respectively, in the manifests directory | ||
|
||
#Do deployment | ||
kubectl create -f manifests/setup/ | ||
kubectl create -f manifests/. | ||
|
||
#All resouce objects are been created under the monitoring namespace, you can check them and their status by run the following command | ||
kubectl -n monitoring get all | ||
``` | ||
|
||
```shell | ||
#Once all resouce objects under the monitoring namespace are in right status, you can obtain svc information for grafana, prometheus, and alertmanager in the following way | ||
root@controller01:~/kube-prometheus# kubectl -n monitoring get svc | egrep "NAME|grafana|prometheus-k8s|alertmanager-main" | ||
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE | ||
alertmanager-main NodePort 10.233.5.65 <none> 9093:30093/TCP,8080:30401/TCP 19h | ||
grafana NodePort 10.233.56.112 <none> 3000:30300/TCP 19h | ||
prometheus-k8s NodePort 10.233.38.113 <none> 9090:30090/TCP,8080:31273/TCP 19h | ||
``` | ||
|
||
If ip address of controller node is 10.0.0.21, then grafana, prometheus, and alertmanager can be accessed using the following urls: http://10.0.0.21:30300 , http://10.0.0.21:30090 , and http://10.0.0.21:30093 , and the default user name and password for accessing grafana are admin | ||
|
||
## Configure grafana | ||
|
||
### Create Datasource ALL | ||
|
||
Go to the "Configuration" -> "Data soutces" page in grafana and create a datasource named "ALL", and keep the value of HTTP.URL be same with the counterpart in default "prometheus" datasource. | ||
|
||
### Import dashboard | ||
|
||
Go to the "Configuration" -> "Data soutces" page in grafana and import the dashboard from https://grafana.com/grafana/dashboards/22043-hami-vgpu-metrics-dashboard/ , and a dashboard page named "hami-vgpu-metrics-dashboard" will be created. 22043-hami-vgpu-metrics-dashboard is valid in grafana8.5.5 and grafana9.1.0, and it's grealty possible that this dashboard is vaild in grafana version later than 9.1.0. Now data of some panels in this dashboard page are missing, which requires you read the rest of the document. | ||
|
||
For versions earlier than grafana8.5.5, such as grafana7.5.17, please refer to:https://grafana.com/grafana/dashboards/21833-hami-vgpu-dashboard/ | ||
|
||
# Deploy dcgm-exporter | ||
|
||
```shell | ||
#Clone dcgm-exporter code repository(the compatibility matrix for dcgm-export and kubernetes is not been founded in its official website, using "tag 3.2.5-3.1.7" here) | ||
git clone -b 3.2.5-3.1.7 https://github.com/NVIDIA/dcgm-exporter.git | ||
cd dcgm-exporter | ||
|
||
#Install dcgm-exporter under monitoring namespace with helm | ||
helm install dcgm-exporter deployment/ -n monitoring | ||
|
||
#Check installation results | ||
root@controller01:~/dcgm-exporter# helm list -n monitoring | ||
NAME NAMESPACE REVISION UPDATED STATUS CHART APP VERSION | ||
dcgm-exporter monitoring 1 2024-10-02 16:32:35.691073696 +0800 CST deployed dcgm-exporter-3.1.7 3.1.7 | ||
``` | ||
|
||
# Create ServiceMonitor | ||
|
||
```shell | ||
#Create the file hami-device-plugin-svc-monitor.yaml | ||
root@controller01:~# touch hami-device-plugin-svc-monitor.yaml | ||
#The content of the file hami-device-plugin-svc-monitor.yaml | ||
root@controller01:~# cat hami-device-plugin-svc-monitor.yaml | ||
apiVersion: monitoring.coreos.com/v1 | ||
kind: ServiceMonitor | ||
metadata: | ||
name: hami-device-plugin-svc-monitor | ||
namespace: kube-system | ||
spec: | ||
selector: | ||
matchLabels: | ||
app.kubernetes.io/component: hami-device-plugin | ||
namespaceSelector: | ||
matchNames: | ||
- "kube-system" | ||
endpoints: | ||
- path: /metrics | ||
port: monitorport | ||
interval: "15s" | ||
honorLabels: false | ||
relabelings: | ||
- sourceLabels: [__meta_kubernetes_endpoints_name] | ||
regex: hami-.* | ||
replacement: $1 | ||
action: keep | ||
- source_labels: [__meta_kubernetes_pod_node_name] | ||
- sourceLabels: [__meta_kubernetes_pod_node_name] | ||
regex: (.*) | ||
target_label: node_name | ||
targetLabel: node_name | ||
replacement: ${1} | ||
action: replace | ||
- source_labels: [__meta_kubernetes_pod_host_ip] | ||
- sourceLabels: [__meta_kubernetes_pod_host_ip] | ||
regex: (.*) | ||
target_label: ip | ||
targetLabel: ip | ||
replacement: $1 | ||
action: replace | ||
- job_name: 'kubernetes-dcgm-exporter' | ||
kubernetes_sd_configs: | ||
- role: endpoints | ||
relabel_configs: | ||
- source_labels: [__meta_kubernetes_endpoints_name] | ||
regex: dcgm-exporter | ||
|
||
#apply the file hami-device-plugin-svc-monitor.yaml | ||
root@controller01:~# kubectl apply -f hami-device-plugin-svc-monitor.yaml | ||
``` | ||
|
||
```shell | ||
#Create the file hami-scheduler-svc-monitor.yaml | ||
root@controller01:~# touch hami-device-plugin-svc-monitor.yaml | ||
#The content of the file hami-scheduler-svc-monitor.yaml | ||
root@controller01:~# cat hami-scheduler-svc-monitor.yaml | ||
apiVersion: monitoring.coreos.com/v1 | ||
kind: ServiceMonitor | ||
metadata: | ||
name: hami-scheduler-svc-monitor | ||
namespace: kube-system | ||
spec: | ||
selector: | ||
matchLabels: | ||
app.kubernetes.io/component: hami-scheduler | ||
namespaceSelector: | ||
matchNames: | ||
- "kube-system" | ||
endpoints: | ||
- path: /metrics | ||
port: monitor | ||
interval: "15s" | ||
honorLabels: false | ||
relabelings: | ||
- sourceLabels: [__meta_kubernetes_endpoints_name] | ||
regex: hami-.* | ||
replacement: $1 | ||
action: keep | ||
- source_labels: [__meta_kubernetes_pod_node_name] | ||
- sourceLabels: [__meta_kubernetes_pod_node_name] | ||
regex: (.*) | ||
target_label: node_name | ||
targetLabel: node_name | ||
replacement: ${1} | ||
action: replace | ||
- source_labels: [__meta_kubernetes_pod_host_ip] | ||
- sourceLabels: [__meta_kubernetes_pod_host_ip] | ||
regex: (.*) | ||
target_label: ip | ||
targetLabel: ip | ||
replacement: $1 | ||
action: replace | ||
|
||
#apply the file hami-scheduler-svc-monitor.yaml | ||
root@controller01:~# kubectl apply -f hami-scheduler-svc-monitor.yaml | ||
``` | ||
|
||
```shell | ||
#Check the servicemonitors | ||
root@controller01:~# kubectl -n kube-system get servicemonitor | ||
NAME AGE | ||
hami-device-plugin-svc-monitor 28h | ||
hami-scheduler-svc-monitor 28h | ||
``` | ||
|
||
- reload promethues: | ||
# Confirm the final monitoring effect | ||
|
||
```bash | ||
curl -XPOST http://{promethuesServer}:{port}/-/reload | ||
```shell | ||
#Create the file gpu-pod.yaml | ||
root@controller01:~# touch gpu-pod.yaml | ||
root@controller01:~# cat gpu-pod.yaml | ||
apiVersion: v1 | ||
kind: Pod | ||
metadata: | ||
name: gpu-pod-01 | ||
spec: | ||
restartPolicy: Never | ||
containers: | ||
- name: cuda-container | ||
image: nvcr.io/nvidia/k8s/cuda-sample:vectoradd-cuda10.2 | ||
resources: | ||
limits: | ||
nvidia.com/vgpu: 2 # requesting 2 vGPUs | ||
nvidia.com/gpumem: 3000 # Each vGPU contains 3000m device memory (Optional,Integer) | ||
nvidia.com/gpucores: 10 # Each vGPU uses 30% of the entire GPU (Optional,Integer) | ||
|
||
#apply the file gpu-pod.yaml | ||
root@controller01:~# kubectl apply -f gpu-pod.yaml | ||
root@controller01:~# kubectl get pods -o wide | ||
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES | ||
gpu-pod-01 0/1 Completed 0 52s 10.233.81.70 controller01 <none> <none> | ||
``` | ||
|
||
You can see the monitoring details in the dashboard. The contents are as follows: | ||
|
||
![image-20241003215400685](../imgs/hami-vgpu-metrics-dashboard.png) | ||
|
Oops, something went wrong.