This command probes the selected GPU's device status every second:
$ nvidia-smi dmon -s pucvmet -i -0
The following screenshot shows the result of the previous command. The device we are monitoring states that it has a GPU device status of 0:
The collected information can be specified with the -s option, as follows:
- p: Power usage and temperature
- u: Utilization
- c: Proc and mem clocks
- v: Power and thermal violations
- m: FB and Bar1 memory
- e: ECC errors and PCIe replay errors
- t: PCIe Rx and Tx throughput