Epstein Files

EFTA00282798.pdf

dataset_9 pdf 1.7 MB Feb 3, 2026 26 pages
nvidia—smi(1) NVIDIA nvidia—smi(l) NAME nvidia—smi — NVIDIA System Management Interface program SYNOPSIS nvidia-smi rOPTIONI [OFTION2 [ARG2]] DESCRIPTION nvidia-smi (also NVSMI) provides monitoring and management capabilities for each of NVIDIA's Tesla, Quadro, GRID and GeForce devices from Fermi and higher architecture families. GeForce Titan series devices are supported for most functions with very limited information provided for the remainder of the Geforce brand. NVSMI is a cross platform tool that supports all standard NVIDIA driver-supported Linux distros, as well as 64bit versions of Windows starting with Windows Server 2008 R2. Metrics can be con- sumed directly by users via stdout, or provided by file via CSV and XML formats for scripting purposes. Note that much of the functionality of NVSMI is provided by the underlying NVML C-based library. See the NVIDIA developer website link below for more information about NVML. NVML-based python bind- ings are also available. The output of NVSMI is not guaranteed to be backwards compatible. However, both NVML and the Python bindings are backwards compatible, and should be the first choice when writing any tools that must be maintained across NVIDIA driver releases. NVML SDK: http://developernvidia.conthavidia-management-library•mml/ Python bindings: http:/lpypipython.org/pypihrvidia-ml-py/ OPTIONS GENERAL OPTIONS —h, —help Print usage information and exit. SUMMARY OPTIONS —L, —list—gpus List each of the NVIDIA GPUs in the system, along with their UUlDs. QUERY OPTIONS —q, —query Display GPU or Unit info. Displayed info includes all data listed in the (GPU ATTRIBUTES) or (UNIT ATTRIBUTES) sections of this document. Some devices and/or environments don't support all possible information. Any unsupported data is indicated by a "N/A" in the output. By default information for all available GPUs or Units is displayed. Use the —i option to restrict the output to a single GPU or Unit. [plus optional] —u, —unit Display Unit data instead of GPU data. Unit data is only available for NVIDIA S—class Tesla enclosures. - —id=ID Display data for a single specified GPU or Unit. The specified id may be the GPU/Unit's 0—based index in the natural enumeration returned by the driver, the GPU's board serial number, the GPU's QUID, or the GPU's PCI bus ID (as domain:bus:device.function in hex). It is recommended that users desiring consis- tency use either UUID or PCI bus ID, since device enumeration ordering is not guaranteed to be consistent between reboots and board serial number might be shared between multiple GPUs on the same board. nvidia—smi 352.26 2015/7/7 EFTA00282798 nvidia—smi(1) NVIDIA nvidia—smi(1) —f FILE, —filename=FILE Redirect query output to the specified file in place of the default stdout. The specified file will be overwrit- ten. —x, --xml—format Produce XML output in place of the default human—readable format. Both GPU and Unit query outputs conform to corresponding DTDs. These are available via the —dtd flag. - -dtd Use with —x. Embed the DTD in the XML output. - -debug=FILE Produces an encrypted debug log for use in submission of bugs back to NVIDIA. - d TYPE, -display=TYPE Display only selected information: MEMORY, UTILIZATION, ECC, TEMPERATURE, POWER, CLOCK, COMPUTE, PIDS, PERFORMANCE, SUPPORTEDSLOCKS, PAGE_RETIREMENT, ACCOUNTING Flags can be combined with comma e.g. "MEMORY,ECC". Sampling data with max, min and avg is also returned for POWER, UTILIZATION and CLOCK display types. Doesn't work with -u/--unit or -x/--xml-format flags. - 1 SEC, —loop=SEC Continuously report query data at the specified interval, rather than the default of just once. The applica- tion will sleep in—between queries. Note that on Linux ECC error or XID enor events will print out during the sleep period if the -x flag was not specified. Pressing Ctrl+C at any time will abort the loop, which will otherwise run indefinitely. If no argument is specified for the —I form a default interval of 5 seconds is used. SELECTIVE QUERY OPTIONS Allows the caller to pass an explicit list of properties to query. [one of) --query—gpu= Information about GPU. Pass comma separated list of properties you want to query. e.g. --query—gpu=pci.bus_id,persistence_mode. Call —help—query—gpu for more info. --query—supported—clocks= List of supported clocks. Call --help—query—supported—clocks for more info. - -query—compute—apps= List of currently active compute processes. Call —help—query—compute—apps for more info. --query—accounted—apps= List of accounted compute processes. Call —help—query—accounted—apps for more info. - -query—retired—page= List of GPU device memory pages that have been retired. Call --help—query—retired—pages for more info. nvidia—smi 352.26 2015/717 EFTA00282799 nvidia-smi( I ) NVIDIA nvidia-smi(1) [mandatory] —-format= Comma separated list of format options: csv - comma separated values (MANDATORY) noheader - skip first line with column headers nounits - don't print units for numerical values [plus any of] - —id=ID Display data for a single specified GPU. The specified id may be the GPU's 0—based index in the natural enumeration returned by the driver, the GPU's board serial number, the GPU's UUID, or the CPU's PCI bus ID (as domain:bus:device.function in hex). It is recommended that users desiring consistency use either UUID or PCI bus ID, since device enumeration ordering is not guaranteed to be consistent between reboots and board serial number might be shared between multiple GPUs on the same board. —f FILE, —filename=FILE Redirect query output to the specified file in place of the default stdout. The specified file will be overwrit- ten. —I SEC, —loop=SEC Continuously report query data at the specified interval, rather than the default of just once. The applica- tion will sleep in—between queries. Note that on Linux ECC error or XID error events will print out during the sleep period if the -A- flag was not specified. Pressing Ctrl+C at any time will abort the loop, which will otherwise run indefinitely. If no argument is specified for the —I form a default interval of 5 seconds is used. —Ims ms, --loop—ms=ms Same as —I.--loop but in milliseconds. DEVICE MODIFICATION OPTIONS [any one of] —pm, --persistence—mode=MODE Set the persistence mode for the target GPUs. See the (GPU ATTRIBUTES) section for a description of persistence mode. Requires root. Will impact all GPUs unless a single GPU is specified using the —i argu- ment. The effect of this operation is immediate. However, it does not persist across reboots. After each reboot persistence mode will default to "Disabled". Available on Linux only. —e, —ecc—config=CONFIG Set the ECC mode for the target GPUs. See the (GPU ATTRIBUTES) section for a description of ECC mode. Requires root. Will impact all GPUs unless a single GPU is specified using the —i argument. This setting takes effect after the next reboot and is persistent. —p, —reset—ece—errors=TYPE Reset the ECC error counters for the target GPUs. See the (GPU ATTRIBUTES) section for a description of ECC error counter types. Available arguments are 01VOLATILE or IIAGGREGATE. Requires root. Will impact all GPUs unless a single GPU is specified using the —i argument. The effect of this operation is immediate. nvidia—smi 352.26 2015/7/7 3 EFTA00282800 nvidia—smi( I ) NVIDIA nvidia—smi(1) —c, —compute—modMODE Set the compute mode for the target GPUs. See the (GPU ATTRIBUTES) section for a description of com- pute mode. Requires root. Will impact all GPUs unless a single GPU is specified using the —i argument. The effect of this operation is immediate. However, it does not persist across reboots. After each reboot compute mode will reset to "DEFAULT". —dm TYPE, —driver —model=TYPE —Mm TYPE, --force—driver—model=TYPE Enable or disable TCC driver model. For Windows only. Requires administrator privileges. —dm will fail if a display is attached, but —fdnt will force the driver model to change. Will impact all GPUs unless a sin- gle GPU is specified using the —i argument. A reboot is required for the change to take place. See Driver Model for more information on Windows driver models. —gom=MODE Set GPU Operation Mode: 0/ALL_ON, I/COMPUTE, 2/LOW_DP Supported on GK I 10 M-class and X- class Testa products from the Kepler family. Not supported on Quadro and Tesla C-class products. LOW_DP and ALL_ON are the only modes supported on GeForce Titan devices. Requires administrator privileges. See GPU Operation Mode for more information about GOM. GOM changes take effect after reboot. The reboot requirement might be removed in the future. Compute only GOMs don't support WDDM (Windows Display Driver Model) —r, --gpu—reset Trigger a reset of the GPU. Can be used to clear GPU HW and SW state in situations that would otherwise require a machine reboot. Typically useful if a double bit ECC error has occurred. Requires —i switch to target specific device. Requires root. There can't be any applications using this particular device (e.g. CUDA application, graphics application like X server, monitoring application like other instance of nvidia- smi). There also can't be any compute applications running on any other GPU in the system. Only on sup- ported devices from Fermi and Kepler family running on Linux. GPU reset is not guaranteed to work in all cases. It is not recommended for production environments at this time. In some situations there may be HW components on the board that fail to revert back to an initial state following the reset request. This is more likely to be seen on Fermi-generation products vs. Kepler, and more likely to be seen if the reset is being performed on a hung GPU. Following a reset, it is recommended that the health of the GPU be verified before further use. The nvidia- healthmon tool is a good choice for this test. If the GPU is not healthy a complete reset should be insti- gated by power cycling the node. Visit http://developernvidia.com/gpu•deployment•kit to download the GDK and nvidia-hcalthmon. —ac, --applications—clock lEM_CLOCIC,GRAPHICS_CLOCK Specifies maximum <memorygraphics> clocks as a pair (e.g. 2000,800) that defines GPU's speed while running applications on a GPU. For Testa devices from the Kepler+ family and Maxwell-based GeForce Titan. Requires root unless restrictions am relaxed with the —acp command.. —rac, —reset—applications—clocks Resets the applications clocks to the default value. For Tesla devices from the Kepler+ family and Max- well-based GeForce Titan. Requires root unless restrictions are relaxed with the —acp command. —acp, —applications—clocks—permission=MODE Toggle whether applications clocks can be changed by all users or only by root. Available arguments are 0ILINRESTRICTED. I 'RESTRICTED. For Tesla devices from the Kepler+ family and Maxwell-based nvidia—smi 352.26 2015/7/7 4 EFTA00282801 nvidia—smi(1) NVIDIA nvidia—smi(1) GeForce Titan. Requires root. —pl, —power—limit=POWER_LIMIT Specifies maximum power limit in watts. Accepts integer and floating point numbers. Only on supported devices from Kepler family. Requires administrator privileges. Value needs to be between Min and Max Power Limit as reported by nvidia-smi. —am, --accounting—mode=MODE Enables or disables GPU Accounting. With GPU Accounting one can keep track of usage of resources throughout lifespan of a single process. Only on supported devices from Kepler family. Requires adminis- trator privileges. Available arguments are OIDISABLED or I/ENABLED. —eaa, —clear—accounted—apps Clears all processes accounted so far. Only on supported devices from Kepler family. Requires administra- tor privileges. —auto—boost—default=MODE Set the default auto boost policy to 0/DISABLED or UENABLED, enforcing the change only after the last boost client has exited. Only on certain Testa devices from the Kepler+ family and Maxwell-based GeForce devices. Requires root. —auto—boost—default—foree=NIODE Set the default auto boost policy to 0/DISABLED or I/ENABLED, enforcing the change immediately. Only on certain Testa devices from the Kepler+ family and Maxwell-based GeForce devices. Requires root. --auto—boost—permission=MODE Allow non-admin/root control over auto boost mode. Available arguments are OIUNRESTRICTED, I RESTRICTED. Only on certain Testa devices from the Kepler+ family and Maxwell-based GeForce devices. Requires root. [plus optional] - --id=ID Modify a single specified GPU. The specified id may be the GPU/Unit's 0—based index in the natural enu- meration returned by the driver, the CPU's board serial number, the CPU's UUID, or the GPU's PCI bus ID (as domain:bus:device.function in hex). It is recommended that users desiring consistency use either UUID or PCI bus ID, since device enumeration ordering is not guaranteed to be consistent between reboots and board serial number might be shared between multiple GPUs on the same board. UNIT MODIFICATION OPTIONS —t, —toggle—led=STATE Set the LED indicator state on the front and back of the unit to the specified color. See the (UNIT ATTRIBUTES) section for a description of the LED states. Allowed colors are OIGREEN and IIAMBER. Requires root. [plus optional] - --id=ID Modify a single specified Unit. The specified id is the Unit's 0-based index in the natural enumeration returned by the driver. nvidia—smi 352.26 2015/7/7 5 EFTA00282802 nvidia—smi(1) NVIDIA nvidia—smi(1) SHOW DTD OPTIONS —dtd Display Device or Unit DTD. [plus optional] —f FILE, —filename=FILE Redirect query output to the specified file in place of the default stdout. The specified file will be overwrit- ten. —u, —unit Display Unit DTD instead of device DTD. stats Display statistics information about the GPU. Use "nvidia-smi stats -h" for more information. Linux only. topo Display topology information about the system. Use "nvidia•smi topo -h" for more information. Linux only. Shows all GPUs NVML is able to detect but CPU affinity information will only be shown for GPUs with Kepler or newer architectures. Note: GPU enumeration is the same as NVML. RETURN VALUE Return code reflects whether the operation succeeded or failed and what was the reason of failure. Return code 0 — Success Return code 2 — A supplied argument or flag is invalid Return code 3 — The requested operation is not available on target device Return code 4 — The current user does not have permission to access this device or perform this operation Return code 6 — A query to find an object was unsuccessful Return code 8 — A device's external power cables are not properly attached Return code 9 — NVIDIA driver is not loaded Return code 10 — NVIDIA Kernel detected an interrupt issue with a GPU Return code 12 — NVML Shared Library couldn't be found or loaded Return code 13 — Local version of NVML doesn't implement this function Return code l4 — infoROM is corrupted Return code 15 — The GPU has fallen off the bus or has otherwise become inaccessible Return code 255 — Other error or internal driver error occurred GPU ATTRIBUTES The following list describes all possible data returned by the —q device query option. Unless otherwise noted all numerical results are base 10 and unitless. runestamp The current system timestamp at the time nvidia—smi was invoked. Format is "Day—of—week Month Day HH:MM:SS Year". nvidia—smi 352.26 2015/7/7 6 EFTA00282803 nvidia—smi(1) NVIDIA nvidia—smi(l) Driver Version The version of the installed NVIDIA display driver. This is an alphanumeric string. Attached GPUs The number of NVIDIA GPUs in the system. Product Name The official product name of the GPU. This is an alphanumeric string. For all products. Display Mode A flag that indicates whether a physical display (e.g. monitor) is currently connected to any of the CPU's connectors. "Enabled" indicates an attached display. "Disabled" indicates otherwise. Display Active A flag that indicates whether a display is initialized on the GPU's (e.g. memory is allocated on the device for display). Display can be active even when no monitor is physically attached. "Enabled" indicates an active display. "Disabled" indicates otherwise. Persistence Mode A flag that indicates whether persistence mode is enabled for the GPU. Value is either "Enabled" or "Dis- abled". When persistence mode is enabled the NVIDIA driver remains loaded even when no active clients, such as XII or nvidia-smi, exist. This minimizes the driver load latency associated with running dependent apps, such as CUDA programs. For all CUDA-capable products. Linux only. Accounting Mode A flag that indicates whether accounting mode is enabled for the GPU Value is either When accounting is enabled statistics are calculated for each compute process running on the GPU. Statistics can be queried during the lifetime or after termination of the process. The execution time of process is reported as 0 while the process is in running state and updated to actual execution time after the process has terminated. See --help-query•accounted-apps for more info. Accounting Mode Buffer Size Returns the size of the circular buffer that holds list of processes that can be queried for accounting stats. This is the maximum number of processes that accounting information will be stored for before information about oldest processes will get overwritten by information about new processes. Driver Model On Windows, the TCC and WDDM driver models are supported. The driver model can be changed with the (—dm) or (—film) flags. The TCC driver model is optimized for compute applications. I.E. kernel launch times will be quicker with TCC. The WDDM driver model is designed for graphics applications and is not recommended for compute applications. Linux does not support multiple driver models, and will always have the value of "N/A". Current The driver model currently in use. Always "N/A" on Linux. Pending The driver model that will be used on the next reboot. Always "NIA" on Linux. Serial Number This number matches the serial number physically printed on each board. It is a globally unique immutable alphanumeric value. nvidia—smi 352.26 2015/7/7 7 EFTA00282804 nvidia—smi(1) NVIDIA nvidia—smi(l) GPU UUID This value is the globally unique immutable alphanumeric identifier of the GPU. It does not correspond to any physical label on the board. Minor Number The minor number for the device is such that the Nvidia device node file for each GPU will have the form /devinvidialminor number]. Available only on Linux platform. VBIOS Version The BIOS of the GPU board. MultiGPU Board Whether or not this GPU is part of a multiGPU board. Board ID The unique board ID assigned by the driver. If two or more GPUs have the same board ID and the above "MultiGPU" field is true then the GPUs are on the same board. Inforom Version Version numbers for each object in the GPU board's inforom storage. The inforom is a small, persistent store of configuration and state data for the GPU. All inforom version fields are numerical. It can be useful to know these version numbers because some GPU features are only available with inforoms of a certain version or higher. If any of the fields below return Unknown Error additional Inforom verification check is performed and appropriate warning message is displayed. Image Version Global version of the infoROM image. Image version just like VBIOS version uniquely describes the exact version of the infoROM flashed on the board in contrast to infoROM object version which is only an indicator of supported features. OEM Object Version for the OEM configuration data. ECC Object Version for the ECC recording data. Power Object Version for the power management data. GPU Operation Mode GOM allows to reduce power usage and optimize GPU throughput by disabling GPU features. Each GOM is designed to meet specific user needs. In "All On" mode everything is enabled and running at full speed. The "Compute" mode is designed for running only compute tasks. Graphics operations are not allowed. The "Low Double Precision" mode is designed for running graphics applications that don't require high bandwidth double precision. nvidia—smi 352.26 20151117 8 EFTA00282805 nvidia—smi(1) NVIDIA nvidia—smi(l) GOM can be changed with the (--gom) flag. Supported on GK110 M•class and X•class Tesla products from the Kepler family. Not supported on Quadro and Tesla C•class products. Low Double Precision and All On modes are the only modes available for supported GeForce Titan products. Current The GOM currently in use. Pending The GOM that will be used on the next reboot. PCI Basic PCI info for the device. Some of this information may change whenever cards are added/removed/moved in a system. For all products. Bus PCI bus number, in hex Device PCI device number, in hex Domain PCI domain number, in hex Device Id PCI vendor device id, in hex Sub System Id PCI Sub System id, in hex Bus Id PCI bus id as "domain:bus:device.function", in hex GPU Link information The PCIe link generation and bus width Current The current link generation and width. These may be reduced when the GPU is not in use. Maximum The maximum link generation and width possible with this GPU and system configura- tion. For example, if the GPU supports a higher PCIe generation than the system sup- ports then this reports the system PCIe generation. Bridge Chip Information related to Bridge Chip on the device. The bridge chip firmware is only present on certain boards and may display "N/A" for some newer multiGPUs boards. Type The type of bridge chip. Reported as N/A if doesn't exist Firmware Version The firmware version of the bridge chip. Reported as N/A if doesn't exist. Replay counter This is the internal counter that records various errors on the PM bus. nvidia—smi 352.26 2015/7/7 9 EFTA00282806 nvidia—smi(1) NVIDIA nvidia—smi(l) 'Dc Throughput The CPU-centric transmission throughput across the PCIe bus in MB/s over the past 20ms. Only supported on Maxwell architectures and newer. Rx Throughput The CPU-centric receive throughput across the PCIe bus in MB/s over the past 20ms. Only supported on Maxwell architectures and newer. Fan Speed The fan speed value is the percent of maximum speed that the device's fan is currently intended to run at. It ranges from 0 to 100%. Note: The reported speed is the intended fan speed. If the fan is physically blocked and unable to spin, this output will not match the actual fan speed. Many parts do not report fan speeds because they rely on cooling via fans in the surrounding enclosure. For all discrete products with dedicated fans. Performance State The current performance state for the GPU. States range from PO (maximum performance) to P12 (mini- mum performance). Clocks Throttle Reasons Retrieves information about factors that are reducing the frequency of clocks. Only on supported Tesla devices from Kepler family. If all throttle reasons are returned as "Not Active" it means that clocks are running as high as possible. Idle Nothing is running on the GPU and the clocks are dropping to Idle state. This limiter may be removed in a later release. Application Clocks Setting GPU clocks are limited by applications clocks setting. E.g. can be changed using nvidia—smi --applications—clocks= SW Power Cap SW Power Scaling algorithm is reducing the clocks below requested clocks because the GPU is consuming too much power. E.g. SW power cap limit can be changed with nvidia—smi --power—limit= HW Slowdown 14W Slowdown (reducing the core clocks by a factor of 2 or more) is engaged. This is an indicator of: • Temperature being too high • External Power Brake Assertion is triggered (e.g. by the system power supply) • Power draw is too high and Fast Trigger protection is reducing the clocks Unknown Some other unspecified factor is reducing the clocks. FB Memory Usage On-board frame buffer memory information. Reported total memory is affected by ECC state. If ECC is enabled the total available memory is decreased by several percent, due to the requisite parity bits. The driver may also reserve a small amount of memory for internal use, even without active work on the GPU. For all products. nvidia—smi 352.26 2015/717 10 EFTA00282807 nvidia—smi(1) NVIDIA nvidia—smi(1) Total Total size of FB memory. Used Used size of FB memory. Free Available size of FB memory. BARI Memory Usage BARI is used to map the FB (device memory) so that it can be directly accessed by the CPU or by 3rd party devices (peer-to-peer on the PCIe bus). Total Total size of BARI memory. Used Used size of BAR I memory. Free Available size of BARI memory. Compute Mode The compute mode flag indicates whether individual or multiple compute applications may run on the GPU. "Default" means multiple contexts are allowed per device. "Exclusive Thread" means only one context is allowed per device, usable from one thread at a time (depre- cated in r352). "Exclusive Process" means only one context is allowed per device, usable from multiple threads at a time. "Prohibited" means no contexts are allowed per device (no compute apps). "EXCLUSIVE_PROCESS" was added in CUDA 4.0. Prior CUDA releases supported only one exclusive mode, which is equivalent to "EXCLUSIVE_THREAD" in CUDA 4.0 and beyond. For all CUDA-capable products. Utilization Utilization rates report how busy each GPU is over time, and can be used to determine how much an appli- cation is using the GPUs in the system. Note: During driver initialization when ECC is enabled one can see high GPU and Memory Utilization readings. This is caused by ECC Memory Scrubbing mechanism that is performed during driver initializa- tion. GPU Percent of time over the past sample period during which one or more kernels was exe- cuting on the GPU. The sample period may be between I second and 1/6 second depending on the product. Memory Percent of time over the past sample period during which global (device) memory was being read or written. The sample period may be between I second and 116 second depending on the product. nvidia—smi 352.26 201517/7 II EFTA00282808 nvidia—smi(1) NVIDIA nvidia—smi(1) Encoder Percent of time over the past sample period during which the GPU's video encoder was being used. The sampling rate is variable and can be obtained directly via the nvmlDe- viceGetEncoderUtilization() API Decoder Percent of time over the past sample period during which the GPU's video decoder was being used. The sampling rate is variable and can be obtained directly via the nvinIDe- viceGetDecoderUtilization0 API Ecc Mode A flag that indicates whether ECC support is enabled. May be either "Enabled" or "Disabled". Changes to ECC mode require a reboot. Requires Inforom ECC object version 1.0 or higher. Current The ECC mode that the GPU is currently operating under. Pending The ECC mode that the GPU will operate under after the next reboot. ECC Errors NVIDIA GPUs can provide error counts for various types of ECC errors. Some ECC errors are either sin• gle or double bit, where single bit errors are corrected and double bit errors are uncorrectable. Texture memory errors may be correctable via resend or uncorrectable if the resend fails. These errors are available across two timescales (volatile and aggregate). Single bit ECC errors are automatically corrected by the HW and do not result in data corruption. Double bit errors are detected but not corrected. Please see the ECC documents on the web for information on compute application behavior when double bit errors occur. Volatile error counters track the number of errors detected since the last driver load. Aggregate error counts persist indefinitely and thus act as a lifetime counter. A note about volatile counts: On Windows this is once per boot. On Linux this can be more frequent. On Linux the driver unloads when no active clients exist. Hence, if persistence mode is enabled or there is always a driver client active (e.g. X II), then Linux also sees per-boot behavior. If not, volatile counts are reset each time a compute app is run. Tesla and Quadro products from the Fermi and Kepler family can display total ECC error counts, as well as a breakdown of errors based on location on the chip. The locations are described below. Location—based data for aggregate error counts requires Inforom ECC object version 2.0. All other ECC counts require ECC object version 1.0. Device Memory Errors detected in global device memory. Register File Errors detected in register file memory. LI Cache Errors detected in the LI cache. L2 Cache Errors detected in the L2 cache. Texture Memory Parity errors detected in texture memory. Total Total errors detected across entire chip. Sum of Device Memory. Register File. LI Cache, 12 Cache and Texture Memory. nvidia—smi 352.26 2015/7/7 12 EFTA00282809 nvidia—smi(1) NVIDIA nvidia—smi(1) Page Retirement NVIDIA CPUs can retire pages of GPU device memory when they become unreliable. This can happen when multiple single bit ECC errors occur for the same page, or on a double bit ECC error. When a page is retired. the NVIDIA driver will hide it such that no driver, or application memory allocations can access it. Double Bit ECC The number of GPU device memory pages that have been retired due to a double bit ECC error. Single Bit ECC The number of GPU device memory pages that have been retired due to multiple single bit ECC errors. Pending Checks if any GPU device memory pages are pending retirement on the next reboot. Pages that are pending retirement can still be allocated, and may cause further reliability issues. Temperature Readings from temperature sensors on the board. All readings are in degrees C. Not all products support all reading types. In particular, products in module form factors that rely on case fans or passive cooling do not usually provide temperature readings. See below for restrictions. GPU Core GPU temperature. For all discrete and S-class products. Shutdown Temp The temperature at which a GPU will shutdown. Slowdown Temp The temperature at which a GPU will begin slowing itself down in order to cool. Power Readings Power readings help to shed light on the current power usage of the CPU, and the factors that affect that usage. When power management is enabled the GPU limits power draw under load to fit within a prede- fined power envelope by manipulating the current performance state. See below for limits of availability. Power State Power State is deprecated and has been renamed to Performance State in 2.285. To maintain XML compatibility, in XML format Performance State is listed in both places. Power Management A flag that indicates whether power management is enabled. Either "Supported" or "N/A". Requires Inforom PWR object version 3.0 or higher or Kepler device. Power Draw The last measured power draw for the entire board, in watts. Only available if power management is supported. This reading is accurate to within +/- 5 watts. Requires Inforom PWR object version 3.0 or higher or Kepler device. Power Limit The software power limit, in watts. Set by software such as nvidia-smi. Only available if power management is supported. Requires Inforom PWR object version 3.0 or higher or Kepler device. On Kepler devices Power Limit can be adjusted using —pl.—power—limit= switches. Enforced Power Limit The power management algorithm's power ceiling, in watts. Total board power draw is manipulated by the power management algorithm such that it stays under this value. nvidia—smi 352.26 2015/7/7 13 EFTA00282810 nvidia—smi(1) NVIDIA nvidia—smi(l) This limit is the minimum of various limits such as the software limit listed above. Only available if power management is supported. Requires a Kepler device. Default Power Limit The default power management algorithm's power ceiling, in watts. Power Limit will be set back to Default Power Limit after driver unload. Only on supported devices from Kepler family. Min Power Limit The minimum value in watts that power limit can be set to. Only on supported devices from Kepler family. Max Power Limit The maximum value in watts that power limit can be set to. Only on supported devices from Kepler family. Clocks Current frequency at which parts of the GPU are running. All readings are in MHz. Graphics Current frequency of graphics (shader) clock. SM Current frequency of SM (Streaming Multiprocessor) clock. Memory Current frequency of memory clock Applications Clocks User specified frequency at which applications will be running at. Can be changed with [—ac I —applica- tions—clocks] switches. Graphics User specified frequency of graphics (shader) clock. Memory User specified frequency of memory clock. Default Applications Clocks Default frequency at which applications will be running at. Application clocks can be changed with [—ac I --applications—clocks] switches. Application clocks can be set to default using [—rac I —reset—applica- tions—clocks] switches. Graphics Default frequency of applications graphics (shader) clock. Memory Default frequency of applications memory clock. Max Clocks Maximum frequency at which parts of the GPU are design to run. All readings are in MHz. On GPUs from Fermi family current PO clocks (reported in Clocks section) can differ from max clocks by few MHz. nvidia—smi 352.26 zoisnn 14 EFTA00282811 nvidia—smi(1) NVIDIA nvidia—smi(l) Graphics Maximum frequency of graphics (shader) clock. SM Maximum frequency of SM (Streaming Multiprocessor) clock. Memory Maximum frequency of memory clock. Clock Policy User-specified settings for automated clocking changes such as auto boost. Auto Boost Indicates whether auto boost mode is currently enabled for this GPU (On) or disabled for this GPU (Off). Shows (N/A) if boost is not supported. Auto boost allows dynamic GPU clocking based on power, thermal and utilization. When auto boost is disabled the GPU will attempt to maintain clocks at precisely the Current Application Clocks settings (whenever a CUDA context is active). With auto boost enabled the GPU will still attempt to maintain this floor, but will opportunistically boost to higher clocks when power, ther- mal and utilization headroom allow. This setting persists for the life of the CUDA con- text for which it was requested. Apps can request a particular mode either via an NVML call (see NVML SDK) or by setting the CUDA environment variable CUDA_AUTO_BOOST. Auto Boost Default Indicates the default setting for auto boost mode, either enabled (On) or disabled (Off). Shows (NIA) if boost is not supported. Apps will run in the default mode if they have not explicitly requested a particular mode. Note: Auto Boost settings can only be modified if "Persistence Mode" is enabled, which is NOT by default. Supported clocks List of possible memory and graphics clocks combinations that the GPU can operate on (not taking into account HW brake reduced clocks). These are the only clock combinations that can be passed to --appli- cations—clocks flag. Supported Clocks are listed only when —q —d SUPPORTED_CLOCKS switches are provided or in XML format. Processes List of processes having Compute or Graphics Context on the device. Compute processes are reported on all the fully supported products. Reporting for Graphics processes is limited to the supported products start- ing with Kepler architecture. Each Entry is of format "<GPU Index> <PID> <Type> <Process Name> <GPU Memory Usage>" GPU Index Represents NVML Index of the device. PID Represents Process ID corresponding to the active Compute or Graphics context. Type Displayed as "C" for Compute Process, "G" for Graphics Process, and "C+G" for the process having both Compute and Graphics contexts. Process Name Represents process name for the Compute or Graphics process. nvidia—smi 352.26 2015/7/7 15 EFTA00282812 nvidia—smi(1) NVIDIA nvidia—smi(l) GPU Memory Usage Amount of memory used on the device by the context. Not available on Windows when running in WDDM mode because Windows KMD manages all the memory not NVIDIA driver. Stats (EXPERIMENTAL) List GPU statistics such as power samples, utilization samples, xid events, clock change events and viola- tion counters. Supported on Tesla, GRID and Quadro based products under Linux. Limited to Kepler or newer GPUs. Displays statistics in CSV format as follows: <GPU device index>, <metric name>, <CPU Timestamp in us>, <value for metric> The metrics to display with their units are as follows: Power samples in Watts. GPU, Memory, Encoder and Decoder utilization samples in Percentage. Xid error events reported with Xid error code. The error code is 999 for unknown xid error. Processor and Memory clock changes in MHz. Violation due to Power capping with violation time in ns. (Tesla Only) Violation due to Thermal capping with violation boolean flag (I/0). (Tesla Only) Notes: Any statistic preceded by "#" is a comment. Non supported device is displayed as "#<device Index>, Device not supported". Non supported metric is displayed as "<device index>, <metric name>, NIA, N/A". Violation due to Thermal/Power supported only for Tesla based products. Thermal Violations are limited to Tesla K20 and higher. Device Monitoring The "nvidia•smi dmon" command-line is used to monitor one or more GPUs (up to 4 devices) plugged into the system. This tool allows the user to see one line of monitoring data per monitoring cycle. The output is in concise format and easy to interpret in interactive mode. The output data per line is limited by the termi- nal size. It is supported on Tesla, GRID, Quadro and limited GeForce products for Kepler or newer GPUs under bare metal 64 bits Linux. By default, the monitoring data includes Power Usage, Temperature, SM clocks, Memory clocks and Utilization values for SM, Memory, Encoder and Decoder. It can also be con- figured to report other metrics such as frame buffer memory usage, barl memory usage, power/thermal vio- lations and aggregate single/double bit ecc errors. If any of the metric is not supported on the device or any other error in fetching the metric is reported as ''--" in the output data. The user can also configure monitor- ing frequency and the number of monitoring iterations for each run. There is also an option to include date and time at each line. All the supported options are exclusive and can be used together in any order. Usage: nvidia—smi 352.26 2015/7/7 16 EFTA00282813 nvidia—smi(1) NVIDIA nvidia—smi(1) 1) Default with no arguments nvidia•smi dmon Monitors default metrics for up to 4 supported devices under natural enumeration (starting with GPU index 0) at a frequency of I sec. Runs until terminated with "C. 2) Select one or more devices nvidia-smi dmon <devicel,device2, , deviceN> Reports default metrics for the devices selected by comma separated device list. The tool picks up to 4 sup- ported devices from the list under natural enumeration (starting with GPU index 0). 3) Select metrics to be displayed nvidia•smi dmon -5 <metric_group> <metric_group> can be one or more from the following: p - Power Usage (in Watts) and Temperature (in C) u Utilization (SM, Memory, Encoder and Decoder Utilization in %) c Proc and Mem Clocks (in MHz) v - Power Violations (in %) and

Entities

0 total entities mentioned

No entities found in this document

Document Metadata

Document ID
0c17323c-6c03-4e9c-86ce-eef8222e0b8c
Storage Key
dataset_9/EFTA00282798.pdf
Content Hash
21bc3f13b9706cbe893f59649cced1f4
Created
Feb 3, 2026