Investigation into Nvidia GPU workloads reveals that Tensor cores are being hammered, just incredibly briefly.

An intrepid Reddit poster, going under the handle Bluedot55, leveraged Nvidia’s Nsight Systems GPU metric tools to drill down into the workloads running on various parts of an Nvidia RTX 4090 GPU.

Bluedot55 ran both DLSS and third party scalers on an Nvidia RTX 4090 and measured Tensor core utilisation. Looking at average Tensor core usage, the figures under DLSS were extremely low, less than 1%.

Initial investigations suggested even the peak utilisation registered in the 4-9% range, implying that while the Tensor cores were being used, they probably weren’t actually essential. However, increasing the polling rate revealed that peak utilisation is in fact in excess of 90%, but only for brief periods measured in microseconds.

  • conciselyverbose@kbin.social
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    In the sense that shaders are capable of replicating the operations, sure. But the reason for tensor cores is the same reason as for any other hardware feature. It’s obscenely faster and more efficient to do math you’ll do frequently with dedicated hardware.

    • lustrum
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      I don’t know what to say. That’s what I said in my OP just more succinct.