Analyzing Modern Nvidia GPU Cores

(arxiv.org)

114 points | by mfiguiere 9 hours ago ago

14 comments

  • winwang 8 hours ago ago

    I hope this can help shed the misconception that GPUs are only good at linear algebra and FP arithmetic, which I've been hearing a whole lot!

    Edit: learned a bunch, but the "uniform" registers and 64-bit (memory) performance are some easy standouts.

    • remcob 4 hours ago ago

      It’s well known GPUs are good at cryptography. Starting with hash functions (e.g. crypto mining) but also zero knowledge proofs and multi party computation.

    • YetAnotherNick 41 minutes ago ago

      In a sense, GPUs are only great at matrix-matrix multiplication. For anything else you would only get 7% of the FLOPs/s compared to it(989 vs 67 TFLOP/s for H100)[1].

      [1]: https://www.nvidia.com/en-in/data-center/h100/

      • winwang 9 minutes ago ago

        lol, I haven't thought about it like that, true. though of course, I mean compared to CPUs :P

        I try and use tensor cores for non-obvious things every now and then. The most promising so far seems to be for linear arithmetic in Datalog, but that's just matrix-vector/gemv

    • qwertox 4 hours ago ago

      Wasn't it well known that CUDA cores are programmable cores?

      • winwang 3 hours ago ago

        Haha, if you're the type to toss out the phrase "well known", then yes!

  • kookamamie 2 hours ago ago

    > NVIDIA RTX A6000

    Unfortunately that's already behind the latest GPU by two generations. You'd have these after A6000: 6000 Ada, Pro 6000.

    • pjmlp 2 hours ago ago

      Still better than most folks have access to.

      I bet I can do more CUDA with my lame GeForce MX 150 from 2017, than what most people can reach for to do ROCm, and that is how NVidia keeps being ahead.

      • kookamamie 9 minutes ago ago

        Yeah, kind of. I have an 6000 Ada and 5090 here.

    • flowerthoughts 2 hours ago ago

      It's a major step forward compared to 2006.

      A6000 was released in 2020: https://www.techpowerup.com/gpu-specs/rtx-a6000.c3686

    • KeplerBoy 2 hours ago ago

      Nvidia's Quadro naming scheme really is bad these days, isn't it?

      I bet there are plenty of papers out there claiming to have used a RTX 6000 instead of a RTX 6000 Ada gen.

      • kookamamie 7 minutes ago ago

        The naming scheme is horrible, to be quite frank.

        To understand this, consider these names in the order of release time: Quadro RTX 6000, RTX A6000, RTX 6000 Ada, RTX Pro 6000, RTX Pro 6000 Max-Q.

  • gitroom 2 hours ago ago

    Haha honestly I always thought GPUs were mostly number crunchers, but there's way more under the hood than I realized. Wondering now if anyone really gets the full potential of these cores, or if we're all just scratching the surface most days?

  • gmays 5 hours ago ago

    The special sauce:

    > "GPUs leverage hardware-compiler techniques where the compiler guides hardware during execution."