You must log in or register to comment.
“NVIDIA was once again the only company to run all MLPerf tests. H100 GPUs demonstrated the fastest performance and the greatest scaling in each of the nine benchmarks.”
Wow I wonder how long that would have taken in Turing/Volta era hardware.
What does it mean to train on such-and-such number of tokens?
How much of this performance is just the usual cheap tricks like software optimizations and task specific hardware that doesn’t generalize? If it was an honest test of pure hardware muscle the AMD MI300 would come out on top.
The benchmarks test workloads people are actually interested in running.
Whether good performance is achieved by good hardware, a good software stack, or both doesn’t matter to the vast majority of buyers.