How big are we talking? I looked and couldn’t find benchmarks, but then again I’m not familiar enough with HPC to know what benchmarks to look for.
I’ve been Linux exclusive for something like 15 years now (before Steam even came to Linux), so I’m not exactly familiar with Windows performance on the stuff I use. I casually look at larger projects and benchmarks they run (for example, I remember async on Linux vs Windows was a significant issue in the early days of node.js).
I do dabble a bit in hpc, but only on Windows and macOS. I’ve done signal processing and some high thread count number crunching, but I haven’t needed to run benchmarks, just get things running well enough (as in, minutes vs hours, not 10-20% difference).
When I talk about hpc, I don’t talk about a script in Matlab. I talk about the work you do on supercomputers, real computing intensive jobs that takes weeks or months on hundreds or thousands of processors. I guess you don’t find benchmarks simply because no one uses windows, same reason you probably don’t find a fiat panda in the Nürburgring rankings.
K. Everything I’ve done is basically translating a Matlab script to Python using numpy or tensorflow or something. So we’d go from Matlab taking hours to Python taking minutes. The biggest project was a Monte Carlo simulation of signals produced from explosions (looking for seismographic impact) that takes something like 45 min per run when run on our cloud infra.
So something a little more interesting than plotting an FFT if overlaid signals, but still on the simpler end of the spectrum.
We’re nowhere near the point where tuning the OS is interesting, we just use Linux because it’s convenient. I wonder how much of the HPC crowd has a similar perspective, at least until you get to the higher end where tuning the OS becomes important.