Holy smokes you have more disk i/o than my overclocked DDR5!
I’m assuming you didn’t get your hands on 2x 5090TIs yesterday given the whole “Paper Launch” as Steve at Gamer Nexus called it haha…
I found this thread because a post on r/LocalLLaMA cited your benchmarks. (for transparency I started that post that randomly blew up).
Anyway, given you seem to be interested in “AI Studies”, I’m curious if you have already or were interested in trying to run the very trendy DeepSeek R1 671B model on your beast of a rig? (not the little distill models, but a dynamic quant of the original fp8 etc…)?
For these very large MoE models I’m wondering if instead of using PCIe lanes on GPUs, to spend them on fast read IOPS storage for “cheap VRAM” assuming CPU inferencing is not the bottle neck.
I just started a post on Level1Techs about it here
No pressure, and thanks for sharing all the data points!