DRAFT — COMMENTS WELCOME
Quick Background
I did this setup on my 96 core Threadripper Falcon Northwest system because its the nicest computer that I own. It also has two RTX 5580s in it (which is sort of the 300w “Max Q” equivalent of that generation.
@wFNWtr:~$ nvidia-smi
Sun May 11 16:45:55 2025
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 575.51.03 Driver Version: 575.51.03 CUDA Version: 12.9 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA RTX PRO 6000 Blac... Off | 00000000:C1:00.0 On | Off |
| 30% 37C P0 70W / 600W | 1559MiB / 97887MiB | 2% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
| 0 N/A N/A 5377 G /usr/lib/xorg/Xorg 820MiB |
| 0 N/A N/A 6998 G /usr/bin/gnome-shell 306MiB |
| 0 N/A N/A 8266 G .../6103/usr/lib/firefox/firefox 358MiB |
+-----------------------------------------------------------------------------------------+
You need pytorch / torch cu128 (or newer) and the 126 will not do it. There is also some weird circular dependency with cuda-drivers
as of the time I’m writing this, so the manual driver download from nvidia.com’s driver section or the driver utility in Ubuntu are really not the best way to go.
This page starts with downloading the nvidia keyring driver, and that gets you most of the way there:
First, I did
apt install cuda-toolkit-12-9 nvidia-open
then after that succeded I rebooted. That worked!
To get pyTorch up and running you need something like
bin/pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu128
to grab the cu128 version.
Huge thanks to @eousphoros for helping out in this one, this was the fastest path to sanity on both Ubuntu 25 and Ubuntu 24.04 LTs. Kudos, and good work.
Linux wFNWtr 6.11.0-25-generic #25~24.04.1-Ubuntu SMP PREEMPT_DYNAMIC Tue Apr 15 17:20:50 UTC 2 x86_64 x86_64 x86_64 GNU/Linux
I also like this script, what we usually use in live streams and hangouts, for testing.
Quick setup guide:
# grab Blackwell-aware deps
pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu128
# run script
./bin/python3 ./mamf-finder.py --m_range 0 4096 256 --n_range 0 4096 256 --k_range 0 20480 256 --output_file=2025-05-08-14:50:42.txt
Here There Be Tigers
Nvidia themselves are still debugging. These cards are not widely distributed yet, so there may be bugs, fwiw.
Ballpark Numbers
RTX Pro 6000 @ 600W: Deepseek-R1 Q8 70b Distill: 19.94 t/s
RTX Pro 6000 @ 300W: Deepseek-R1 Q8 70b Distill: 16.4 t/s
MAMF @ 300W: 377.5 (max) TFLOPS (288.4 median)
MAMF @ 600w: 414.4 (max) TFLOPS (404.0 median)
MAMF @ 450w: 391.5 (max) TFLOPS (374.4 median)