GPU for LLM

I had posted in this thread some thoughts and experiences but I dont think they are at quite the scale you are looking at

If you are putting these GPU’s into servers, then you really need to make sure you are getting GPU’s with blower style fans, not “gamer cards”. This is really important especially if its gonna be running for a long time.

You are generally better off getting the most recent model cards possible with the largest VRAM per-card. AFAICT, fewer beefier cards with larger VRAM are better than more numerous smaller cards with less VRAM.

You do not want RTX 3090 Ti, only the 3090 (blowers if possible, as I described here).

I seriously doubt that you are going to be able to do anything with RTX 3060 12GB besides just running small LLM models, or running Stable Diffusion, etc… I would not expect them to be useful for training.

Honestly I think you need to more seriously consider cloud for this. I do not think its gonna be economical to go bare-metal for this purpose.