First of all, greetings to all! I’m new here, just registered to seek some advice and insights. My knowledge of software and hardware is quite basic, but I’ve always been fascinated by it. Now, with the advent of Large Language Models (LLMs), my interest has peaked more than ever.
That’s why I’m planning to build a homelab to experiment with training my own models and running them locally. I’ve managed to get a good deal on an Nvidia A100 40GB, and I’m curious to know to what extent it’s possible to train LLMs with a single A100 and how much time would that take.
Thank you! looking forward to your thoughts and suggestions
Yea I figured it out, so are you telling me that I would need like 40 a100 80gb to train a model like that in 6 months?! oh I expected much different numbers.
Yes and No, it depends what you want the machine to learn. A full fledge AI to Human conversation? No you need a lot of compute.
You want an Objekt recognicion Model you call your own, that could be doable or an Picture creation Model or modefy an existing and fork it, yea more plausable
Alright I’m forming a better picture now, thanks a lot for the help.
So in order to do some fine tuning to bigger LLMs like Dolphin do you have any estimation of what would it take?
but i quick google search it seems you could fine tune LLAMA in the Mid Model size with your VRam, so i can do an eduacte gues that you can dabble in that range
i hostet LLAMA on Unraid with a Mid Model and it worked Ok’ish, but that was about a year ago.