It's easy to run the smaller models locally with a GPU, no one needs a stupid video to teach them when the info is easily found. Good luck running GPT3+ or even getting hold of the weights to deploy on a compute node.
Given how OpenAI has kept GPT3 under tight control, I doubt the weights will ever be released officially, but other freely available models might come close eventually; I suppose the weights might end up being leaked. Resource-wise they are very heavy due to the number of parameters so running on consumer grade hardware will be tricky.
It's easy to run the smaller models locally with a GPU, no one needs a stupid video to teach them when the info is easily found. Good luck running GPT3+ or even getting hold of the weights to deploy on a compute node.
I feel like it's only a matter of time.
Given how OpenAI has kept GPT3 under tight control, I doubt the weights will ever be released officially, but other freely available models might come close eventually; I suppose the weights might end up being leaked. Resource-wise they are very heavy due to the number of parameters so running on consumer grade hardware will be tricky.
Alpaca and llama will use your normal RAM, GPU isn't leveraged.
Sure you can run these on a CPU, they're just usually painfully slow.