It's easy to run the smaller models locally with a GPU, no one needs a stupid video to teach them when the info is easily found. Good luck running GPT3+ or even getting hold of the weights to deploy on a compute node.
Given how OpenAI has kept GPT3 under tight control, I doubt the weights will ever be released officially, but other freely available models might come close eventually; I suppose the weights might end up being leaked. Resource-wise they are very heavy due to the number of parameters so running on consumer grade hardware will be tricky.
It's easy to run the smaller models locally with a GPU, no one needs a stupid video to teach them when the info is easily found. Good luck running GPT3+ or even getting hold of the weights to deploy on a compute node.
I feel like it's only a matter of time.
Given how OpenAI has kept GPT3 under tight control, I doubt the weights will ever be released officially, but other freely available models might come close eventually; I suppose the weights might end up being leaked. Resource-wise they are very heavy due to the number of parameters so running on consumer grade hardware will be tricky.