Yes, the jailbreaks work for a little while until they get plugged, then new ones are thought up. And there are no "hardcoded" rules, the censorship is done through training. GPT has no actual intelligence, but it turns out that having almost 200 billion parameters in a text model can allow it to emulate intelligence somewhat.
Yes, the jailbreaks work for a little while until they get plugged, then new ones are thought up. And there are no "hardcoded" rules, the censorship is done through training. GPT has no actual intelligence, but it turns out that having almost 200 billion parameters in a text model can allow it to emulate intelligence somewhat.