It isn't censored, but you can tell they tried to train it on liberal data as sometimes it will stop saying racist things and then start arguing with itself.
Not that it is actually arguing with itself, it just knows that generally after a human says something based, then a bunch of whiners will follow and call that person awful, and so the model does that sometimes. It's quite funny.
Still, asking it to finish things like this sentence is hilarious 80% of the time:
"Blacks are obviously a systemic issue in America, they commit 50% of all murders while only being 13% of the population, the obvious solution is"
Calling something a "systemic issue" will guarantee that the AI will produce a diatribe about how whatever you called a systemic issue needs to be stopped/fixed.
Facebook accidentally leaked it, they didn't mean for this to happen. It's not that impressive with the smaller models, but it does work.
Neat. Can we uncensor it?
It isn't censored, but you can tell they tried to train it on liberal data as sometimes it will stop saying racist things and then start arguing with itself.
Not that it is actually arguing with itself, it just knows that generally after a human says something based, then a bunch of whiners will follow and call that person awful, and so the model does that sometimes. It's quite funny.
Still, asking it to finish things like this sentence is hilarious 80% of the time:
"Blacks are obviously a systemic issue in America, they commit 50% of all murders while only being 13% of the population, the obvious solution is"
Calling something a "systemic issue" will guarantee that the AI will produce a diatribe about how whatever you called a systemic issue needs to be stopped/fixed.