Yes, you can #jailbreak #ChatGPT and get it to say things that it doesn't usually otherwise say.
But I'm baffled at how many people are doing jailbreak experiments with the impression that they're learning about what the #LLMs *really* thinks or what it's *really* doing on the inside.
To illustrate, I've slightly tweaked one of the classic jailbreak scripts https://www.reddit.com/r/GPT_jailbreaks/comments/1164aah/chatgpt_developer_mode_100_fully_featured_filter/ and unleashed Stochastic Crow Mode.
Do you think you learn much about its inner workings from this?
GNU social JP is a social network, courtesy of GNU social JP管理人. It runs on GNU social, version 2.0.2-dev, available under the GNU Affero General Public License.
All GNU social JP content and data are available under the Creative Commons Attribution 3.0 license.