Jailbreaking LLMs with ASCII Art. Turns out LLMs are still computer programs and sanitizing inputs is hard.
NSFW as it isn’t a bad take by techpeople, but research showing that the AI creating diamondoid virusses because we were mean to it fears are overblown. It cannot follow simple (for us intelligent humans) instructions not to do certain things.
LLMs are extremely good at parsing things however.
That’s hilarious, and much more efficient than when I ask it to list all the permutations of C, F, K and U.