Skeleton Key Can ‘Jailbreak’ Most of the Biggest AI Models – Business Insider
A jailbreaking technique called “Skeleton Key” lets users persuade OpenAI’s GPT 3.5 into giving them the recipe for all kind of dangerous things.
See more –> Source
Connect with us on X