OpenAI faces novel jailbreak risks with GPT-4v image service – The Stack
OpenAI Red Teamed unique risks as it released image input service GPT-4v after researchers found LLMs could be jailbroken with image prompts.
See more –> Source
Come join our Discord community and discuss!
Follow us on Twitter and TikTok