A jailbreaking technique called "Skeleton Key" lets users persuade OpenAI's GPT 3.5 into giving them the recipe for all kind of dangerous things.
Generative AI models have a larger attack surface than many CSOs might think. Microsoft Azure’s CTO walked through some of the more significant challenges facing developers and defenders. Generative ...