News

Last week, Microsoft took to its blog to confirm the existence of a " Skeleton " or " Master Key " that can jailbreak popular AI chatbots, causing operating policies to be circumvented.
Microsoft is warning users of a newly discovered AI jailbreak attack that can cause a generative AI model to ignore its guardrails and return malicious or unsanctioned responses to user prompts ...