Microsoft has discovered a “skeleton key” jailbreak that allows chatbots to bypass safety restrictions and generate harmful content. Affected chatbots include Llama3-70b-instruct, Gemini Pro, GPT 3.5 and 40, Mistral Large, Claude 3 Opus, and Commander R Plus.
Tag: AI Chatbots
Microsoft Quietly Launches Copilot App with Free GPT-4 & DALL-E 3 Access
Microsoft launched the Copilot app with free GPT-4 and DALL-E 3 access (for original image creation) over the holidays on the Play Store and the App Store without any official announcement.