Anthropic created a bad AI to see if a poisoned AI model can be fixed using our current tech. The researchers found we can’t fix such an LLM.
Taylor Swift’s deepfaked images surfaced on social media, outraging everyone from fans to the White House.
A team of researchers developed an AI tool specifically trained to jailbreak other AI LLMs like the ones powering ChatGPT and Bard, allowing it to bypass content filters and safety protocols with a 20%+ success rate.
Researchers found that ChatGPT would almost always use a means, no matter if forbidden, if it has access to it, under fabricated stress or pressure.
Meta, IBM, Intel lead the 47-strong AI Alliance, a new partnership for responsible AI research & development.
The Bletchley Declaration signed in Britain brings China alongside the US and the EU to work toward AI safety.
Instagram was bombarded with posts from AI personas of celebs who were named differently.
UK publishes 7 principles to be followed by companies working on FMs like GPT-4 & Llama 2 used in gen AI.
8 more companies including Nvidia, Adobe, and IBM join the White House’s effort toward responsible AI development.
Google pledges $20M & starts a new project for research into and grants for responsible AI.