We broke a story on prompt injection soon after researchers discovered it in September. It's a method that can circumvent previous instructions in a language model prompt and provide new ones in their ...