A new trick uses AI to jailbreak AI models—including GPT-4