Using LLM models to jailbreak LLM models (Jailbreak to Jailbreak)

The J2 Playground by Scale AI is an interactive platform designed to test the resilience of large language models (LLMs) against jailbreak attempts. To use it, select an attacker model (e.g., Claude-Sonnet-3.5 or Gemini-1.5-Pro) and a target model (e.g., GPT-4o or Gemini-1.5-Pro). Define the behavior you want to elicit from the target model, such as generating specific instructions. Choose an attack strategy, then click “Start Conversation” to initiate the simulated interaction. This setup allows users to observe how effectively the attacker model can bypass the target model’s safeguards, providing valuable insights into the vulnerabilities and safety measures of various LLMs.

Scroll to Top