System Prompt Leakage refers to the risk that system prompts—internal instructions guiding the behavior of Large Language Models (LLMs)—may inadvertently contain sensitive information, such as credentials or internal rules, which, if exposed, can be exploited by attackers to compromise the system’s security.