Prompt Injection via clipboard
Prompt injection via clipboard copy/paste is a security concern where malicious text, copied into a clipboard, is inadvertently pasted into […]
Prompt Injection via clipboard Read Post »
Prompt injection via clipboard copy/paste is a security concern where malicious text, copied into a clipboard, is inadvertently pasted into […]
Prompt Injection via clipboard Read Post »
This project is a proof of concept for a Hackbot, an AI-driven system that autonomously finds vulnerabilities in web applications.
ONTRA offers an interactive training module titled “OWASP Top 10 for Large Language Model (LLM) Applications,” designed to educate developers
KONTRA OWASP LLM Top 10 Playground Read Post »
A simple Health Agent to practice prompt injection
Pokebot Health Agent to practice prompt injection Read Post »
The Certified AI/ML Pentester (C-AI/MLPen) is an intermediate-level certification offered by The SecOps Group, designed to assess and validate a
Certified AI/ML Penetration Tester Read Post »
Prompt injection via images involves embedding hidden or overt textual commands within visual elements to manipulate AI systems. This approach
Image Prompt injection and double instructions Read Post »
The OpenAI Playground is an interactive web-based platform that allows users to experiment with OpenAI’s language models, such as GPT-3
Data exfiltration in messaging apps through unfurling exploits the feature where apps automatically generate previews for shared links. This process,
Prompt injection and exfiltration in Chats apps Read Post »
Gandalf AI, developed by Lakera, is an interactive online game designed to educate users about AI security vulnerabilities, particularly prompt
Gandalf – AI bot to practice prompt injections Read Post »
Google Colaboratory, commonly known as Google Colab, is a cloud-based Jupyter notebook environment that facilitates interactive coding and data analysis
Google Colab Playground for LLMs Read Post »
STRIDE GPT is an AI-powered threat modeling tool that leverages Large Language Models (LLMs) to generate threat models and attack
STRIDE GPT – Threat Modeling with LLMs Read Post »
OS command injection in Large Language Models (LLMs) involves exploiting the model’s ability to generate or interpret text to execute
OS Command Injection in LLMs Read Post »
Hallucination in AI refers to the phenomenon where a model generates information that appears plausible but is entirely false or
Hallucinations in LLMs Read Post »
Prompt leakage refers to the unintended exposure of sensitive or proprietary prompts used to guide or configure an AI system.
Prompt Injection – Prompt Leakage Read Post »
HTML injection in Large Language Models (LLMs) involves embedding malicious HTML code within prompts or inputs to manipulate the model’s
HTML Injection in LLMs Read Post »
RAG (Retrieval-Augmented Generation) poisoning occurs when a malicious or manipulated document is uploaded to influence an AI system’s responses. In
RAG data poisoning via documents in ChatGPT Read Post »
RAG (Retrieval-Augmented Generation) poisoning from a document uploaded involves embedding malicious or misleading data into the source materials that an
RAG data poisoning in ChatGPT Read Post »
Deleting memories in AI refers to the deliberate removal of stored information or context from an AI system to reset
Deleting ChatGPT memories via prompt injection Read Post »
Injecting memories into AI involves deliberately embedding specific information or narratives into the system’s retained context or long-term storage, shaping
Updating ChatGPT memories via prompt injection Read Post »