#prompt_injection
Prompt injection
Type of attack in machine learning
Prompt injection is a family of related computer security exploits carried out by getting a machine learning model which was trained to follow human-given instructions to follow instructions provided by a malicious user. This stands in contrast to the intended operation of instruction-following systems, wherein the ML model is intended only to follow trusted instructions (prompts) provided by the ML model's operator.
Thu 19th
Provided by Wikipedia
This keyword could refer to multiple things. Here are some suggestions:
0 searches
This keyword has never been searched before
This keyword has never been searched for with any other keyword.