Prompt injection is a type of vulnerability that specifically targets machine learning models employing prompt-based learning. It exploits the model's inability to distinguish between instructions and ...