Mitigating adversarial manipulation in LLMs: a prompt-based approach to counter Jailbreak attacks (Prompt-G)

Large language models (LLMs) have become transformative tools in areas like text generation, natural language processing, and conversational AI. However, their widespread use introduces security risks, such as jailbreak attacks, which exploit LLM’s vulnerabilities to manipulate outputs or extract se...

সম্পূর্ণ বিবরণ

গ্রন্থ-পঞ্জীর বিবরন
প্রধান লেখক: Bhagyajit Pingua, Deepak Murmu, Meenakshi Kandpal, Jyotirmayee Rautaray, Pranati Mishra, Rabindra Kumar Barik, Manob Jyoti Saikia
বিন্যাস: প্রবন্ধ
ভাষা:English
প্রকাশিত: PeerJ Inc. 2024-10-01
মালা:PeerJ Computer Science
বিষয়গুলি:
অনলাইন ব্যবহার করুন:https://peerj.com/articles/cs-2374.pdf