Mitigating adversarial manipulation in LLMs: a prompt-based approach to counter Jailbreak attacks (Prompt-G)
Large language models (LLMs) have become transformative tools in areas like text generation, natural language processing, and conversational AI. However, their widespread use introduces security risks, such as jailbreak attacks, which exploit LLM’s vulnerabilities to manipulate outputs or extract se...
প্রধান লেখক: | , , , , , , |
---|---|
বিন্যাস: | প্রবন্ধ |
ভাষা: | English |
প্রকাশিত: |
PeerJ Inc.
2024-10-01
|
মালা: | PeerJ Computer Science |
বিষয়গুলি: | |
অনলাইন ব্যবহার করুন: | https://peerj.com/articles/cs-2374.pdf |