4.9 (596) · € 25.00 · En Stock
Large Language Models can craft poetry, answer queries, and even write code. Yet, with immense power comes inherent risks. The same prompts that enable LLMs to engage in meaningful dialogue can be manipulated with malicious intent. Hacking, misuse, and a lack of comprehensive security protocols can turn these marvels of technology into tools of deception.
Prompt Hacking and Misuse of LLMs
LLM Security Concerns Shine a Light on Existing Data Vulnerabilities
OWASP Top 10 for LLMs: An Overview with SOCRadar
Exploring Prompt Injection Attacks, NCC Group Research Blog
Adversarial Robustness Could Help Prevent Catastrophic Misuse — AI Alignment Forum
What you didn't want to know about prompt injections in LLM apps, by Jakob Cassiman
OWASP Top 10 for Large Language Model Applications Explained: A Practical Guide
TrustLLM: Trustworthiness in Large Language Models
Prompt Hacking and Misuse of LLMs
7 methods to secure LLM apps from prompt injections and jailbreaks [Guest]
Not What You've Signed Up For: Compromising Real-World LLM-Integrated Applications With Indirect Prompt Injection, PDF, Malware
What Are Large Language Models Capable Of: The Vulnerability of LLMs to Adversarial Attacks
🟢 Jailbreaking Learn Prompting: Your Guide to Communicating with AI
RLAIF: Scaling Reinforcement Learning from AI feedback
Data Privacy Challenges Around Generative AI Models Like LLMs