As LLMs begin to integrate multimodal capabilities, attackers could use hidden instructions in images and audio to get a chatbot to respond the way they want, say researchers at Black Hat Europe 2023.
Google Dorks Database |
Exploits Vulnerability |
Exploit Shellcodes |
CVE List |
Tools/Apps |
News/Aarticles |
Phishing Database |
Deepfake Detection |
Trends/Statistics & Live Infos |
Tags:
LLMs Open to Manipulation Using Doctored Images, Audio