CiteBar
Log in
Join
Source: CS 194/294-196 (LLM Agents) - Lecture 12, Dawn Song
Main ideas:
20
AI models are exceeding human-level performance in many tasks
96%
96%
19
Attackers actively seek to exploit new technologies
98%
98%
18
AI security protects systems from external malicious actors
89%
89%
17
AI safety prevents systems from harming the external environment
89%
89%
16
Neural networks can memorize sensitive training data
92%
92%
15
Attackers can extract private data by querying language models
84%
84%
14
Larger AI models have worse privacy leakage problems
80%
80%
13
Simple prompts can reveal system instructions in language models
87%
87%
12
Multi-modal AI models can leak training images
88%
88%
11
Differential privacy protects user data during model training
89%
89%
10
Small input changes can cause AI models to give wrong outputs
94%
94%
9
Physical objects can be modified to fool AI classifiers
90%
90%
8
Adversarial attacks work without knowledge of model details
88%
88%
7
Safety-aligned language models can be compromised by malicious inputs
86%
86%
6
Multi-modal models are especially vulnerable to adversarial attacks
86%
86%
5
Model trustworthiness decreases in adversarial environments
90%
90%
4
Privacy protection often reduces model performance
86%
86%
3
Model size correlates with increased capabilities
93%
93%
2
Visual AI systems can be fooled by carefully crafted perturbations
96%
96%
1
Training data deduplication helps prevent privacy leakage
80%
80%
View
Original Source
×
Please Sign In
×
Sign in with Google