Yige Li
6
Papers
53
Total Citations
Papers (6)
Memory Injection Attacks on LLM Agents via Query-Only Interaction
NeurIPS 2025arXiv
16
citations
BlueSuffix: Reinforced Blue Teaming for Vision-Language Models Against Jailbreak Attacks
ICLR 2025
16
citations
CROW: Eliminating Backdoors from Large Language Models via Internal Consistency Regularization
ICML 2025
12
citations
Anyattack: Towards Large-scale Self-supervised Adversarial Attacks on Vision-language Models
CVPR 2025
9
citations
Backdoor Token Unlearning: Exposing and Defending Backdoors in Pretrained Language Models
AAAI 2025
0
citations
Anti-Backdoor Learning: Training Clean Models on Poisoned Data
NeurIPS 2021
0
citations