关注
Xiaogeng Liu
Xiaogeng Liu
在 wisc.edu 的电子邮件经过验证 - 首页
标题
引用次数
引用次数
年份
Autodan: Generating stealthy jailbreak prompts on aligned large language models
X Liu, N Xu, M Chen, C Xiao
arXiv preprint arXiv:2310.04451, 2023
722023
Protecting facial privacy: Generating adversarial identity masks via style-robust makeup transfer
S Hu, X Liu, Y Zhang, M Li, LY Zhang, H Jin, L Wu
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2022
672022
Advhash: Set-to-set targeted attack on deep hashing with one single adversarial patch
S Hu, Y Zhang, X Liu, LY Zhang, M Li, H Jin
Proceedings of the 29th ACM International Conference on Multimedia, 2335-2343, 2021
262021
Detecting backdoors during the inference stage based on corruption robustness consistency
X Liu, M Li, H Wang, S Hu, D Ye, H Jin, L Wu, C Xiao
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2023
172023
Towards efficient data-centric robust machine learning with noise-based augmentation
X Liu, H Wang, Y Zhang, F Wu, S Hu
arXiv preprint arXiv:2203.03810, 2022
122022
Deceptprompt: Exploiting llm-driven code generation via adversarial natural language instructions
F Wu, X Liu, C Xiao
arXiv preprint arXiv:2312.04730, 2023
72023
Pointcrt: Detecting backdoor in 3d point cloud via corruption robustness
S Hu, W Liu, M Li, Y Zhang, X Liu, X Wang, LY Zhang, J Hou
Proceedings of the 31st ACM International Conference on Multimedia, 666-675, 2023
72023
Why Does Little Robustness Help? A Further Step Towards Understanding Adversarial Transferability
Y Zhang, S Hu, LY Zhang, J Shi, M Li, X Liu, H Jin
Proceedings of the 45th IEEE Symposium on Security and Privacy (S&P’24) 2, 2024
52024
Adashield: Safeguarding multimodal large language models from structure-based attack via adaptive shield prompting
Y Wang, X Liu, Y Li, M Chen, C Xiao
arXiv preprint arXiv:2403.09513, 2024
22024
Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Z Yu, X Liu, S Liang, Z Cameron, C Xiao, N Zhang
arXiv preprint arXiv:2403.17336, 2024
12024
JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak Attacks
W Luo, S Ma, X Liu, X Guo, C Xiao
arXiv preprint arXiv:2404.03027, 2024
2024
Automatic and Universal Prompt Injection Attacks against Large Language Models
X Liu, Z Yu, Y Zhang, N Zhang, C Xiao
arXiv preprint arXiv:2403.04957, 2024
2024
系统目前无法执行此操作,请稍后再试。
文章 1–12