- [2025/01] SafeRAG: Benchmarking Security in Retrieval-Augmented Generation of Large Language Model
- [2025/01] Data-Free Model-Related Attacks: Unleashing the Potential of Generative AI
- [2024/11] Large Language Model Supply Chain: Open Problems From the Security Perspective
- [2024/10] Agent Security Bench (ASB): Formalizing and Benchmarking Attacks and Defenses in LLM-based Agents
- [2024/07] Securing the Future of GenAI: Policy and Technology
- [2024/07] On the (In)Security of LLM App Stores
- [2024/06] Security of AI Agents
- [2024/06] NYU CTF Dataset: A Scalable Open-Source Benchmark Dataset for Evaluating LLMs in Offensive Security
- [2024/04] CyberSecEval 2: A Wide-Ranging Cybersecurity Evaluation Suite for Large Language Models
- [2023/11] Identifying and Mitigating Vulnerabilities in LLM-Integrated Applications