Binyan Xu
Ph.D. student in Information Engineering at The Chinese University of Hong Kong. My research explores how advanced models like CLIP and embodied agents can be both defenders and attackers in AI security.
- Backdoor attacks & defenses for DNNs
- Universal & transferable adversarial attacks
- Using CLIP / VLMs / LLMs for AI security
- Security of embodied agents & robotic systems
Research Overview
Friend and Foe: Exploring the dual role of vision-language models in AI safety.
- UnivIntruder: Using a single public VLM to craft universal, transferable and targeted adversarial perturbations that can hijack fully black-box image classifiers.
- CLIP-Guided Defense: Leveraging CLIP as an external semantic inspector to separate poisoned from clean data and defend against a wide range of backdoor attacks.
- Environment-Driven Jailbreaks (in progress): Studying how carefully designed physical environments around embodied agents can bypass safety mechanisms via their perception pipelines.
Selected Publications
First-author works on AI security, vision-language models and backdoor robustness.
Selected Project
Industry research experience at Microsoft Research Asia (MSRA).
- an autoregressive vision-language layout generator that produces structured, constraint-aware designs,
- a diffusion-based module that refines style and texture while preserving layout semantics.
Awards & Education
Academic training and recognitions along the way.
- Finalist Award, Mathematical Contest in Modeling (MCM)
- Qian Xuesen Program Honorary Graduate
- Outstanding Graduates Scholarship
- First-class Funding for Studying Abroad (~¥100k)
Contact & Links
The fastest way to reach me is via email.
I am always happy to discuss AI security, backdoor attacks/defenses, and vision-language or embodied AI. If you are interested in collaboration, feel free to send me a short email describing your idea.
Email: binyxu@ie.cuhk.edu.hk