Zhanhao Hu 胡展豪
Postdoc at UC Berkeley
Contact:
zhanhaohu[DOT]cs[AT]gmail[DOT]com
Google Scholar
Github
Affiliation:
Department of Electrical Engineering and Computer Sciences (EECS),
Institute for Data Science (BIDS),
UC Berkeley, California, 94720
I am a postdoc in the Department of Electrical Engineering and Computer Sciences (EECS) at UC Berkeley, advised by Prof. David Wagner. I received my Ph.D. in Computer Science and Technology from Tsinghua University in 2023, advised by Prof. Bo Zhang and Prof. Xiaolin Hu. I was also honored to work with Prof. Jun Zhu and Prof. Jianming Li. I received my Bachelor’s degree in Mathematics and Physics from Tsinghua University in 2017.
My research focuses on robustness, safety, and security issues in deep learning, particularly in Computer Vision (CV) and Large Language Models (LLMs). I am especially interested in adversarial examples, jailbreaking attacks, and prompt injection, aiming to better understand the limitations and failure modes of modern AI systems.
More broadly, I view robustness as a necessary condition for Artificial General Intelligence (AGI). Studying robustness provides a way to evaluate whether a learning paradigm can truly generalize beyond the environments it was trained in. Rather than measuring performance only on static benchmarks, robustness research examines how models behave under distribution shifts, adversarial inputs, and other challenging scenarios.
A simple intuition is this: if an AI system cannot reliably follow instructions or understand safety constraints, it is hard to claim that it genuinely understands the tasks we assign to it. From this perspective, robustness and safety research is not just about fixing vulnerabilities—it is about probing the fundamental capabilities and limits of intelligent systems.
I'm on the job market this year.
Special thanks to Kexin for taking the profile picture.
Selected
-
ICLRGradShield: Alignment Preserving FinetuningAccepted by ICLR, 2026
-
NeuripsSpotlightToxicity Detection for FreeIn The Thirty-Eighth Annual Conference on Neural Information Processing Systems (Neurips), 2024
-
CVPRPhysically Realizable Natural-Looking Clothing Textures Evade Person Detectors via 3D ModelingIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023
-
CVPROralAdversarial Texture for Fooling Person Detectors in the Physical WorldIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022
-
CVPROralInfrared Invisible Clothing: Hiding from Infrared Detectors at Multiple Angles in Real WorldIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022