Postdoc, The Hong Kong University of Science and Technology
1 paper at NeurIPS 2025
This paper reveals that backdoored LLM are more susceptible to new trigger injection during in-context learning (ICL) and proposes a targeted ICL-based lightweight backdoor detection framework for black-box LLMs used for generative tasks.