Researcher, UK AI Security Institute
1 paper at NeurIPS 2025
Defences against LLM misuse fine-tuning attacks that aim to detect individual malicious or suspicious samples are insufficient.