Researcher, IBM Research AI
1 paper at NeurIPS 2025
We discover model-generated constitutions and train language models to intrinsically self-correct their responses by using these principles; repeating this process iteratively enables self-improvement.