3 papers across 3 sessions
LLMs show a utilitarian boost in moral judgment when reasoning in groups, similar to humans, but driven by distinct model-specific mechanisms, highlighting key considerations for multi-agent alignment and moral reasoning.
We argue that lethal autonomous weapon systems that use AI are a unique class of weapon system with unique risks that needs novel regulation informed by the AI community in order to mitigate those risks.
We measure directional bias amplification in balanced and unbalanced classification datasets