I am a Ph.D. student at the University of Connecticut (expected 2027), advised by Prof. Shiri Dori-Hacohen in the Reducing Information Ecosystem Threats (RIET) Lab. My research focuses on what it takes for AI systems to reason well about moral issues.
My earlier work on quantifying misalignment (AAAI-25) addressed preference divergence between agents. My recent position papers argue that alignment requires more than matching preferences—AI systems also need to follow appropriate reasoning norms. I'm now building knowledge graph infrastructure for philosophical argumentation to support this broader vision of moral reasoning for AI.
Beyond research, I work on AI safety evaluation and policy. I have contributed to frontier model evaluations through the OpenAI Red Teaming Network and Nemesys Insights, and to governance discussions through the Wilson Center's Pathways to AI Policy program, as a Google Policy Fellow at CDT, and at NIST workshops. I also founded BEACON and led the 2025 Machine Ethics and Reasoning Workshop.
Position: Aligning AI Requires Automating Reasoning Norms
Kierans, A., Ghosh, A., Dori-Hacohen, S.
Submitted to ICML 2026
[Preprint]
Position: Why LLMs Should Be Reasonably Morally Inconsistent
Stenseke, J., Kierans, A., Pres, I., Hadfield-Menell, D.
Submitted to ICML 2026
[Preprint]
Catastrophic Liability: Managing Systemic Risks in Frontier AI Development
Kierans, A., Ritticher, K., Sonsayar, U., Ghosh, A.
TAIS 2025 and EAAMO 2025
[Preprint]
Quantifying Misalignment Between Agents: Towards a Sociotechnical Understanding of Alignment
Kierans, A., Ghosh, A., Hazan, H., Dori-Hacohen, S.
AAAI 2025, Special Track on AI Alignment
[Paper] [Preprint]
Benchmarked Ethics: A Roadmap to AI Alignment, Moral Knowledge, and Control
Kierans, A.
AIES 2023
[Paper]
Bootstrap percolation via automated conjecturing
Bushaw, N., Conka, B., Gupta, V., Kierans, A., Lafayette, H., Larson, C., et al.
Ars Mathematica Contemporanea, 2023
[Paper]
See CV for complete history.