Modeling Moral Choices in Social Dilemmas with Multi-Agent Reinforcement Learning

Published in The 32nd International Joint Conference On Artificial Intelligence (IJCAI'23), 2023

Recommended citation: Tennant, E., Hailes, S., Musolesi, M. (2023). "Modeling Moral Choices in Social Dilemmas with Multi-Agent Reinforcement Learning." The 32nd International Joint Conference On Artificial Intelligence (IJCAI'23) https://doi.org/10.24963/ijcai.2023/36

We define intrisic rewards for reinforcement learning agents based on various classic moral philosophies, and study agent behaviours and emerging outcomes in (multi-agent) social dilemma settings.

We focus especially on modelling and evaluating interactions between agents who differ in their moral values, attempting to create insights for morally diverse societies (which are arguably more like the real world). The applications of this research could be two-fold: 1) design of more ethical AI agents for the real world, and 2) insights for human moral behaviour in societies.

This is the first public piece of work from my PhD, so I’m very keen to hear any feedback!

Conference paper

Appendix only or arXiv version (with Appendix)

Slides

Poster

Poster