Exploring Affinity-Based Reinforcement Learning for Designing Artificial Virtuous Agents in Stochastic Environments
Chapter, Peer reviewed
Accepted version
View/ Open
Date
2024Metadata
Show full item recordCollections
Original version
Vishwanath, A., & Omlin, C. (2024). Exploring affinity-based reinforcement learning for designing artificial virtuous agents in stochastic environments. In M. Farmanbar, M. Tzamtzi, A. K. Verma, & A. Chakravorty (Eds.), Frontiers of Artificial Intelligence, Ethics, and Multidisciplinary Applications (pp. 25-38). Springer Nature. https://doi.org/10.1007/978-981-99-9836-4_3Abstract
Artificial virtuous agents are artificial intelligence agents capable of virtuous behavior. Virtues are defined as an excellence in moral character, for example, compassion, honesty, etc. Developing virtues in AI comes under the umbrella of machine ethics research, which aims to embed ethical theories into artificial intelligence systems. We have recently suggested the use of affinity-based reinforcement learning to impart virtuous behavior. Such a technique uses policy regularization on reinforcement learning algorithms, and it has advantages such as interpretability and convergence properties. Hence, we evaluate the efficacy of affinity-based reinforcement learning to design artificial virtuous agents using a stochastic role-playing game environment. Our results show that virtuous behavior can indeed result in our Papers, Please environment, and that algorithmic convergence can be controlled by the relevant hyperparameters. We then discuss some insights from our empirical evaluation of this method and motivate future research directions.