Abstract
One of the possible ways to embed values into autonomous agents is through reasoning over the norms that govern the MAS where agents are situated. Unfortunately, most previous research on value alignment of norms does not take into consideration the strong social dimension of values. Here, we take the stance that agents should be able to reason not exclusively about their own values, but also take into account the values that others in their community hold and how they interpret them. In this work, we present a novel functionality for autonomous agents to compute the perspective-dependent value alignment of norms. We build upon and integrate previous work on value representation, normative reasoning and Theory of Mind (the ability to perceive and interpret others in terms of their mental states, such as beliefs). This novel functionality enables an agent to compute the alignment of a set of norms with respect to a set of values not exclusively from its opinion perspective, but to switch its value structure and perception of the world at run-time using Theory of Mind, to estimate the alignment that another agent may have for the same set of norms. Our proposal opens new grounds for research on value-based negotiation over normative systems, where agents can perform better if they can estimate the opinion that their interlocutors have on the proposals they make.
Supported by the Spanish-funded VAE project (#TED2021-131295B-C31) and RYHMAS project (#PID2020-113594RB-100), the EU VALAWAI project (HORIZON #101070930), and the EU TAILOR project (H2020 #952215).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
The clauses displayed in this section are written in Jason agent code. It follows a syntax very similar to that of Prolog, but using ampersand “ &” instead of comma for conjunction. Additionally, Jason built-in predicates are preceded by a dot “.”.
- 2.
- 3.
References
Andrighetto, G., Governatori, G., Noriega, P., van der Torre, L.: Normative multi-agent systems (Dagstuhl Seminar 12111). Dagstuhl Rep. 2(3), 23–49 (2012). https://doi.org/10.4230/DagRep.2.3.23. http://drops.dagstuhl.de/opus/volltexte/2012/3535
Atkinson, K., Bench-Capon, T.: States, goals and values: revisiting practical reasoning. Argum. Comput. 7(2–3), 135–154 (2016). https://doi.org/10.3233/aac-160011
Benzmüller, C., Parent, X., van der Torre, L.: Designing normative theories for ethical and legal reasoning: LogiKEy framework, methodology, and tool support. Artif. Intell. 287, 103348 (2020). https://doi.org/10.1016/j.artint.2020.103348
Berreby, F., Bourgne, G., Ganascia, J.G.: A declarative modular framework for representing and applying ethical principles. In: Proceedings of the 16th Conference on Autonomous Agents and MultiAgent Systems. AAMAS 2017, International Foundation for Autonomous Agents and Multiagent Systems, Richland, SC, pp. 96–104 (2017)
Bordini, R.H.: Programming Multi-agent Systems in AgentSpeak Using Jason. John Wiley, Hoboken (2007)
Cervantes, J.A., López, S., Rodríguez, L.F., Cervantes, S., Cervantes, F., Ramos, F.: Artificial moral agents: a survey of the current status. Sci. Eng. Ethics 26(2), 501–532 (2019). https://doi.org/10.1007/s11948-019-00151-x
Conte, R., Castelfranchi, C.: From conventions to prescription: towards an integrated view of norms. Artif. Intell. Law 7(4), 323–340 (1999). https://doi.org/10.1023/a:1008310107755
Cranefield, S., Winikoff, M., Dignum, V., Dignum, F.: No pizza for you: value-based plan selection in BDI agents. In: Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence. International Joint Conferences on Artificial Intelligence Organization (2017). https://doi.org/10.24963/ijcai.2017/26
Frith, C., Frith, U.: Theory of mind. Curr. Biol. 15(17), R644–R645 (2005). https://doi.org/10.1016/j.cub.2005.08.041
González-Díaz, J., García-Jurado, I., Fiestras-Janeiro, M.G.: An introductory course on mathematical game theory. American Mathematical Society and Real Sociedad Matemática Española, Providence, Rhode Island, USA and Madrid (2010)
Lorini, E.: A logic for reasoning about moral agents. Logique et Analyse 58(230), 177–218 (2015). http://www.jstor.org/stable/44085321
Montes, N., Luck, M., Osman, N., Rodrigues, O., Sierra, C.: Combining theory of mind and abductive reasoning in agent-oriented programming. Auton. Agents Multi-Agent Syst. 37(2) (2023). https://doi.org/10.1007/s10458-023-09613-w
Montes, N., Osman, N., Sierra, C.: A computational model of Ostrom’s institutional analysis and development framework. Artif. Intell. 311, 103756 (2022). https://doi.org/10.1016/j.artint.2022.103756
Montes, N., Sierra, C.: Synthesis and properties of optimally value-aligned normative systems. J. Artif. Intell. Res. 74, 1739–1774 (2022). https://doi.org/10.1613/jair.1.13487
Ostrom, E.: Governing the Commons. Cambridge University Press (1990). https://doi.org/10.1017/cbo9780511807763
Ostrom, E.: Understanding Institutional Diversity. Princeton University Press, Princeton (2005)
Ostrom, E., Gardner, R., Walker, J.: Rules, Games, and Common-Pool Resources. University of Michigan Press, Ann Arbor (1994). https://doi.org/10.3998/mpub.9739
Rohan, M.J.: A rose by any name? the values construct. Pers. Soc. Psychol. Rev. 4(3), 255–277 (2000). https://doi.org/10.1207/s15327957pspr0403_4
Schwartz, S.H.: Universals in the content and structure of values: theoretical advances and empirical tests in 20 countries. In: Advances in Experimental Social Psychology, pp. 1–65. Elsevier (1992).https://doi.org/10.1016/s0065-2601(08)60281-6
Serramià, M., López-Sánchez, M., Rodríguez-Aguilar, J.A.: A qualitative approach to composing value-aligned norm systems. In: Proceedings of the 19th International Conference on Autonomous Agents and MultiAgent Systems. AAMAS 2020, International Foundation for Autonomous Agents and Multiagent Systems, Richland, SC, pp. 1233-1241 (2020)
Serramià, M., et al.: Exploiting moral values to choose the right norms. In: Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society. ACM (2018). https://doi.org/10.1145/3278721.3278735
Sierra, C., Osman, N., Noriega, P., Sabater-Mir, J., Perelló-Moragues, A.: Value alignment: a formal approach. In: Responsible Artificial Intelligence Agents Workshop (RAIA) in AAMAS 2019 (2019)
Szabo, J., Such, J.M., Criado, N.: Understanding the role of values and norms in practical reasoning. In: Bassiliades, N., Chalkiadakis, G., de Jonge, D. (eds.) EUMAS/AT -2020. LNCS (LNAI), vol. 12520, pp. 431–439. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-66412-1_27
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Montes, N., Osman, N., Sierra, C. (2024). Perspective-Dependent Value Alignment of Norms. In: Osman, N., Steels, L. (eds) Value Engineering in Artificial Intelligence. VALE 2023. Lecture Notes in Computer Science(), vol 14520. Springer, Cham. https://doi.org/10.1007/978-3-031-58202-8_4
Download citation
DOI: https://doi.org/10.1007/978-3-031-58202-8_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-58204-2
Online ISBN: 978-3-031-58202-8
eBook Packages: Computer ScienceComputer Science (R0)