Adherence Improves Cooperation in Sequential Social Dilemmas
Social dilemmas have guided research on mutual cooperation for decades, especially the two-person social dilemma. Most famously, Tit-for-Tat performs very well in tournaments of the Prisoner’s Dilemma. Nevertheless, they treat the options to cooperate or defect only as an atomic action, which cannot...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-08-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/12/16/8004 |
_version_ | 1827600931091382272 |
---|---|
author | Yuyu Yuan Ting Guo Pengqian Zhao Hongpu Jiang |
author_facet | Yuyu Yuan Ting Guo Pengqian Zhao Hongpu Jiang |
author_sort | Yuyu Yuan |
collection | DOAJ |
description | Social dilemmas have guided research on mutual cooperation for decades, especially the two-person social dilemma. Most famously, Tit-for-Tat performs very well in tournaments of the Prisoner’s Dilemma. Nevertheless, they treat the options to cooperate or defect only as an atomic action, which cannot satisfy the complexity of the real world. In recent research, these options to cooperate or defect were temporally extended. Here, we propose a novel adherence-based multi-agent reinforcement learning algorithm for achieving cooperation and coordination by rewarding agents who adhere to other agents. The evaluation of adherence is based on counterfactual reasoning. During training, each agent observes the changes in the actions of other agents by replacing its current action, thereby calculating the degree of adherence of other agents to its behavior. Using adherence as an intrinsic reward enables agents to consider the collective, thus promoting cooperation. In addition, the adherence rewards of all agents are calculated in a decentralized way. We experiment in sequential social dilemma environments, and the results demonstrate the potential for the algorithm to enhance cooperation and coordination and significantly increase the scores of the deep RL agents. |
first_indexed | 2024-03-09T04:45:06Z |
format | Article |
id | doaj.art-8cb4314113e549828269f3e0dcf18a81 |
institution | Directory Open Access Journal |
issn | 2076-3417 |
language | English |
last_indexed | 2024-03-09T04:45:06Z |
publishDate | 2022-08-01 |
publisher | MDPI AG |
record_format | Article |
series | Applied Sciences |
spelling | doaj.art-8cb4314113e549828269f3e0dcf18a812023-12-03T13:16:41ZengMDPI AGApplied Sciences2076-34172022-08-011216800410.3390/app12168004Adherence Improves Cooperation in Sequential Social DilemmasYuyu Yuan0Ting Guo1Pengqian Zhao2Hongpu Jiang3School of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, ChinaSchool of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, ChinaSchool of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, ChinaSchool of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, ChinaSocial dilemmas have guided research on mutual cooperation for decades, especially the two-person social dilemma. Most famously, Tit-for-Tat performs very well in tournaments of the Prisoner’s Dilemma. Nevertheless, they treat the options to cooperate or defect only as an atomic action, which cannot satisfy the complexity of the real world. In recent research, these options to cooperate or defect were temporally extended. Here, we propose a novel adherence-based multi-agent reinforcement learning algorithm for achieving cooperation and coordination by rewarding agents who adhere to other agents. The evaluation of adherence is based on counterfactual reasoning. During training, each agent observes the changes in the actions of other agents by replacing its current action, thereby calculating the degree of adherence of other agents to its behavior. Using adherence as an intrinsic reward enables agents to consider the collective, thus promoting cooperation. In addition, the adherence rewards of all agents are calculated in a decentralized way. We experiment in sequential social dilemma environments, and the results demonstrate the potential for the algorithm to enhance cooperation and coordination and significantly increase the scores of the deep RL agents.https://www.mdpi.com/2076-3417/12/16/8004multi-agent reinforcement learningmulti-agent systemintrinsic rewardcounterfactual reasoningsequential social dilemmas |
spellingShingle | Yuyu Yuan Ting Guo Pengqian Zhao Hongpu Jiang Adherence Improves Cooperation in Sequential Social Dilemmas Applied Sciences multi-agent reinforcement learning multi-agent system intrinsic reward counterfactual reasoning sequential social dilemmas |
title | Adherence Improves Cooperation in Sequential Social Dilemmas |
title_full | Adherence Improves Cooperation in Sequential Social Dilemmas |
title_fullStr | Adherence Improves Cooperation in Sequential Social Dilemmas |
title_full_unstemmed | Adherence Improves Cooperation in Sequential Social Dilemmas |
title_short | Adherence Improves Cooperation in Sequential Social Dilemmas |
title_sort | adherence improves cooperation in sequential social dilemmas |
topic | multi-agent reinforcement learning multi-agent system intrinsic reward counterfactual reasoning sequential social dilemmas |
url | https://www.mdpi.com/2076-3417/12/16/8004 |
work_keys_str_mv | AT yuyuyuan adherenceimprovescooperationinsequentialsocialdilemmas AT tingguo adherenceimprovescooperationinsequentialsocialdilemmas AT pengqianzhao adherenceimprovescooperationinsequentialsocialdilemmas AT hongpujiang adherenceimprovescooperationinsequentialsocialdilemmas |