Abstract
Agents that perform intelligent tasks interacting with humans in a seamless manner are becoming a reality. In contexts in which interactions among agents repeat over time, they might evolve from a cooperative to a competitive attitude, and vice versa, depending on environmental factors and other contextual circumstances. We provide a framework to model transitions between competition and cooperation in a community of agents. Competition is dealt with through the paradigm of adversarial risk analysis, which provides a disagreement solution; implicitly, we minimize the distance to such solution. Cooperation is handled through a concept of maximal separation from the disagreement solution. Mixtures of both problems are used to refer to in-between behaviour. We illustrate the ideas with several simulations in relation with a group of robots. Our motivation is the constitution of communities of robotic agents that interact among them and with one or more users.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
By this, we understand \(x_i \ge d_i\), \(\forall i = 1,\ldots ,n\); we shall also use \(x > d\) meaning \(x_i > d_i\), \(\forall i = 1,\ldots ,n\).
Since \(w_i^1\) and \(w_i^2\) are complementary, we could use just one of them. However, we preserve both for clarity purposes.
References
AiSoy1 KiK. AiSoy Robotics. Social robot. AiSoy Robotics. http://www.aisoy.es. Accessed 30 October 2018.
Gelman, A., Carlin, J. B., Stern, H. S., Dunson, D. B., Vehtari, A., & Rubin, D. B. (1995). Bayesian data analysis (2013th ed.). New York: CRC Press.
Busemeyer, J. R., Dimperio, E., & Jessup, R. K. (2007). Integrating emotional processes into decision-making models. In W. D. Gray (Ed.), Integrated models of cognitive systems (pp. 213–229). New York: Oxford University Press.
Chesbrough, H. W. (2003). The era of open innovation. MIT Sloan Management Review, 44(3), 35–41.
Clemen, R. T., & Reilly, T. (2004). Making hard decisions with decisiontools (2013th ed.). Mason: Cengage Learning.
Dyer, J. S., & Sarin, R. K. (1979). Measurable multiattribute value functions. Operations Research, 27(4), 810–822.
Esteban, P. G., & Ríos Insua, D. (2014). Supporting an autonomous social agent within a competitive environment. Cybernetics and Systems, 45(3), 241–253.
Esteban, P. G., & Ríos Insua, D. (2015). Designing societies of robots. In T. V. Guy, M. Karny, & D. H. Wolpert (Eds.), Decision making: Uncertainty, imperfection, deliberation and scalability (pp. 33–53). Cham: Springer.
Fong, T., Nourbakhsh, I., & Dautenhahn, K. (2003). A survey of socially interactive robots. Robotics and Autonomous Systems, 42(3–4), 143–166.
Gerber, A. (2005). Reference functions and solutions to bargaining problems with and without claims. Social Choice and Welfare, 24(3), 527–541.
Gibbons, R. (1992). Game theory for applied economists (1992nd ed.). Princeton: Princeton University Press.
González-Ortega, J., Radovic, V., & Ríos Insua, D. (2018). Utility elicitation. In L. C. Dias, A. Morton, & J. Quigley (Eds.), Elicitation: The science and art of structuring judgement (pp. 241–264). Cham: Springer.
Hargreaves-Heap, S. P., & Varoufakis, Y. (1995). Game theory: A critical introduction (2004th ed.). New York: Routledge.
Harrington, J. E. (2009). Games, strategies and decision making (2014th ed.). New York: Macmillan Publishers.
Harsanyi, J. C. (1982). Comment subjective probability and the theory of games: Comments on kadane and larkey’s paper. Management Science, 29(2), 120–124.
Iocchi, L., Holz, D., Ruiz-del-Solar, J., Sugiura, K., & van der Zant, T. (2015). RoboCup@Home: Analysis and results of evolving competitions for domestic and service robots. Artificial Intelligence, 229(1), 258–281.
Jefferies, N., Mitchell, C., & Walker, M. (1996). A proposed architecture for trusted third party services. In E. Dawson & J. Golić (Eds.), Cryptography: Policy and algorithms (pp. 98–104). Berlin: Springer.
Kadane, J. B., & Larkey, P. D. (1982). Subjective probability and the theory of games. Management Science, 28(2), 113–120.
Levy, H. (1998). Stochastic dominance: Investment decision making under uncertainty (2016th ed.). Cham: Springer.
Lin, P., Abney, K., & Bekey, G. (2011). Robot ethics: Mapping the issues for a mechanized world. Artificial Intelligence, 175(5–6), 942–949.
Lippman, S. A., & McCardle, K. F. (2012). Embedded Nash bargaining: Risk aversion and impatience. Decision Analysis, 9(1), 31–40.
Maschler, M., Solan, E., & Zamir, S. (2013). Game theory. New York: Cambridge University Press.
Menache, I., & Ozdaglar, A. (2011). Network games: Theory, models, and dynamics. Synthesis Lectures on Communication Networks, 4(1), 1–159.
Myerson, R. B. (1991). Game theory: Analysis of conflict (1997th ed.). Cambridge: Harvard University Press.
Nash, J. F. (1950). The bargaining problem. Econometrica, 18(2), 155–162.
Nash, J. F. (1953). Two-person cooperative games. Econometrica, 21(1), 128–140.
Nisan, N., Roughgarden, T., Tardos, É., & Vazirani, V. V. (2007). Algorithmic game theory (2007th ed.). New York: Cambridge University Press.
Pfingsten, A., & Wagener, A. (2003). Bargaining solutions as social compromises. Theory and Decision, 55(4), 359–389.
Raiffa, H. (1982). The art and science of negotiation (2003rd ed.). Cambridge: Harvard University Press.
Raiffa, H., Richardson, J., & Metcalfe, D. (2002). Negotiation analysis: The science and art of collaborative decision making (2002nd ed.). Cambridge: Harvard University Press.
Ray, P. (1973). Independence of irrelevant alternatives. Econometrica, 41(5), 987–991.
Ríos Insua, D., Banks, D. L., & Ríos, J. (2016). Modeling opponents in adversarial risk analysis. Risk Analysis, 36(4), 742–755.
Ríos Insua, D., Ríos, J., & Banks, D. L. (2009). Adversarial risk analysis. Journal of the American Statistical Association, 104(486), 841–854.
Rothkopf, M. H. (2007). Decision analysis: The right tool for auctions. Decision Analysis, 4(3), 167–172.
Stahl, D. O., & Wilson, P. W. (1995). On players’ models of other players: Theory and experimental evidence. Games and Economic Behavior, 10(1), 218–254.
Thomson, W. (1981). Nash’s bargaining solution and utilitarian choice rules. Econometrica, 49(2), 535–538.
Thomson, W. (1994). Cooperative models of bargaining. In R. Aumann & S. Hart (Eds.), Handbook of game theory with economic applications (Vol. 2, pp. 1237–1284). Amsterdam: North-Holland.
Thomson, W. (2010). Bargaining and the theory of cooperative games: John Nash and beyond (2010th ed.). Northampton: Edward Elgar Publishing.
Wu, H. (2007). Finite Bargaining Problems. PhD dissertation. Georgia State University. https://scholarworks.gsu.edu/econ_diss/32. Accessed 30 Oct 2018.
Yu, P. L. (1973). A class of solutions for group decision problems. Management Science, 19(8), 936–946.
Acknowledgements
Research is supported by the Spanish Ministry of Economy and Innovation programs MTM2014-56949-C3-1-R and MTM2017-86875-C3-1-R and the INNPACTO project HAUS. The work of DRI is funded by the AXA-ICMAT Chair on Adversarial Risk Analysis. Besides, JGO’s research is financed by the Spanish Ministry of Economy and Competitiveness under FPI SO grant agreement BES-2015-072892. This work has also been partially supported by the Spanish Ministry of Economy and Competitiveness, through the “Severo Ochoa” Program for Centers of Excellence in R&D (SEV-2015-0554). We are grateful for discussion to Diego García from AiSoy Robotics S.L., Jesus Ríos and David Banks. We are also grateful to numeorus suggestions by the referees.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This research is supported by Singapore Ministry of Education Academic Research Fund Tier 2, MOE2016-T2-2-156.
This is one of the several papers published in Autonomous Robots comprising the Special Issue on Multi-Robot and Multi-Agent Systems.
Experimental setup
Experimental setup
We describe in some detail the setup for our experiments. The chosen robotic platform has been selected for being a low-cost solution affordable for being at home and schools. This robot has several sensors, including a camera to detect objects or persons within a scene; a microphone used to recognize and understand what the user says through an ASR component; touch sensors to interpret when it has been stroked or attacked; an inclination sensor to know whether it is in vertical position; and several actuators, including servos that allow the robot to move some of its parts, although it mostly uses a TTS system and a simple screen to simulate a mouth when talking.
Our social agents will be dynamically aware of their external context which comprises their environment E as well as the actions performed by a user B during interactions. The evolution of environmental conditions due to the agents’ and user’s actions is assumed to be perceptible through the sensors installed in the agents as described above. The agents’ decisions a will be regulated and planned within the environment, which changes with the user’s actions b, made within an action set \({{\mathcal {B}}}\), leading to an environmental state e within a set \({{\mathcal {E}}}\).
The global loop of the robots covers the stages of: (i)sensing and forecasting; and (ii) decision making.
At time t, the agents step into the first stage. They collect signals from their sensors to interpret the environmental conditions and user actions. The forecasting model is used in expected utility calculations to determine optimal decisions. Once the agents perform their actions, the user responds and the environment evolves. As soon as the agents receive the user responses, they assess the actual consequences of all decisions. Then, the control values are adapted and the environmental states updated, with the time mechanism forwarded in last.
We provide now some details of all the relevant elements. The underlying decision making model uses multi-attribute expected utilities with probabilities based on the ARA framework, adapting ideas from Esteban and Ríos Insua (2014, 2015). We refer to the action sets of the agents and the user, respectively, as \({{\mathcal {A}}}_i =\{a_1,\ldots ,a_{15}\}\) and \({{\mathcal {B}}} = \{b_1,\ldots ,b_{14}\}\). The simulated agents may thus perform 15 actions, divided in four groups labeled attention-seeking, complaining, unresponsive and interactive. They may also detect 14 user actions which we have divided as affective, aggressive, interactive, unresponsive and updating, as reflected in Table 6.
At time t, depending on the actions \(a_t\) of the agents, the action \(b_t\) of the user and the environmental state \(e_t \in {{\mathcal {E}}} = \{e_1,\ldots ,e_r\}\), the agents obtain the multi-attribute consequences \(c^i(a_t,b_t,e_t)\), \(i = 1,\ldots ,l\). Specifically, we set \(l = 5\), being the objectives:
- \(u_1\):
Being sufficiently charged.
- \(u_2\):
Remain secure, in relation with the noise, light and temperature conditions surrounding the agent.
- \(u_3\):
Interact with identified users.
- \(u_4\):
Having fun with identified users.
- \(u_5\):
Having the software updated.
The utility function adopts the multi-attribute additive form for the i-th agent
where \(\omega _i^k \ge 0\) and \(\sum _{k=1}^{5} {w_i^k} = 1\). \(w_i^k\) represents the weight of the i-th agent’s k-th objective and \(u_i^k\) represents the corresponding component utility function. We set \(w_1> w_2> w_3> w_4 > w_5\) to stress the hierarchical nature of the objectives.
The agents’ beliefs are regulated within the ARA framework, more specifically within the level-1 thinking approach. Given the past history of the agents’ and user’s actions, environmental states and the agents’ potential action \(a_t\), each agent forecasts the user’s action and environment state through
where we limit memory to two periods for computational reasons. We decompose (9) through
Then, under standard conditions, the agents are designed to choose the action with Maximum Expected Utility (MEU), that is, they will solve
Note that, for computational reasons, we just plan one period ahead. The ideas may be easily extended to longer planning periods through dynamic programming. In such case, a more stable class of weights could be obtained as the average of a few of the last utilities attained.
Rights and permissions
About this article
Cite this article
Gómez Esteban, P., Liu, S., Ríos Insua, D. et al. Competition and cooperation in a community of autonomous agents. Auton Robot 44, 533–546 (2020). https://doi.org/10.1007/s10514-019-09867-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10514-019-09867-y