Skip to main content
Log in

Competition and cooperation in a community of autonomous agents

  • Published:
Autonomous Robots Aims and scope Submit manuscript

Abstract

Agents that perform intelligent tasks interacting with humans in a seamless manner are becoming a reality. In contexts in which interactions among agents repeat over time, they might evolve from a cooperative to a competitive attitude, and vice versa, depending on environmental factors and other contextual circumstances. We provide a framework to model transitions between competition and cooperation in a community of agents. Competition is dealt with through the paradigm of adversarial risk analysis, which provides a disagreement solution; implicitly, we minimize the distance to such solution. Cooperation is handled through a concept of maximal separation from the disagreement solution. Mixtures of both problems are used to refer to in-between behaviour. We illustrate the ideas with several simulations in relation with a group of robots. Our motivation is the constitution of communities of robotic agents that interact among them and with one or more users.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Notes

  1. By this, we understand \(x_i \ge d_i\), \(\forall i = 1,\ldots ,n\); we shall also use \(x > d\) meaning \(x_i > d_i\), \(\forall i = 1,\ldots ,n\).

  2. Since \(w_i^1\) and \(w_i^2\) are complementary, we could use just one of them. However, we preserve both for clarity purposes.

References

  • AiSoy1 KiK. AiSoy Robotics. Social robot. AiSoy Robotics. http://www.aisoy.es. Accessed 30 October 2018.

  • Gelman, A., Carlin, J. B., Stern, H. S., Dunson, D. B., Vehtari, A., & Rubin, D. B. (1995). Bayesian data analysis (2013th ed.). New York: CRC Press.

    Book  Google Scholar 

  • Busemeyer, J. R., Dimperio, E., & Jessup, R. K. (2007). Integrating emotional processes into decision-making models. In W. D. Gray (Ed.), Integrated models of cognitive systems (pp. 213–229). New York: Oxford University Press.

    Chapter  Google Scholar 

  • Chesbrough, H. W. (2003). The era of open innovation. MIT Sloan Management Review, 44(3), 35–41.

    Google Scholar 

  • Clemen, R. T., & Reilly, T. (2004). Making hard decisions with decisiontools (2013th ed.). Mason: Cengage Learning.

    Google Scholar 

  • Dyer, J. S., & Sarin, R. K. (1979). Measurable multiattribute value functions. Operations Research, 27(4), 810–822.

    Article  MathSciNet  Google Scholar 

  • Esteban, P. G., & Ríos Insua, D. (2014). Supporting an autonomous social agent within a competitive environment. Cybernetics and Systems, 45(3), 241–253.

    Article  Google Scholar 

  • Esteban, P. G., & Ríos Insua, D. (2015). Designing societies of robots. In T. V. Guy, M. Karny, & D. H. Wolpert (Eds.), Decision making: Uncertainty, imperfection, deliberation and scalability (pp. 33–53). Cham: Springer.

    Chapter  Google Scholar 

  • Fong, T., Nourbakhsh, I., & Dautenhahn, K. (2003). A survey of socially interactive robots. Robotics and Autonomous Systems, 42(3–4), 143–166.

    Article  Google Scholar 

  • Gerber, A. (2005). Reference functions and solutions to bargaining problems with and without claims. Social Choice and Welfare, 24(3), 527–541.

    Article  MathSciNet  Google Scholar 

  • Gibbons, R. (1992). Game theory for applied economists (1992nd ed.). Princeton: Princeton University Press.

    Book  Google Scholar 

  • González-Ortega, J., Radovic, V., & Ríos Insua, D. (2018). Utility elicitation. In L. C. Dias, A. Morton, & J. Quigley (Eds.), Elicitation: The science and art of structuring judgement (pp. 241–264). Cham: Springer.

    Chapter  Google Scholar 

  • Hargreaves-Heap, S. P., & Varoufakis, Y. (1995). Game theory: A critical introduction (2004th ed.). New York: Routledge.

    Book  Google Scholar 

  • Harrington, J. E. (2009). Games, strategies and decision making (2014th ed.). New York: Macmillan Publishers.

    Google Scholar 

  • Harsanyi, J. C. (1982). Comment subjective probability and the theory of games: Comments on kadane and larkey’s paper. Management Science, 29(2), 120–124.

    Article  Google Scholar 

  • Iocchi, L., Holz, D., Ruiz-del-Solar, J., Sugiura, K., & van der Zant, T. (2015). RoboCup@Home: Analysis and results of evolving competitions for domestic and service robots. Artificial Intelligence, 229(1), 258–281.

    Article  MathSciNet  Google Scholar 

  • Jefferies, N., Mitchell, C., & Walker, M. (1996). A proposed architecture for trusted third party services. In E. Dawson & J. Golić (Eds.), Cryptography: Policy and algorithms (pp. 98–104). Berlin: Springer.

    Chapter  Google Scholar 

  • Kadane, J. B., & Larkey, P. D. (1982). Subjective probability and the theory of games. Management Science, 28(2), 113–120.

    Article  MathSciNet  Google Scholar 

  • Levy, H. (1998). Stochastic dominance: Investment decision making under uncertainty (2016th ed.). Cham: Springer.

    Book  Google Scholar 

  • Lin, P., Abney, K., & Bekey, G. (2011). Robot ethics: Mapping the issues for a mechanized world. Artificial Intelligence, 175(5–6), 942–949.

    Article  Google Scholar 

  • Lippman, S. A., & McCardle, K. F. (2012). Embedded Nash bargaining: Risk aversion and impatience. Decision Analysis, 9(1), 31–40.

    Article  MathSciNet  Google Scholar 

  • Maschler, M., Solan, E., & Zamir, S. (2013). Game theory. New York: Cambridge University Press.

    Book  Google Scholar 

  • Menache, I., & Ozdaglar, A. (2011). Network games: Theory, models, and dynamics. Synthesis Lectures on Communication Networks, 4(1), 1–159.

    Article  Google Scholar 

  • Myerson, R. B. (1991). Game theory: Analysis of conflict (1997th ed.). Cambridge: Harvard University Press.

    MATH  Google Scholar 

  • Nash, J. F. (1950). The bargaining problem. Econometrica, 18(2), 155–162.

    Article  MathSciNet  Google Scholar 

  • Nash, J. F. (1953). Two-person cooperative games. Econometrica, 21(1), 128–140.

    Article  MathSciNet  Google Scholar 

  • Nisan, N., Roughgarden, T., Tardos, É., & Vazirani, V. V. (2007). Algorithmic game theory (2007th ed.). New York: Cambridge University Press.

    Book  Google Scholar 

  • Pfingsten, A., & Wagener, A. (2003). Bargaining solutions as social compromises. Theory and Decision, 55(4), 359–389.

    Article  MathSciNet  Google Scholar 

  • Raiffa, H. (1982). The art and science of negotiation (2003rd ed.). Cambridge: Harvard University Press.

    Google Scholar 

  • Raiffa, H., Richardson, J., & Metcalfe, D. (2002). Negotiation analysis: The science and art of collaborative decision making (2002nd ed.). Cambridge: Harvard University Press.

    Google Scholar 

  • Ray, P. (1973). Independence of irrelevant alternatives. Econometrica, 41(5), 987–991.

    Article  MathSciNet  Google Scholar 

  • Ríos Insua, D., Banks, D. L., & Ríos, J. (2016). Modeling opponents in adversarial risk analysis. Risk Analysis, 36(4), 742–755.

    Article  Google Scholar 

  • Ríos Insua, D., Ríos, J., & Banks, D. L. (2009). Adversarial risk analysis. Journal of the American Statistical Association, 104(486), 841–854.

    Article  MathSciNet  Google Scholar 

  • Rothkopf, M. H. (2007). Decision analysis: The right tool for auctions. Decision Analysis, 4(3), 167–172.

    Article  Google Scholar 

  • Stahl, D. O., & Wilson, P. W. (1995). On players’ models of other players: Theory and experimental evidence. Games and Economic Behavior, 10(1), 218–254.

    Article  MathSciNet  Google Scholar 

  • Thomson, W. (1981). Nash’s bargaining solution and utilitarian choice rules. Econometrica, 49(2), 535–538.

    Article  MathSciNet  Google Scholar 

  • Thomson, W. (1994). Cooperative models of bargaining. In R. Aumann & S. Hart (Eds.), Handbook of game theory with economic applications (Vol. 2, pp. 1237–1284). Amsterdam: North-Holland.

    Chapter  Google Scholar 

  • Thomson, W. (2010). Bargaining and the theory of cooperative games: John Nash and beyond (2010th ed.). Northampton: Edward Elgar Publishing.

    Book  Google Scholar 

  • Wu, H. (2007). Finite Bargaining Problems. PhD dissertation. Georgia State University. https://scholarworks.gsu.edu/econ_diss/32. Accessed 30 Oct 2018.

  • Yu, P. L. (1973). A class of solutions for group decision problems. Management Science, 19(8), 936–946.

    Article  MathSciNet  Google Scholar 

Download references

Acknowledgements

Research is supported by the Spanish Ministry of Economy and Innovation programs MTM2014-56949-C3-1-R and MTM2017-86875-C3-1-R and the INNPACTO project HAUS. The work of DRI is funded by the AXA-ICMAT Chair on Adversarial Risk Analysis. Besides, JGO’s research is financed by the Spanish Ministry of Economy and Competitiveness under FPI SO grant agreement BES-2015-072892. This work has also been partially supported by the Spanish Ministry of Economy and Competitiveness, through the “Severo Ochoa” Program for Centers of Excellence in R&D (SEV-2015-0554). We are grateful for discussion to Diego García from AiSoy Robotics S.L., Jesus Ríos and David Banks. We are also grateful to numeorus suggestions by the referees.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Pablo Gómez Esteban.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

This research is supported by Singapore Ministry of Education Academic Research Fund Tier 2, MOE2016-T2-2-156.

This is one of the several papers published in Autonomous Robots comprising the Special Issue on Multi-Robot and Multi-Agent Systems.

Experimental setup

Experimental setup

We describe in some detail the setup for our experiments. The chosen robotic platform has been selected for being a low-cost solution affordable for being at home and schools. This robot has several sensors, including a camera to detect objects or persons within a scene; a microphone used to recognize and understand what the user says through an ASR component; touch sensors to interpret when it has been stroked or attacked; an inclination sensor to know whether it is in vertical position; and several actuators, including servos that allow the robot to move some of its parts, although it mostly uses a TTS system and a simple screen to simulate a mouth when talking.

Our social agents will be dynamically aware of their external context which comprises their environment E as well as the actions performed by a user B during interactions. The evolution of environmental conditions due to the agents’ and user’s actions is assumed to be perceptible through the sensors installed in the agents as described above. The agents’ decisions a will be regulated and planned within the environment, which changes with the user’s actions b, made within an action set \({{\mathcal {B}}}\), leading to an environmental state e within a set \({{\mathcal {E}}}\).

The global loop of the robots covers the stages of: (i)sensing and forecasting; and (ii) decision making.

At time t, the agents step into the first stage. They collect signals from their sensors to interpret the environmental conditions and user actions. The forecasting model is used in expected utility calculations to determine optimal decisions. Once the agents perform their actions, the user responds and the environment evolves. As soon as the agents receive the user responses, they assess the actual consequences of all decisions. Then, the control values are adapted and the environmental states updated, with the time mechanism forwarded in last.

We provide now some details of all the relevant elements. The underlying decision making model uses multi-attribute expected utilities with probabilities based on the ARA framework, adapting ideas from Esteban and Ríos Insua (2014, 2015). We refer to the action sets of the agents and the user, respectively, as \({{\mathcal {A}}}_i =\{a_1,\ldots ,a_{15}\}\) and \({{\mathcal {B}}} = \{b_1,\ldots ,b_{14}\}\). The simulated agents may thus perform 15 actions, divided in four groups labeled attention-seeking, complaining, unresponsive and interactive. They may also detect 14 user actions which we have divided as affective, aggressive, interactive, unresponsive and updating, as reflected in Table 6.

At time t, depending on the actions \(a_t\) of the agents, the action \(b_t\) of the user and the environmental state \(e_t \in {{\mathcal {E}}} = \{e_1,\ldots ,e_r\}\), the agents obtain the multi-attribute consequences \(c^i(a_t,b_t,e_t)\), \(i = 1,\ldots ,l\). Specifically, we set \(l = 5\), being the objectives:

\(u_1\):

Being sufficiently charged.

\(u_2\):

Remain secure, in relation with the noise, light and temperature conditions surrounding the agent.

\(u_3\):

Interact with identified users.

\(u_4\):

Having fun with identified users.

\(u_5\):

Having the software updated.

The utility function adopts the multi-attribute additive form for the i-th agent

$$\begin{aligned} u(c^1,\ldots ,c^5) = \sum _{k=1}^{5} {\omega _i^k\,u_i^k}, \end{aligned}$$

where \(\omega _i^k \ge 0\) and \(\sum _{k=1}^{5} {w_i^k} = 1\). \(w_i^k\) represents the weight of the i-th agent’s k-th objective and \(u_i^k\) represents the corresponding component utility function. We set \(w_1> w_2> w_3> w_4 > w_5\) to stress the hierarchical nature of the objectives.

The agents’ beliefs are regulated within the ARA framework, more specifically within the level-1 thinking approach. Given the past history of the agents’ and user’s actions, environmental states and the agents’ potential action \(a_t\), each agent forecasts the user’s action and environment state through

$$\begin{aligned} p(e_t,b_t\,|\,a_t,(e_{t-1},a_{t-1},b_{t-1}),(e_{t-2},a_{t-2},b_{t-2})), \end{aligned}$$
(9)

where we limit memory to two periods for computational reasons. We decompose (9) through

$$\begin{aligned} p(e_t\,|\,b_t,e_{t-1},e_{t-2}) \cdot p(b_t\,|\,a_t,b_{t-1},b_{t-2}). \end{aligned}$$

Then, under standard conditions, the agents are designed to choose the action with Maximum Expected Utility (MEU), that is, they will solve

$$\begin{aligned} \psi _t^*= & {} \max _{a_t \in {{\mathcal {A}}}} \psi (a_t) \\ \displaystyle= & {} \iint u(a_t,b_t,e_t)\,p(e_t\,|\,b_t,e_{t-1},e_{t-2}) \\&\qquad \qquad \displaystyle p(b_t\,|\,a_t,b_{t-1},b_{t-2})\,\text {d}b_t\,\text {d}de_t. \end{aligned}$$

Note that, for computational reasons, we just plan one period ahead. The ideas may be easily extended to longer planning periods through dynamic programming. In such case, a more stable class of weights could be obtained as the average of a few of the last utilities attained.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Gómez Esteban, P., Liu, S., Ríos Insua, D. et al. Competition and cooperation in a community of autonomous agents. Auton Robot 44, 533–546 (2020). https://doi.org/10.1007/s10514-019-09867-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10514-019-09867-y

Keywords

Navigation