Results for 'Multi-agent reinforcement learning'

1000+ found
Order:
  1. Multi-Agent Reinforcement Learning: Weighting and Partitioning.Ron Sun & Todd Peterson - unknown
    This paper addresses weighting and partitioning in complex reinforcement learning tasks, with the aim of facilitating learning. The paper presents some ideas regarding weighting of multiple agents and extends them into partitioning an input/state space into multiple regions with di erential weighting in these regions, to exploit di erential characteristics of regions and di erential characteristics of agents to reduce the learning complexity of agents (and their function approximators) and thus to facilitate the learning overall. (...)
     
    Export citation  
     
    Bookmark   6 citations  
  2.  5
    Multi-agent reinforcement learning based algorithm detection of malware-infected nodes in IoT networks.Marcos Severt, Roberto Casado-Vara, Ángel Martín del Rey, Héctor Quintián & Jose Luis Calvo-Rolle - forthcoming - Logic Journal of the IGPL.
    The Internet of Things (IoT) is a fast-growing technology that connects everyday devices to the Internet, enabling wireless, low-consumption and low-cost communication and data exchange. IoT has revolutionized the way devices interact with each other and the internet. The more devices become connected, the greater the risk of security breaches. There is currently a need for new approaches to algorithms that can detect malware regardless of the size of the network and that can adapt to dynamic changes in the network. (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  3.  42
    Safe multi-agent reinforcement learning for multi-robot control.Shangding Gu, Jakub Grudzien Kuba, Yuanpei Chen, Yali Du, Long Yang, Alois Knoll & Yaodong Yang - 2023 - Artificial Intelligence 319 (C):103905.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  4. Automatic Partitioning for Multi-Agent Reinforcement Learning.Ron Sun - unknown
    This paper addresses automatic partitioning in complex reinforcement learning tasks with multiple agents, without a priori domain knowledge regarding task structures. Partitioning a state/input space into multiple regions helps to exploit the di erential characteristics of regions and di erential characteristics of agents, thus facilitating learning and reducing the complexity of agents especially when function approximators are used. We develop a method for optimizing the partitioning of the space through experience without the use of a priori domain (...)
     
    Export citation  
     
    Bookmark  
  5. Bidding in Reinforcement Learning: A Paradigm for Multi-Agent Systems.Chad Sessions - unknown
    The paper presents an approach for developing multi-agent reinforcement learning systems that are made up of a coalition of modular agents. We focus on learning to segment sequences (sequential decision tasks) to create modular structures, through a bidding process that is based on reinforcements received during task execution. The approach segments sequences (and divides them up among agents) to facilitate the learning of the overall task. Notably, our approach does not rely on a priori (...)
     
    Export citation  
     
    Bookmark   1 citation  
  6.  31
    Instilling moral value alignment by means of multi-objective reinforcement learning.Juan Antonio Rodriguez-Aguilar, Maite Lopez-Sanchez, Marc Serramia & Manel Rodriguez-Soto - 2022 - Ethics and Information Technology 24 (1).
    AI research is being challenged with ensuring that autonomous agents learn to behave ethically, namely in alignment with moral values. Here, we propose a novel way of tackling the value alignment problem as a two-step process. The first step consists on formalising moral values and value aligned behaviour based on philosophical foundations. Our formalisation is compatible with the framework of (Multi-Objective) Reinforcement Learning, to ease the handling of an agent’s individual and ethical objectives. The second step (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  7.  22
    Instilling moral value alignment by means of multi-objective reinforcement learning.M. Rodriguez-Soto, M. Serramia, M. Lopez-Sanchez & J. Antonio Rodriguez-Aguilar - 2022 - Ethics and Information Technology 24 (9).
    AI research is being challenged with ensuring that autonomous agents learn to behave ethically, namely in alignment with moral values. Here, we propose a novel way of tackling the value alignment problem as a two-step process. The first step consists on formalising moral values and value aligned behaviour based on philosophical foundations. Our formalisation is compatible with the framework of (Multi-Objective) Reinforcement Learning, to ease the handling of an agent’s individual and ethical objectives. The second step (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  8. Integrating reinforcement learning, bidding and genetic algorithms.Ron Sun - unknown
    This paper presents a GA-based multi-agent reinforce- ment learning bidding approach (GMARLB) for perform- ing multi-agent reinforcement learning. GMARLB inte- grates reinforcement learning, bidding and genetic algo- rithms. The general idea of our multi-agent systems is as follows: There are a number of individual agents in a team, each agent of the team has two modules: Q module and CQ module. Each agent can select actions to be (...)
     
    Export citation  
     
    Bookmark  
  9.  54
    Too Many Cooks: Bayesian Inference for Coordinating MultiAgent Collaboration.Sarah A. Wu, Rose E. Wang, James A. Evans, Joshua B. Tenenbaum, David C. Parkes & Max Kleiman-Weiner - 2021 - Topics in Cognitive Science 13 (2):414-432.
    Collaboration requires agents to coordinate their behavior on the fly, sometimes cooperating to solve a single task together and other times dividing it up into sub‐tasks to work on in parallel. Underlying the human ability to collaborate is theory‐of‐mind (ToM), the ability to infer the hidden mental states that drive others to act. Here, we develop Bayesian Delegation, a decentralized multiagent learning mechanism with these abilities. Bayesian Delegation enables agents to rapidly infer the hidden intentions of others (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   4 citations  
  10.  81
    An evolutionary game theoretic perspective on learning in multi-agent systems.Karl Tuyls, Ann Nowe, Tom Lenaerts & Bernard Manderick - 2004 - Synthese 139 (2):297 - 330.
    In this paper we revise Reinforcement Learning and adaptiveness in Multi-Agent Systems from an Evolutionary Game Theoretic perspective. More precisely we show there is a triangular relation between the fields of Multi-Agent Systems, Reinforcement Learning and Evolutionary Game Theory. We illustrate how these new insights can contribute to a better understanding of learning in MAS and to new improved learning algorithms. All three fields are introduced in a self-contained manner. Each (...)
    Direct download (5 more)  
     
    Export citation  
     
    Bookmark  
  11.  71
    Learning with neighbours: Emergence of convention in a society of learning agents.Roland Mühlenbernd - 2011 - Synthese 183 (S1):87-109.
    I present a game-theoretical multi-agent system to simulate the evolutionary process responsible for the pragmatic phenomenon division of pragmatic labour (DOPL), a linguistic convention emerging from evolutionary forces. Each agent is positioned on a toroid lattice and communicates via signaling games , where the choice of an interlocutor depends on the Manhattan distance between them. In this framework I compare two learning dynamics: reinforcement learning (RL) and belief learning (BL). An agent’s experiences (...)
    Direct download (5 more)  
     
    Export citation  
     
    Bookmark   8 citations  
  12. Universal Agent Mixtures and the Geometry of Intelligence.Samuel Allen Alexander, David Quarel, Len Du & Marcus Hutter - 2023 - Aistats.
    Inspired by recent progress in multi-agent Reinforcement Learning (RL), in this work we examine the collective intelligent behaviour of theoretical universal agents by introducing a weighted mixture operation. Given a weighted set of agents, their weighted mixture is a new agent whose expected total reward in any environment is the corresponding weighted average of the original agents' expected total rewards in that environment. Thus, if RL agent intelligence is quantified in terms of performance across (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  13.  36
    The Puzzle of Evaluating Moral Cognition in Artificial Agents.Madeline G. Reinecke, Yiran Mao, Markus Kunesch, Edgar A. Duéñez-Guzmán, Julia Haas & Joel Z. Leibo - 2023 - Cognitive Science 47 (8):e13315.
    In developing artificial intelligence (AI), researchers often benchmark against human performance as a measure of progress. Is this kind of comparison possible for moral cognition? Given that human moral judgment often hinges on intangible properties like “intention” which may have no natural analog in artificial agents, it may prove difficult to design a “like‐for‐like” comparison between the moral behavior of artificial and human agents. What would a measure of moral behavior for both humans and AI look like? We unravel the (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  14.  17
    強化学習エージェントへの階層化意志決定法の導入―追跡問題を例に―.輿石 尚宏 謙吾 片山 - 2004 - Transactions of the Japanese Society for Artificial Intelligence 19:279-291.
    Reinforcement Learning is a promising technique for creating agents that can be applied to real world problems. The most important features of RL are trial-and-error search and delayed reward. Thus, agents randomly act in the early learning stage. However, such random actions are impractical for real world problems. This paper presents a novel model of RL agents. A feature of our learning agent model is to integrate the Analytic Hierarchy Process into the standard RL (...) model, which consists of three modules: state recognition, learning, and action selecting modules. In our model, the AHP module is designed with {\\it primary knowledge} that humans intrinsically have in a process until a goal state is attained. This integration aims at increasing promising actions instead of completely random actions in the standard RL algorithms. Profit Sharing is adopted as a RL method for our model, since PS is known to be useful even in multi-agent environments. To evaluate our approach in a multi-agent environment, we test a PS RL method with our agent model on a pursuit problem in a grid world. Computational results show that our approach outperforms the standard PS in terms of learning speed in the earlier stages of learning. We also show that the learning performance of our approach is superior at least competitive to that of the standard one in the final stages of learning. (shrink)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  15. HCI Model with Learning Mechanism for Cooperative Design in Pervasive Computing Environment.Hong Liu, Bin Hu & Philip Moore - 2015 - Journal of Internet Technology 16.
    This paper presents a human-computer interaction model with a three layers learning mechanism in a pervasive environment. We begin with a discussion around a number of important issues related to human-computer interaction followed by a description of the architecture for a multi-agent cooperative design system for pervasive computing environment. We present our proposed three- layer HCI model and introduce the group formation algorithm, which is predicated on a dynamic sharing niche technology. Finally, we explore the cooperative (...) learning and fusion algorithms; the paper closes with concluding observations and a summary of the principal work and contributions of this paper. (shrink)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  16.  14
    環境状況に応じて自己の報酬を操作する学習エージェントの構築.沼尾 正行 森山 甲一 - 2002 - Transactions of the Japanese Society for Artificial Intelligence 17:676-683.
    The authors aim at constructing an agent which learns appropriate actions in a Multi-Agent environment with and without social dilemmas. For this aim, the agent must have nonrationality that makes it give up its own profit when it should do that. Since there are many studies on rational learning that brings more and more profit, it is desirable to utilize them for constructing the agent. Therefore, we use a reward-handling manner that makes internal evaluation (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  17.  14
    If multi-agent learning is the answer, what is the question?Yoav Shoham, Rob Powers & Trond Grenager - 2007 - Artificial Intelligence 171 (7):365-377.
  18.  45
    Reinforcement learning and artificial agency.Patrick Butlin - 2024 - Mind and Language 39 (1):22-38.
    There is an apparent connection between reinforcement learning and agency. Artificial entities controlled by reinforcement learning algorithms are standardly referred to as agents, and the mainstream view in the psychology and neuroscience of agency is that humans and other animals are reinforcement learners. This article examines this connection, focusing on artificial reinforcement learning systems and assuming that there are various forms of agency. Artificial reinforcement learning systems satisfy plausible conditions for minimal (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  19. Q(st at):= (I — o')Q(st at) + o'(r(st+1).Ron Sun - unknown
    Straightforward reinforcement learning for multi-agent co-learning settings often results in poor outcomes. Meta-learning processes beyond straightforward reinforcement learning may be necessary to achieve good (or optimal) outcomes. Algorithmic processes of meta-learning, or "manipulation", will be described, which is a cognitively realistic and effective means for learning cooperation. We will discuss various "manipulation" routines that address the issue of improving multi-agent co-learning. We hope to develop better adaptive means (...)
     
    Export citation  
     
    Bookmark  
  20.  6
    Interestingness elements for explainable reinforcement learning: Understanding agents' capabilities and limitations.Pedro Sequeira & Melinda Gervasio - 2020 - Artificial Intelligence 288:103367.
  21.  7
    Multi-agent learning and the descriptive value of simple models.Ido Erev & Alvin E. Roth - 2007 - Artificial Intelligence 171 (7):423-428.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  22.  8
    Multi-agent learning for engineers.Shie Mannor & Jeff S. Shamma - 2007 - Artificial Intelligence 171 (7):417-422.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  23.  7
    Multi-robot inverse reinforcement learning under occlusion with estimation of state transitions.Kenneth Bogert & Prashant Doshi - 2018 - Artificial Intelligence 263 (C):46-73.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  24.  20
    Counterfactual state explanations for reinforcement learning agents via generative deep learning.Matthew L. Olson, Roli Khanna, Lawrence Neal, Fuxin Li & Weng-Keen Wong - 2021 - Artificial Intelligence 295 (C):103455.
  25.  50
    Reinforcement learning: A brief guide for philosophers of mind.Julia Haas - 2022 - Philosophy Compass 17 (9):e12865.
    In this opinionated review, I draw attention to some of the contributions reinforcement learning can make to questions in the philosophy of mind. In particular, I highlight reinforcement learning's foundational emphasis on the role of reward in agent learning, and canvass two ways in which the framework may advance our understanding of perception and motivation.
    No categories
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  26.  23
    Toward a Psychology of Deep Reinforcement Learning Agents Using a Cognitive Architecture.Konstantinos Mitsopoulos, Sterling Somers, Joel Schooler, Christian Lebiere, Peter Pirolli & Robert Thomson - 2022 - Topics in Cognitive Science 14 (4):756-779.
    We argue that cognitive models can provide a common ground between human users and deep reinforcement learning (Deep RL) algorithms for purposes of explainable artificial intelligence (AI). Casting both the human and learner as cognitive models provides common mechanisms to compare and understand their underlying decision-making processes. This common grounding allows us to identify divergences and explain the learner's behavior in human understandable terms. We present novel salience techniques that highlight the most relevant features in each model's decision-making, (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark  
  27. Using Reinforcement Learning to Examine Dynamic Attention Allocation During Reading.Yanping Liu, Erik D. Reichle & Ding-Guo Gao - 2013 - Cognitive Science 37 (8):1507-1540.
    A fundamental question in reading research concerns whether attention is allocated strictly serially, supporting lexical processing of one word at a time, or in parallel, supporting concurrent lexical processing of two or more words (Reichle, Liversedge, Pollatsek, & Rayner, 2009). The origins of this debate are reviewed. We then report three simulations to address this question using artificial reading agents (Liu & Reichle, 2010; Reichle & Laurent, 2006) that learn to dynamically allocate attention to 1–4 words to “read” as efficiently (...)
    Direct download  
     
    Export citation  
     
    Bookmark   1 citation  
  28.  6
    Reinforcement Learning with Probabilistic Boolean Network Models of Smart Grid Devices.Pedro Juan Rivera Torres, Carlos Gershenson García, María Fernanda Sánchez Puig & Samir Kanaan Izquierdo - 2022 - Complexity 2022:1-15.
    The area of smart power grids needs to constantly improve its efficiency and resilience, to provide high quality electrical power in a resilient grid, while managing faults and avoiding failures. Achieving this requires high component reliability, adequate maintenance, and a studied failure occurrence. Correct system operation involves those activities and novel methodologies to detect, classify, and isolate faults and failures and model and simulate processes with predictive algorithms and analytics. In this paper, we showcase the application of a complex-adaptive, self-organizing (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  29.  6
    Agendas for multi-agent learning.Geoffrey J. Gordon - 2007 - Artificial Intelligence 171 (7):392-401.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  30.  4
    Foundations of multi-agent learning: Introduction to the special issue.Rakesh V. Vohra & Michael P. Wellman - 2007 - Artificial Intelligence 171 (7):363-364.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  31. Individual action and collective function: From sociology to multi-agent learning.Ron Sun - manuscript
    Co-learning of multiple agents has been studied in co-learning settings, and how do they help, or many different disciplines under various guises. For hamper, learning and cooperation? example, the issue has been tackled by distributed • How do we characterize the process and the artificial intelligence, parallel and distributed com- dynamics of co-learning, conceptually, mathe- puting, cognitive psychology, social psychology, matically, or computationally? game theory (and other areas of mathematical econ- • how do social structures and (...)
     
    Export citation  
     
    Bookmark   3 citations  
  32.  19
    Exploration-exploitation in multi-agent learning: Catastrophe theory meets game theory.Stefanos Leonardos & Georgios Piliouras - 2022 - Artificial Intelligence 304 (C):103653.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  33.  53
    Cognitive science meets multi-agent systems: A prolegomenon.Ron Sun - 2001 - Philosophical Psychology 14 (1):5 – 28.
    In the current research on multi-agent systems (MAS), many theoretical issues related to sociocultural processes have been touched upon. These issues are in fact intellectually profound and should prove to be significant for MAS. Moreover, these issues should have equally significant impact on cognitive science, if we ever try to understand cognition in the broad context of sociocultural environments in which cognitive agents exist. Furthermore, cognitive models as studied in cognitive science can help us in a substantial way (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   15 citations  
  34.  16
    Enforcing ethical goals over reinforcement-learning policies.Guido Governatori, Agata Ciabattoni, Ezio Bartocci & Emery A. Neufeld - 2022 - Ethics and Information Technology 24 (4):1-19.
    Recent years have yielded many discussions on how to endow autonomous agents with the ability to make ethical decisions, and the need for explicit ethical reasoning and transparency is a persistent theme in this literature. We present a modular and transparent approach to equip autonomous agents with the ability to comply with ethical prescriptions, while still enacting pre-learned optimal behaviour. Our approach relies on a normative supervisor module, that integrates a theorem prover for defeasible deontic logic within the control loop (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  35. The Archimedean trap: Why traditional reinforcement learning will probably not yield AGI.Samuel Allen Alexander - 2020 - Journal of Artificial General Intelligence 11 (1):70-85.
    After generalizing the Archimedean property of real numbers in such a way as to make it adaptable to non-numeric structures, we demonstrate that the real numbers cannot be used to accurately measure non-Archimedean structures. We argue that, since an agent with Artificial General Intelligence (AGI) should have no problem engaging in tasks that inherently involve non-Archimedean rewards, and since traditional reinforcement learning rewards are real numbers, therefore traditional reinforcement learning probably will not lead to AGI. (...)
    Direct download  
     
    Export citation  
     
    Bookmark   1 citation  
  36.  19
    The change of signaling conventions in social networks.Roland Mühlenbernd - 2019 - AI and Society 34 (4):721-734.
    To depict the mechanisms that have enabled the emergence of semantic conventions, philosophers and researchers particularly access a game-theoretic model: the signaling game. In this article I argue that this model is also quite appropriate to analyze not only the emergence of a semantic convention, but also its change. I delineate how the application of signaling games helps to reproduce and depict mechanisms of semantic change. For that purpose I present a model that combines a signaling game with innovative (...) learning; in simulation runs I conduct this game repeatedly within a multi-agent setup, where agents are arranged in social network structures. The results of these runs are contrasted with an attested theory from sociolinguistics: the ‘weak tie’ theory. Analyses of the produced data target a deeper understanding of the role of environmental variables for the promotion of semantic change or solidity of semantic conventions. (shrink)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  37.  20
    Reward tampering problems and solutions in reinforcement learning: a causal influence diagram perspective.Tom Everitt, Marcus Hutter, Ramana Kumar & Victoria Krakovna - 2021 - Synthese 198 (Suppl 27):6435-6467.
    Can humans get arbitrarily capable reinforcement learning agents to do their bidding? Or will sufficiently capable RL agents always find ways to bypass their intended objectives by shortcutting their reward signal? This question impacts how far RL can be scaled, and whether alternative paradigms must be developed in order to build safe artificial general intelligence. In this paper, we study when an RL agent has an instrumental goal to tamper with its reward process, and describe design principles (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   4 citations  
  38.  17
    Pessimistic value iteration for multi-task data sharing in Offline Reinforcement Learning.Chenjia Bai, Lingxiao Wang, Jianye Hao, Zhuoran Yang, Bin Zhao, Zhen Wang & Xuelong Li - 2024 - Artificial Intelligence 326 (C):104048.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  39.  5
    An economist's perspective on multi-agent learning.Drew Fudenberg & David K. Levine - 2007 - Artificial Intelligence 171 (7):378-381.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  40.  9
    Learning, detection and representation of multi-agent events in videos.Asaad Hakeem & Mubarak Shah - 2007 - Artificial Intelligence 171 (8-9):586-605.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  41.  43
    A model for updates in a multi-agent setting.John Cantwell - 2007 - Journal of Applied Non-Classical Logics 17 (2):183-196.
    A formal model for updates—the result of learning that the world has changed—in a multi-agent setting is presented and completely axiomatized. The model allows that several agents simultaneously are informed of an event in the world in such a way that it becomes common knowledge among the agents that the event has occurred. The model shares many features with the model for common announcements—an announcement about the state of the world in which it becomes common knowledge among (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  42. Online adaptation of computer games agents: A reinforcement learning approach.Gustavo Andrade, Hugo Santana, André Furtado, Arga Leitão & Geber Ramalho - 2004 - Scientia 15 (2).
    No categories
     
    Export citation  
     
    Bookmark  
  43.  32
    Does Feedback-Related Brain Response during Reinforcement Learning Predict Socio-motivational (In-)dependence in Adolescence?Diana Raufelder, Rebecca Boehme, Lydia Romund, Sabrina Golde, Robert C. Lorenz, Tobias Gleich & Anne Beck - 2016 - Frontiers in Psychology 7:190427.
    This multi-methodological study applied functional magnetic resonance imaging to investigate neural activation in a group of adolescent students ( N = 88) during a probabilistic reinforcement learning task. We related patterns of emerging brain activity and individual learning rates to socio-motivational (in-)dependence manifested in four different motivation types (MTs): (1) peer-dependent MT, (2) teacher-dependent MT, (3) peer-and-teacher-dependent MT, (4) peer-and-teacher-independent MT. A multinomial regression analysis revealed that the individual learning rate predicts students’ membership to the (...)
    Direct download (6 more)  
     
    Export citation  
     
    Bookmark   4 citations  
  44.  4
    The possible and the impossible in multi-agent learning.H. Peyton Young - 2007 - Artificial Intelligence 171 (7):429-433.
  45.  12
    Cognitive prediction of obstacle's movement for reinforcement learning pedestrian interacting model.Masaomi Kimura & Thanh-Trung Trinh - 2022 - Journal of Intelligent Systems 31 (1):127-147.
    Recent studies in pedestrian simulation have been able to construct a highly realistic navigation behaviour in many circumstances. However, when replicating the close interactions between pedestrians, the replicated behaviour is often unnatural and lacks human likeness. One of the possible reasons is that the current models often ignore the cognitive factors in the human thinking process. Another reason is that many models try to approach the problem by optimising certain objectives. On the other hand, in real life, humans do not (...)
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  46.  29
    Bayesian learning for cooperation in multi-agent systems.Mair Allen-Williams & Nicholas R. Jennings - 2009 - In L. Magnani (ed.), computational intelligence. pp. 321--360.
    Direct download  
     
    Export citation  
     
    Bookmark  
  47.  14
    A Stable Distributed Neural Controller for Physically Coupled Networked Discrete-Time System via Online Reinforcement Learning.Jian Sun & Jie Li - 2018 - Complexity 2018:1-15.
    The large scale, time varying, and diversification of physically coupled networked infrastructures such as power grid and transportation system lead to the complexity of their controller design, implementation, and expansion. For tackling these challenges, we suggest an online distributed reinforcement learning control algorithm with the one-layer neural network for each subsystem or called agents to adapt the variation of the networked infrastructures. Each controller includes a critic network and action network for approximating strategy utility function and desired control (...)
    No categories
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark  
  48. Artificial virtuous agents in a multi-agent tragedy of the commons.Jakob Stenseke - 2022 - AI and Society:1-18.
    Although virtue ethics has repeatedly been proposed as a suitable framework for the development of artificial moral agents, it has been proven difficult to approach from a computational perspective. In this work, we present the first technical implementation of artificial virtuous agents in moral simulations. First, we review previous conceptual and technical work in artificial virtue ethics and describe a functionalistic path to AVAs based on dispositional virtues, bottom-up learning, and top-down eudaimonic reward. We then provide the details of (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  49.  15
    Learning agents that acquire representations of social groups.Joel Z. Leibo, Alexander Sasha Vezhnevets, Maria K. Eckstein, John P. Agapiou & Edgar A. Duéñez-Guzmán - 2022 - Behavioral and Brain Sciences 45.
    Humans are learning agents that acquire social group representations from experience. Here, we discuss how to construct artificial agents capable of this feat. One approach, based on deep reinforcement learning, allows the necessary representations to self-organize. This minimizes the need for hand-engineering, improving robustness and scalability. It also enables “virtual neuroscience” research on the learned representations.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  50.  46
    Probabilistic rule-based argumentation for norm-governed learning agents.Régis Riveret, Antonino Rotolo & Giovanni Sartor - 2012 - Artificial Intelligence and Law 20 (4):383-420.
    This paper proposes an approach to investigate norm-governed learning agents which combines a logic-based formalism with an equation-based counterpart. This dual formalism enables us to describe the reasoning of such agents and their interactions using argumentation, and, at the same time, to capture systemic features using equations. The approach is applied to norm emergence and internalisation in systems of learning agents. The logical formalism is rooted into a probabilistic defeasible logic instantiating Dung’s argumentation framework. Rules of this logic (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   5 citations  
1 — 50 / 1000