Results for 'Cross‐entropy reinforcement learning'

991 found
Order:
  1.  41
    Interrogating Feature Learning Models to Discover Insights Into the Development of Human Expertise in a Real‐Time, Dynamic Decision‐Making Task.Catherine Sibert, Wayne D. Gray & John K. Lindstedt - 2017 - Topics in Cognitive Science 9 (2):374-394.
    Tetris provides a difficult, dynamic task environment within which some people are novices and others, after years of work and practice, become extreme experts. Here we study two core skills; namely, choosing the goal or objective function that will maximize performance and a feature-based analysis of the current game board to determine where to place the currently falling zoid so as to maximize the goal. In Study 1, we build cross-entropy reinforcement learning models to determine whether different goals (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   7 citations  
  2.  39
    Interrogating Feature Learning Models to Discover Insights Into the Development of Human Expertise in a Real‐Time, Dynamic Decision‐Making Task.Catherine Sibert, Wayne D. Gray & John K. Lindstedt - 2016 - Topics in Cognitive Science 8 (4).
    Tetris provides a difficult, dynamic task environment within which some people are novices and others, after years of work and practice, become extreme experts. Here we study two core skills; namely, choosing the goal or objective function that will maximize performance and a feature-based analysis of the current game board to determine where to place the currently falling zoid so as to maximize the goal. In Study 1, we build cross-entropy reinforcement learning models to determine whether different goals (...)
    Direct download  
     
    Export citation  
     
    Bookmark   7 citations  
  3.  25
    Pragmatically Framed Cross-Situational Noun Learning Using Computational Reinforcement Models.Shamima Najnin & Bonny Banerjee - 2018 - Frontiers in Psychology 9.
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  4.  17
    Evolutionary Reinforcement Learning for Adaptively Detecting Database Intrusions.Seul-Gi Choi & Sung-Bae Cho - 2020 - Logic Journal of the IGPL 28 (4):449-460.
    Relational database management system is the most popular database system. It is important to maintain data security from information leakage and data corruption. RDBMS can be attacked by an outsider or an insider. It is difficult to detect an insider attack because its patterns are constantly changing and evolving. In this paper, we propose an adaptive database intrusion detection system that can be resistant to potential insider misuse using evolutionary reinforcement learning, which combines reinforcement learning and (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  5.  13
    Deep Reinforcement Learning for UAV Intelligent Mission Planning.Longfei Yue, Rennong Yang, Ying Zhang, Lixin Yu & Zhuangzhuang Wang - 2022 - Complexity 2022:1-13.
    Rapid and precise air operation mission planning is a key technology in unmanned aerial vehicles autonomous combat in battles. In this paper, an end-to-end UAV intelligent mission planning method based on deep reinforcement learning is proposed to solve the shortcomings of the traditional intelligent optimization algorithm, such as relying on simple, static, low-dimensional scenarios, and poor scalability. Specifically, the suppression of enemy air defense mission planning is described as a sequential decision-making problem and formalized as a Markov decision (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  6.  31
    The Pursuit of Word Meanings.Jon Scott Stevens, Lila R. Gleitman, John C. Trueswell & Charles Yang - 2017 - Cognitive Science 41 (S4):638-676.
    We evaluate here the performance of four models of cross-situational word learning: two global models, which extract and retain multiple referential alternatives from each word occurrence; and two local models, which extract just a single referent from each occurrence. One of these local models, dubbed Pursuit, uses an associative learning mechanism to estimate word-referent probability but pursues and tests the best referent-meaning at any given time. Pursuit is found to perform as well as global models under many conditions (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   11 citations  
  7.  53
    Learning Representations of Animated Motion Sequences—A Neural Model.Georg Layher, Martin A. Giese & Heiko Neumann - 2014 - Topics in Cognitive Science 6 (1):170-182.
    The detection and categorization of animate motions is a crucial task underlying social interaction and perceptual decision making. Neural representations of perceived animate objects are partially located in the primate cortical region STS, which is a region that receives convergent input from intermediate-level form and motion representations. Populations of STS cells exist which are selectively responsive to specific animated motion sequences, such as walkers. It is still unclear how and to what extent form and motion information contribute to the generation (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  8.  35
    Reinforcing robot perception of multi-modal events through repetition and redundancy and repetition and redundancy.Paul Fitzpatrick, Artur Arsenio & Eduardo R. Torres-Jara - 2006 - Interaction Studies. Social Behaviour and Communication in Biological and Artificial Systemsinteraction Studies / Social Behaviour and Communication in Biological and Artificial Systemsinteraction Studies 7 (2):171-196.
    For a robot to be capable of development it must be able to explore its environment and learn from its experiences. It must find opportunities to experience the unfamiliar in ways that reveal properties valid beyond the immediate context. In this paper, we develop a novel method for using the rhythm of everyday actions as a basis for identifying the characteristic appearance and sounds associated with objects, people, and the robot itself. Our approach is to identify and segment groups of (...)
    Direct download (5 more)  
     
    Export citation  
     
    Bookmark  
  9.  19
    Perceived Nexus Between Non-Invigilated Summative Assessment and Mental Health Difficulties: A Cross Sectional Studies.Amanda Graf, Esther Adama, Ebenezer Afrifa-Yamoah & Kwadwo Adusei-Asante - 2023 - Journal of Academic Ethics 21 (4):609-623.
    The COVID-19 pandemic rapidly led to changes in the mode of teaching, learning and assessments in most tertiary institutions worldwide. Notably, non-invigilated summative assessments became predominant. These changes heightened anxiety and depression, especially among individuals with less resilient coping mechanism. We explored the perceptions and experiences of mental health difficulties of students in tertiary education regarding non-invigilated alternative assessments in comparison to invigilated assessments. A pragmatic, mixed method cross sectional design was conducted online via Qualtrics. Thematic analysis of text (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark  
  10.  2
    LSTM vs CNN in real ship trajectory classification.Juan Pedro Llerena, Jesús García & José Manuel Molina - forthcoming - Logic Journal of the IGPL.
    Ship-type identification in a maritime context can be critical to the authorities to control the activities being carried out. Although Automatic Identification Systems has been mandatory for certain vessels, if a vessel does not have them voluntarily or not, it can lead to a whole set of problems, which is why the use of tracking alternatives such as radar is fully complementary for a vessel monitoring systems. However, radars provide positions, but not what they are detecting. Having systems capable of (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  11.  8
    Lights Off, Spot On: Carbon Literacy Training Crossing Boundaries in the Television Industry.Wendy Chapple, Petra Molthan-Hill, Rachel Welton & Michael Hewitt - 2020 - Journal of Business Ethics 162 (4):813-834.
    Proclaimed the “greenest television programme in the world,” the award-winning soap opera Coronation Street is seen as an industry success story. This paper explores how the integration of carbon literacy training led to a widespread transformational change of practice within Coronation Street. Using the theoretical lens of Communities of Practice, this study examines the nature of social learning and the enablers and barriers to change within the organization. Specifically, how boundary spanning practices, objects and people led to the transformation (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  12. IN-cross Entropy Based MAGDM Strategy under Interval Neutrosophic Set Environment.Shyamal Dalapati, Surapati Pramanik, Shariful Alam, Florentin Smarandache & Tapan Kumar Roy - 2017 - Neutrosophic Sets and Systems 18:43-57.
    Cross entropy measure is one of the best way to calculate the divergence of any variable from the priori one variable. We define a new cross entropy measure under interval neutrosophic set environment.
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   5 citations  
  13.  45
    Reinforcement Learning and Counterfactual Reasoning Explain Adaptive Behavior in a Changing Environment.Yunfeng Zhang, Jaehyon Paik & Peter Pirolli - 2015 - Topics in Cognitive Science 7 (2):368-381.
    Animals routinely adapt to changes in the environment in order to survive. Though reinforcement learning may play a role in such adaptation, it is not clear that it is the only mechanism involved, as it is not well suited to producing rapid, relatively immediate changes in strategies in response to environmental changes. This research proposes that counterfactual reasoning might be an additional mechanism that facilitates change detection. An experiment is conducted in which a task state changes over time (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark  
  14. Advances and Applications of DSmT for Information Fusion. Collected Works, Volume 5.Florentin Smarandache - 2023 - Edited by Smarandache Florentin, Dezert Jean & Tchamova Albena.
    This fifth volume on Advances and Applications of DSmT for Information Fusion collects theoretical and applied contributions of researchers working in different fields of applications and in mathematics, and is available in open-access. The collected contributions of this volume have either been published or presented after disseminating the fourth volume in 2015 in international conferences, seminars, workshops and journals, or they are new. The contributions of each part of this volume are chronologically ordered. First Part of this book presents some (...)
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  15.  43
    Reinforcement learning and artificial agency.Patrick Butlin - 2024 - Mind and Language 39 (1):22-38.
    There is an apparent connection between reinforcement learning and agency. Artificial entities controlled by reinforcement learning algorithms are standardly referred to as agents, and the mainstream view in the psychology and neuroscience of agency is that humans and other animals are reinforcement learners. This article examines this connection, focusing on artificial reinforcement learning systems and assuming that there are various forms of agency. Artificial reinforcement learning systems satisfy plausible conditions for minimal (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  16.  22
    Psychology: Contemporary Perspectives.Paul Okami - 2013 - Oxford University Press USA.
    Research-based but highly accessible, this fresh, contemporary, and engaging volume helps students appreciate the science of psychology and understand how its principles apply to their own lives. Key features: Contemporary perspectives and references: giving careful consideration to the field's historical foundations, Psychology: Contemporary Perspectives provides a unique balance of traditional and contemporary perspectives. This approach invites students to develop a modern appraisal of psychology. Current research: the book covers the latest in evolutionary psychology and behavior genetics, ecological and evolutionary theories (...)
    Direct download  
     
    Export citation  
     
    Bookmark   1 citation  
  17.  18
    A Review on Five Recent and Near-Future Developments in Computational Processing of Emotion in the Human Voice.Dagmar M. Schuller & Björn W. Schuller - 2020 - Emotion Review 13 (1):44-50.
    We provide a short review on the recent and near-future developments of computational processing of emotion in the voice, highlighting self-learning of representations moving continuously away from traditional expert-crafted or brute-forced feature representations to end-to-end learning, a movement towards the coupling of analysis and synthesis of emotional voices to foster better mutual understanding, weakly supervised learning at a large scale, transfer learning from related domains such as speech recognition or cross-modal transfer learning, and reinforced (...) through interactive applications at a large scale. For each of these trends, we shortly explain their implications and potential use such as for interpretation in psychological studies and usage in digital health and digital psychology applications. We also discuss further potential development. (shrink)
    Direct download  
     
    Export citation  
     
    Bookmark   2 citations  
  18.  37
    Morals, ethics, and the technology capabilities and limitations of automated and self-driving vehicles.Joshua Siegel & Georgios Pappas - 2023 - AI and Society 38 (1):213-226.
    We motivate the desire for self-driving and explain its potential and limitations, and explore the need for—and potential implementation of—morals, ethics, and other value systems as complementary “capabilities” to the Deep Technologies behind self-driving. We consider how the incorporation of such systems may drive or slow adoption of high automation within vehicles. First, we explore the role for morals, ethics, and other value systems in self-driving through a representative hypothetical dilemma faced by a self-driving car. Through the lens of engineering, (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  19.  22
    How to Make the Most out of Very Little.Charles Yang - 2020 - Topics in Cognitive Science 12 (1):136-152.
    Yang returns to the problem of referential ambiguity, addressed in the opening paper by Gleitman and Trueswell. Using a computational approach, he argues that “big data” approaches to resolving referential ambiguity are destined to fail, because of the inevitable computational explosion needed to keep track of contextual associations present when a word is uttered. Yang tests several computational models, two of which depend on one‐trial learning, as described in Gleitman and Trueswell’s paper. He concludes that such models outperform cross‐situational (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  20.  10
    Reinforcement Learning-Based Collision Avoidance Guidance Algorithm for Fixed-Wing UAVs.Yu Zhao, Jifeng Guo, Chengchao Bai & Hongxing Zheng - 2021 - Complexity 2021:1-12.
    A deep reinforcement learning-based computational guidance method is presented, which is used to identify and resolve the problem of collision avoidance for a variable number of fixed-wing UAVs in limited airspace. The cooperative guidance process is first analyzed for multiple aircraft by formulating flight scenarios using multiagent Markov game theory and solving it by machine learning algorithm. Furthermore, a self-learning framework is established by using the actor-critic model, which is proposed to train collision avoidance decision-making neural (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  21. Reinforcement learning.Chris Jch Watkins & Peter Dayan - 2003 - In L. Nadel (ed.), Encyclopedia of Cognitive Science. Nature Publishing Group.
  22.  10
    Reinforcement Learning for Production‐Based Cognitive Models.Adrian Brasoveanu & Jakub Dotlačil - 2021 - Topics in Cognitive Science 13 (3):467-487.
    We investigate how Reinforcement Learning methods can be used to solve the production selection and production ordering problem in ACT‐R. We focus on four algorithms from the Q learning family, tabular Q and three versions of Deep Q Networks, as well as the ACT‐R utility learning algorithm, which provides a baseline for the Q algorithms. We compare the performance of these five algorithms in a range of lexical decision tasks framed as sequential decision problems.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  23.  3
    Expanding horizons in reinforcement learning for curious exploration and creative planning.Dale Zhou & Aaron M. Bornstein - 2024 - Behavioral and Brain Sciences 47:e118.
    Curiosity and creativity are expressions of the trade-off between leveraging that with which we are familiar or seeking out novelty. Through the computational lens of reinforcement learning, we describe how formulating the value of information seeking and generation via their complementary effects on planning horizons formally captures a range of solutions to striking this balance.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  24.  49
    Reinforcement learning: A brief guide for philosophers of mind.Julia Haas - 2022 - Philosophy Compass 17 (9):e12865.
    In this opinionated review, I draw attention to some of the contributions reinforcement learning can make to questions in the philosophy of mind. In particular, I highlight reinforcement learning's foundational emphasis on the role of reward in agent learning, and canvass two ways in which the framework may advance our understanding of perception and motivation.
    No categories
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  25. Reinforcement learning: A brief guide for philosophers of mind.Julia Haas - 2022 - Philosophy Compass 17 (9):e12865.
    I argue for the role of reinforcement learning in the philosophy of mind. To start, I make several assumptions about the nature of reinforcement learning and its instantiation in minds like ours. I then review some of the contributions of reinforcement learning methods have made across the so-called 'decision sciences.' Finally, I show how principles from reinforcement learning can shape philosophical debates regarding the nature of perception and characterisations of desire.
    Direct download (5 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  26.  42
    Reconciling reinforcement learning models with behavioral extinction and renewal: Implications for addiction, relapse, and problem gambling.A. David Redish, Steve Jensen, Adam Johnson & Zeb Kurth-Nelson - 2007 - Psychological Review 114 (3):784-805.
    No categories
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   21 citations  
  27. Using Reinforcement Learning to Examine Dynamic Attention Allocation During Reading.Yanping Liu, Erik D. Reichle & Ding-Guo Gao - 2013 - Cognitive Science 37 (8):1507-1540.
    A fundamental question in reading research concerns whether attention is allocated strictly serially, supporting lexical processing of one word at a time, or in parallel, supporting concurrent lexical processing of two or more words (Reichle, Liversedge, Pollatsek, & Rayner, 2009). The origins of this debate are reviewed. We then report three simulations to address this question using artificial reading agents (Liu & Reichle, 2010; Reichle & Laurent, 2006) that learn to dynamically allocate attention to 1–4 words to “read” as efficiently (...)
    Direct download  
     
    Export citation  
     
    Bookmark   1 citation  
  28.  18
    Predictive Movements and Human Reinforcement Learning of Sequential Action.Roy Kleijn, George Kachergis & Bernhard Hommel - 2018 - Cognitive Science 42 (S3):783-808.
    Sequential action makes up the bulk of human daily activity, and yet much remains unknown about how people learn such actions. In one motor learning paradigm, the serial reaction time (SRT) task, people are taught a consistent sequence of button presses by cueing them with the next target response. However, the SRT task only records keypress response times to a cued target, and thus it cannot reveal the full time‐course of motion, including predictive movements. This paper describes a mouse (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  29.  28
    Predictive Movements and Human Reinforcement Learning of Sequential Action.Roy de Kleijn, George Kachergis & Bernhard Hommel - 2018 - Cognitive Science 42 (S3):783-808.
    Sequential action makes up the bulk of human daily activity, and yet much remains unknown about how people learn such actions. In one motor learning paradigm, the serial reaction time (SRT) task, people are taught a consistent sequence of button presses by cueing them with the next target response. However, the SRT task only records keypress response times to a cued target, and thus it cannot reveal the full time‐course of motion, including predictive movements. This paper describes a mouse (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  30. Integrating reinforcement learning, bidding and genetic algorithms.Ron Sun - unknown
    This paper presents a GA-based multi-agent reinforce- ment learning bidding approach (GMARLB) for perform- ing multi-agent reinforcement learning. GMARLB inte- grates reinforcement learning, bidding and genetic algo- rithms. The general idea of our multi-agent systems is as follows: There are a number of individual agents in a team, each agent of the team has two modules: Q module and CQ module. Each agent can select actions to be performed at each step, which are done by (...)
     
    Export citation  
     
    Bookmark  
  31.  59
    Competitive Processes in Cross‐Situational Word Learning.Daniel Yurovsky, Chen Yu & Linda B. Smith - 2013 - Cognitive Science 37 (5):891-921.
    Cross-situational word learning, like any statistical learning problem, involves tracking the regularities in the environment. However, the information that learners pick up from these regularities is dependent on their learning mechanism. This article investigates the role of one type of mechanism in statistical word learning: competition. Competitive mechanisms would allow learners to find the signal in noisy input and would help to explain the speed with which learners succeed in statistical learning tasks. Because cross-situational word (...)
    No categories
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   23 citations  
  32.  23
    Using reinforcement learning to understand the emergence of "intelligent" eye-movement behavior during reading.Erik D. Reichle & Patryk A. Laurent - 2006 - Psychological Review 113 (2):390-408.
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   6 citations  
  33.  63
    Novelty and Inductive Generalization in Human Reinforcement Learning.Samuel J. Gershman & Yael Niv - 2015 - Topics in Cognitive Science 7 (3):391-415.
    In reinforcement learning, a decision maker searching for the most rewarding option is often faced with the question: What is the value of an option that has never been tried before? One way to frame this question is as an inductive problem: How can I generalize my previous experience with one set of options to a novel option? We show how hierarchical Bayesian inference can be used to solve this problem, and we describe an equivalence between the Bayesian (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  34.  5
    Reinforcement learning of non-Markov decision processes.Steven D. Whitehead & Long-Ji Lin - 1995 - Artificial Intelligence 73 (1-2):271-306.
  35.  6
    Reinforcement Learning with Probabilistic Boolean Network Models of Smart Grid Devices.Pedro Juan Rivera Torres, Carlos Gershenson García, María Fernanda Sánchez Puig & Samir Kanaan Izquierdo - 2022 - Complexity 2022:1-15.
    The area of smart power grids needs to constantly improve its efficiency and resilience, to provide high quality electrical power in a resilient grid, while managing faults and avoiding failures. Achieving this requires high component reliability, adequate maintenance, and a studied failure occurrence. Correct system operation involves those activities and novel methodologies to detect, classify, and isolate faults and failures and model and simulate processes with predictive algorithms and analytics. In this paper, we showcase the application of a complex-adaptive, self-organizing (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  36.  12
    Reinforcement Learning With Parsimonious Computation and a Forgetting Process.Asako Toyama, Kentaro Katahira & Hideki Ohira - 2019 - Frontiers in Human Neuroscience 13.
  37.  27
    Deep Reinforcement Learning for Vectored Thruster Autonomous Underwater Vehicle Control.Tao Liu, Yuli Hu & Hui Xu - 2021 - Complexity 2021:1-25.
    Autonomous underwater vehicles are widely used to accomplish various missions in the complex marine environment; the design of a control system for AUVs is particularly difficult due to the high nonlinearity, variations in hydrodynamic coefficients, and external force from ocean currents. In this paper, we propose a controller based on deep reinforcement learning in a simulation environment for studying the control performance of the vectored thruster AUV. RL is an important method of artificial intelligence that can learn behavior (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  38.  15
    Reinforcement Learning in Autism Spectrum Disorder.Manuela Schuetze, Christiane S. Rohr, Deborah Dewey, Adam McCrimmon & Signe Bray - 2017 - Frontiers in Psychology 8.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  39.  11
    “Reconciling reinforcement learning models with behavioral extinction and renewal: Implications for addiction, relapse, and problem gambling”: Correction.David A. Redish, Steve Jensen, Adam Johnson & Zeb Kurth-Nelson - 2009 - Psychological Review 116 (3):518-518.
  40.  20
    Reinforcement learning for Golog programs with first-order state-abstraction.D. Beck & G. Lakemeyer - 2012 - Logic Journal of the IGPL 20 (5):909-942.
  41.  29
    Network formation by reinforcement learning: The long and medium run.Brian Skyrms - unknown
    We investigate a simple stochastic model of social network formation by the process of reinforcement learning with discounting of the past. In the limit, for any value of the discounting parameter, small, stable cliques are formed. However, the time it takes to reach the limiting state in which cliques have formed is very sensitive to the discounting parameter. Depending on this value, the limiting result may or may not be a good predictor for realistic observation times.
    Direct download  
     
    Export citation  
     
    Bookmark   8 citations  
  42. Multi-Agent Reinforcement Learning: Weighting and Partitioning.Ron Sun & Todd Peterson - unknown
    This paper addresses weighting and partitioning in complex reinforcement learning tasks, with the aim of facilitating learning. The paper presents some ideas regarding weighting of multiple agents and extends them into partitioning an input/state space into multiple regions with di erential weighting in these regions, to exploit di erential characteristics of regions and di erential characteristics of agents to reduce the learning complexity of agents (and their function approximators) and thus to facilitate the learning overall. (...)
     
    Export citation  
     
    Bookmark   6 citations  
  43.  15
    Integrating reinforcement learning, equilibrium points, and minimum variance to understand the development of reaching: A computational model.Daniele Caligiore, Domenico Parisi & Gianluca Baldassarre - 2014 - Psychological Review 121 (3):389-421.
    No categories
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  44.  55
    Deep Reinforcement Learning as Foundation for Artificial General Intelligence.Itamar Arel - 2012 - In Pei Wang & Ben Goertzel (eds.), Theoretical Foundations of Artificial General Intelligence. Springer. pp. 89--102.
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  45.  19
    A reinforcement learning approach to gait training improves retention.Christopher J. Hasson, Julia Manczurowsky & Sheng-Che Yen - 2015 - Frontiers in Human Neuroscience 9.
  46.  26
    Reinforcement learning and higher level cognition: Introduction to special issue.Nathaniel D. Daw & Michael J. Frank - 2009 - Cognition 113 (3):259-261.
  47.  42
    Can reinforcement learning explain variation in early infant crying?Arnon Lotem & David W. Winkler - 2004 - Behavioral and Brain Sciences 27 (4):468-468.
    We welcome Soltis' use of evolutionary signaling theory, but question his interpretations of colic as a signal of vigor and his explanation of abnormal high-pitched crying as a signal of poor infant quality. Instead, we suggest that these phenomena may be suboptimal by-products of a generally adaptive learning process by which infants adjust their crying levels in relation to parental responsiveness.
    Direct download (7 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  48.  10
    Relational reinforcement learning with guided demonstrations.David Martínez, Guillem Alenyà & Carme Torras - 2017 - Artificial Intelligence 247 (C):295-312.
  49. Reinforcement learning with raw image pixels as state input.D. Ernst, R. Marée & L. Wehenkel - 2006 - In O. Stock & M. Schaerf (eds.), Lecture Notes In Computer Science. Springer Verlag. pp. 4153.
  50.  18
    Cross-Modal Transfer Learning From EEG to Functional Near-Infrared Spectroscopy for Classification Task in Brain-Computer Interface System.Yuqing Wang, Zhiqiang Yang, Hongfei Ji, Jie Li, Lingyu Liu & Jie Zhuang - 2022 - Frontiers in Psychology 13.
    The brain-computer interface based on functional near-infrared spectroscopy has received more and more attention due to its vast application potential in emotion recognition. However, the relatively insufficient investigation of the feature extraction algorithms limits its use in practice. In this article, to improve the performance of fNIRS-based BCI, we proposed a method named R-CSP-E, which introduces EEG signals when computing fNIRS signals’ features based on transfer learning and ensemble learning theory. In detail, we used the Independent Component Analysis (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
1 — 50 / 991