Results for 'Q-learning, dynamic structuring of exploration space, reinforcement learning, genetic algorithm'

1000+ found
Order:
  1.  21
    Ga により探索空間の動的生成を行う Q 学習.Matsuno Fumitoshi Ito Kazuyuki - 2001 - Transactions of the Japanese Society for Artificial Intelligence 16:510-520.
    Reinforcement learning has recently received much attention as a learning method for complicated systems, e.g., robot systems. It does not need prior knowledge and has higher capability of reactive and adaptive behaviors. However increase in dimensionality of the action-state space makes it diffcult to accomplish learning. The applicability of the existing reinforcement learning algorithms are effective for simple tasks with relatively small action-state space. In this paper, we propose a new reinforcement learning algorithm: “Q-learning with (...) Structuring of Exploration Space Based on Genetic Algorithm ”. The algorithm is applicable to systems with high dimensional action and interior state spaces, for example a robot with many redundant degrees of freedom. To demonstrate the effectiveness of the proposed algorithm simulations of obstacle avoidance by a 50 links manipulator have been carried out. It is shown that effective behavior can be learned by using the proposed algorithm. (shrink)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  2.  23
    Qdsega による多足ロボットの歩行運動の獲得.Matsuno Fumitoshi Ito Kazuyuki - 2002 - Transactions of the Japanese Society for Artificial Intelligence 17:363-372.
    Reinforcement learning is very effective for robot learning. Because it does not need priori knowledge and has higher capability of reactive and adaptive behaviors. In our previous works, we proposed new reinforcement learning algorithm: “Q-learning with Dynamic Structuring of Exploration Space Based on Genetic Algorithm (QDSEGA)”. It is designed for complicated systems with large action-state space like a robot with many redundant degrees of freedom. And we applied it to 50 link manipulator (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  3.  11
    Economic Structure Analysis Based on Neural Network and Bionic Algorithm.Yanjun Dai & Lin Su - 2021 - Complexity 2021:1-11.
    In this article, an in-depth study and analysis of economic structure are carried out using a neural network fusion release algorithm. The method system defines the weight space and structure space of neural networks from the perspective of optimization theory, proposes a bionic optimization algorithm under the weight space and structure space, and establishes a neuroevolutionary method with shallow neural network and deep neural network as the research objects. In the shallow neuroevolutionary, the improved genetic algorithm (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  4. Integrating reinforcement learning, bidding and genetic algorithms.Ron Sun - unknown
    This paper presents a GA-based multi-agent reinforce- ment learning bidding approach (GMARLB) for perform- ing multi-agent reinforcement learning. GMARLB inte- grates reinforcement learning, bidding and genetic algo- rithms. The general idea of our multi-agent systems is as follows: There are a number of individual agents in a team, each agent of the team has two modules: Q module and CQ module. Each agent can select actions to be performed at each step, which are done by the Q (...)
     
    Export citation  
     
    Bookmark  
  5.  20
    Resilience Analysis of Urban Road Networks Based on Adaptive Signal Controls: Day-to-Day Traffic Dynamics with Deep Reinforcement Learning.Wen-Long Shang, Yanyan Chen, Xingang Li & Washington Y. Ochieng - 2020 - Complexity 2020:1-19.
    Improving the resilience of urban road networks suffering from various disruptions has been a central focus for urban emergence management. However, to date the effective methods which may mitigate the negative impacts caused by the disruptions, such as road accidents and natural disasters, on urban road networks is highly insufficient. This study proposes a novel adaptive signal control strategy based on a doubly dynamic learning framework, which consists of deep reinforcement learning and day-to-day traffic dynamic learning, to (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  6.  10
    Causal Structure Learning in Continuous Systems.Zachary J. Davis, Neil R. Bramley & Bob Rehder - 2020 - Frontiers in Psychology 11.
    Real causal systems are complicated. Despite this, causal learning research has traditionally emphasized how causal relations can be induced on the basis of idealized events, i.e. those that have been mapped to binary variables and abstracted from time. For example, participants may be asked to assess the efficacy of a headache-relief pill on the basis of multiple patients who take the pill (or not) and find their headache relieved (or not). In contrast, the current study examines learning via interactions with (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  7.  6
    Reinforcement Learning with Probabilistic Boolean Network Models of Smart Grid Devices.Pedro Juan Rivera Torres, Carlos Gershenson García, María Fernanda Sánchez Puig & Samir Kanaan Izquierdo - 2022 - Complexity 2022:1-15.
    The area of smart power grids needs to constantly improve its efficiency and resilience, to provide high quality electrical power in a resilient grid, while managing faults and avoiding failures. Achieving this requires high component reliability, adequate maintenance, and a studied failure occurrence. Correct system operation involves those activities and novel methodologies to detect, classify, and isolate faults and failures and model and simulate processes with predictive algorithms and analytics. In this paper, we showcase the application of a complex-adaptive, self-organizing (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  8.  20
    Online Optimal Control of Robotic Systems with Single Critic NN-Based Reinforcement Learning.Xiaoyi Long, Zheng He & Zhongyuan Wang - 2021 - Complexity 2021:1-7.
    This paper suggests an online solution for the optimal tracking control of robotic systems based on a single critic neural network -based reinforcement learning method. To this end, we rewrite the robotic system model as a state-space form, which will facilitate the realization of optimal tracking control synthesis. To maintain the tracking response, a steady-state control is designed, and then an adaptive optimal tracking control is used to ensure that the tracking error can achieve convergence in an optimal sense. (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  9.  63
    Novelty and Inductive Generalization in Human Reinforcement Learning.Samuel J. Gershman & Yael Niv - 2015 - Topics in Cognitive Science 7 (3):391-415.
    In reinforcement learning, a decision maker searching for the most rewarding option is often faced with the question: What is the value of an option that has never been tried before? One way to frame this question is as an inductive problem: How can I generalize my previous experience with one set of options to a novel option? We show how hierarchical Bayesian inference can be used to solve this problem, and we describe an equivalence between the Bayesian model (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  10.  17
    Learning the Structure of Bayesian Networks: A Quantitative Assessment of the Effect of Different Algorithmic Schemes.Stefano Beretta, Mauro Castelli, Ivo Gonçalves, Roberto Henriques & Daniele Ramazzotti - 2018 - Complexity 2018:1-12.
    One of the most challenging tasks when adopting Bayesian networks is the one of learning their structure from data. This task is complicated by the huge search space of possible solutions and by the fact that the problem isNP-hard. Hence, a full enumeration of all the possible solutions is not always feasible and approximations are often required. However, to the best of our knowledge, a quantitative analysis of the performance and characteristics of the different heuristics to solve this problem has (...)
    No categories
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark  
  11.  9
    Foundations of algorithms.Richard E. Neapolitan - 2015 - Burlington, MA: Jones & Bartlett Learning.
    Foundations of Algorithms, Fifth Edition offers a well-balanced presentation of algorithm design, complexity analysis of algorithms, and computational complexity. Ideal for any computer science students with a background in college algebra and discrete structures, the text presents mathematical concepts using standard English and simple notation to maximize accessibility and user-friendliness. Concrete examples, appendices reviewing essential mathematical concepts, and a student-focused approach reinforce theoretical explanations and promote learning and retention. C++ and Java pseudocode help students better understand complex algorithms. A (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  12.  3
    Multi-agent reinforcement learning based algorithm detection of malware-infected nodes in IoT networks.Marcos Severt, Roberto Casado-Vara, Ángel Martín del Rey, Héctor Quintián & Jose Luis Calvo-Rolle - forthcoming - Logic Journal of the IGPL.
    The Internet of Things (IoT) is a fast-growing technology that connects everyday devices to the Internet, enabling wireless, low-consumption and low-cost communication and data exchange. IoT has revolutionized the way devices interact with each other and the internet. The more devices become connected, the greater the risk of security breaches. There is currently a need for new approaches to algorithms that can detect malware regardless of the size of the network and that can adapt to dynamic changes in the (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  13. Learning to perceive in the sensorimotor approach: Piaget’s theory of equilibration interpreted dynamically.Ezequiel A. Di Paolo, Xabier E. Barandiaran, Michael Beaton & Thomas Buhrmann - 2014 - Frontiers in Human Neuroscience 8:551.
    Learning to perceive is faced with a classical paradox: if understanding is required for perception, how can we learn to perceive something new, something we do not yet understand? According to the sensorimotor approach, perception involves mastery of regular sensorimotor co-variations that depend on the agent and the environment, also known as the “laws” of sensorimotor contingencies (SMCs). In this sense, perception involves enacting relevant sensorimotor skills in each situation. It is important for this proposal that such skills can be (...)
     
    Export citation  
     
    Bookmark  
  14.  13
    Structural-parametric synthesis of deep learning neural networks.Sineglazov V. M. & Chumachenko O. I. - 2020 - Artificial Intelligence Scientific Journal 25 (4):42-51.
    The structural-parametric synthesis of neural networks of deep learning, in particular convolutional neural networks used in image processing, is considered. The classification of modern architectures of convolutional neural networks is given. It is shown that almost every convolutional neural network, depending on its topology, has unique blocks that determine its essential features, Residual block, Inception module, ResNeXt block. It is stated the problem of structural-parametric synthesis of convolutional neural networks, for the solution of which it is proposed to use a (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  15. Automatic Partitioning for Multi-Agent Reinforcement Learning.Ron Sun - unknown
    This paper addresses automatic partitioning in complex reinforcement learning tasks with multiple agents, without a priori domain knowledge regarding task structures. Partitioning a state/input space into multiple regions helps to exploit the di erential characteristics of regions and di erential characteristics of agents, thus facilitating learning and reducing the complexity of agents especially when function approximators are used. We develop a method for optimizing the partitioning of the space through experience without the use of a priori domain knowledge. The (...)
     
    Export citation  
     
    Bookmark  
  16.  56
    The structure of intrinsic complexity of learning.Sanjay Jain & Arun Sharma - 1997 - Journal of Symbolic Logic 62 (4):1187-1201.
    Limiting identification of r.e. indexes for r.e. languages (from a presentation of elements of the language) and limiting identification of programs for computable functions (from a graph of the function) have served as models for investigating the boundaries of learnability. Recently, a new approach to the study of "intrinsic" complexity of identification in the limit has been proposed. This approach, instead of dealing with the resource requirements of the learning algorithm, uses the notion of reducibility from recursion theory to (...)
    Direct download (7 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  17. The structure of egocentric space.Adrian J. T. Alsmith - 2020 - In Frédérique de Vignemont (ed.), The World at Our Fingertips: A Multidisciplinary Exploration of Peripersonal Space. Oxford: Oxford University Press.
    This chapter offers an indirect defence of the Evansian conception of egocentric space, by showing how it resolves a puzzle concerning the unity of egocentric spatial perception. The chapter outlines several common assumptions about egocentric perspectival structure and argues that a subject’s experience, both within and across her sensory modalities, may involve multiple structures of this kind. This raises the question of how perspectival unity is achieved, such that these perspectival structures form a complex whole, rather than merely disunified set (...)
    Direct download  
     
    Export citation  
     
    Bookmark   3 citations  
  18.  16
    Q-Learning Applied to Genetic Algorithm-Fuzzy Approach for On-Line Control in Autonomous Agents.Hengameh Sarmadi - 2009 - Journal of Intelligent Systems 18 (1-2):1-32.
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  19. Are People Successful at Learning Sequences of Actions on a Perceptual Matching Task?Reiko Yakushijin & Robert A. Jacobs - 2011 - Cognitive Science 35 (5):939-962.
    We report the results of an experiment in which human subjects were trained to perform a perceptual matching task. Subjects were asked to manipulate comparison objects until they matched target objects using the fewest manipulations possible. An unusual feature of the experimental task is that efficient performance requires an understanding of the hidden or latent causal structure governing the relationships between actions and perceptual outcomes. We use two benchmarks to evaluate the quality of subjects’ learning. One benchmark is based on (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  20. The Structure of Intrinsic Complexity of Learning.Sanjay Jain & Arun Sharma - 1997 - Journal of Symbolic Logic 62 (4):1187-1201.
    Limiting identification of r.e. indexes for r.e. languages and limiting identification of programs for computable functions have served as models for investigating the boundaries of learnability. Recently, a new approach to the study of "intrinsic" complexity of identification in the limit has been proposed. This approach, instead of dealing with the resource requirements of the learning algorithm, uses the notion of reducibility from recursion theory to compare and to capture the intuitive difficulty of learning various classes of concepts. Freivalds, (...)
     
    Export citation  
     
    Bookmark   1 citation  
  21.  29
    Decolonization Projects.Cornelius Ewuoso - 2023 - Voices in Bioethics 9.
    Photo ID 279661800 © Sidewaypics|Dreamstime.com ABSTRACT Decolonization is complex, vast, and the subject of an ongoing academic debate. While the many efforts to decolonize or dismantle the vestiges of colonialism that remain are laudable, they can also reinforce what they seek to end. For decolonization to be impactful, it must be done with epistemic and cultural humility, requiring decolonial scholars, project leaders, and well-meaning people to be more sensitive to those impacted by colonization and not regularly included in the discourse. (...)
    No categories
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark  
  22.  28
    Controle da diversidade da população em algoritmos genéticos aplicados na predição de estruturas de proteínas.Vinicius Tragante do Ó & Renato Tinos - 2009 - Scientia (Brazil) 20 (2):83-93.
    Control of the population diversity in genetic algorithms applied to the protein structure prediction problem. Genetic Algorithms (GAs), a successful approach for optimization problems, usually fail when employed in the standard configuration in the protein structure prediction problem, since the solution space is very large and the population converges before a reasonable percentage of the possible solutions is explored. Thus, this work investigates the effect of increasing the diversity of the population on this problem by using Hypermutation and (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  23.  45
    The Impact of Goal Specificity on Strategy Use and the Acquisition of Problem Structure.Regina Vollmeyer, Bruce D. Burns & Keith J. Holyoak - 1996 - Cognitive Science 20 (1):75-100.
    Theories of skill acquisition have made radically different predictions about the role of general problem‐solving methods in acquiring rules that promote effective transfer to new problems. Under one view, methods that focus on reaching specific goals, such as means‐ends analysis, are assumed to provide the basis for efficient knowledge compilation (Anderson, 1987), whereas under an alternative view such methods are believed to disrupt rule induction (Sweller, 1988). We suggest that the role of general methods in learning varies with both the (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   17 citations  
  24.  41
    Face recognition algorithms and the other‐race effect: computational mechanisms for a developmental contact hypothesis.Nicholas Furl, P. Jonathon Phillips & Alice J. O'Toole - 2002 - Cognitive Science 26 (6):797-815.
    People recognize faces of their own race more accurately than faces of other races. The “contact” hypothesis suggests that this “other‐race effect” occurs as a result of the greater experience we have with own‐ versus other‐race faces. The computational mechanisms that may underlie different versions of the contact hypothesis were explored in this study. We replicated the other‐race effect with human participants and evaluated four classes of computational face recognition algorithms for the presence of an other‐race effect. Consistent with the (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   10 citations  
  25.  25
    重点サンプリングを用いた Ga による強化学習.Kimura Hajime Tsuchiya Chikao - 2005 - Transactions of the Japanese Society for Artificial Intelligence 20:1-10.
    Reinforcement Learning (RL) handles policy search problems: searching a mapping from state space to action space. However RL is based on gradient methods and as such, cannot deal with problems with multimodal landscape. In contrast, though Genetic Algorithm (GA) is promising to deal with them, it seems to be unsuitable for policy search problems from the viewpoint of the cost of evaluation. Minimal Generation Gap (MGG), used as a generation-alternation model in GA, generates many offspring from two (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  26.  37
    Frontiers of Artificial Intelligence, Ethics, and Multidisciplinary Applications: 1st International Conference on Frontiers of AI, Ethics, and Multidisciplinary Applications (FAIEMA), Greece, 2023.Mina Farmanbar, Maria Tzamtzi, Ajit Kumar Verma & Antorweep Chakravorty (eds.) - 2024 - Springer Nature Singapore.
    This groundbreaking proceedings volume explores the integration of Artificial Intelligence (AI) across key domains—healthcare, finance, education, robotics, industrial and other engineering applications —unveiling its transformative potential and practical implications. With a multidisciplinary lens, it transcends technical aspects, fostering a comprehensive understanding while bridging theory and practice. Approaching the subject matter with depth, the book combines theoretical foundations with real-world case studies, empowering researchers, professionals, and enthusiasts with the knowledge and tools to effectively harness AI. Encompassing diverse AI topics—machine learning, natural (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  27.  20
    Learning Causal Structure through Local Prediction-error Learning.Sarah Wellen & David Danks - unknown
    Research on human causal learning has largely focused on strength learning, or on computational-level theories; there are few formal algorithmic models of how people learn causal structure from covariations. We introduce a model that learns causal structure in a local manner via prediction-error learning. This local learning is then integrated dynamically into a unified representation of causal structure. The model uses computationally plausible approximations of rational learning, and so represents a hybrid between the associationist and rational paradigms in causal learning (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  28.  53
    SA w_ S _u: An Integrated Model of Associative and Reinforcement Learning.Vladislav D. Veksler, Christopher W. Myers & Kevin A. Gluck - 2014 - Cognitive Science 38 (3):580-598.
    Successfully explaining and replicating the complexity and generality of human and animal learning will require the integration of a variety of learning mechanisms. Here, we introduce a computational model which integrates associative learning (AL) and reinforcement learning (RL). We contrast the integrated model with standalone AL and RL models in three simulation studies. First, a synthetic grid‐navigation task is employed to highlight performance advantages for the integrated model in an environment where the reward structure is both diverse and (...). The second and third simulations contrast the performances of the three models in behavioral experiments, demonstrating advantages for the integrated model in accounting for behavioral data. (shrink)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  29. Numerical simulations of the Lewis signaling game: Learning strategies, pooling equilibria, and the evolution of grammar.Jeffrey A. Barrett - unknown
    David Lewis (1969) introduced sender-receiver games as a way of investigating how meaningful language might evolve from initially random signals. In this report I investigate the conditions under which Lewis signaling games evolve to perfect signaling systems under various learning dynamics. While the 2-state/2- term Lewis signaling game with basic urn learning always approaches a signaling system, I will show that with more than two states suboptimal pooling equilibria can evolve. Inhomogeneous state distributions increase the likelihood of pooling equilibria, but (...)
     
    Export citation  
     
    Bookmark   30 citations  
  30.  41
    Experiments on the Accuracy of Algorithms for Inferring the Structure of Genetic Regulatory Networks from Microarray Expression Levels.Joseph Ramsey & Clark Glymour - unknown
    After reviewing theoretical reasons for doubting that machine learning methods can accurately infer gene regulatory networks from microarray data, we test 10 algorithms on simulated data from the sea urchin network, and on microarray data for yeast compared with recent experimental determinations of the regulatory network in the same yeast species. Our results agree with the theoretical arguments: most algorithms are at chance for determining the existence of a regulatory connection between gene pairs, and the algorithms that perform better than (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  31. When, What, and How Much to Reward in Reinforcement Learning-Based Models of Cognition.Christian P. Janssen & Wayne D. Gray - 2012 - Cognitive Science 36 (2):333-358.
    Reinforcement learning approaches to cognitive modeling represent task acquisition as learning to choose the sequence of steps that accomplishes the task while maximizing a reward. However, an apparently unrecognized problem for modelers is choosing when, what, and how much to reward; that is, when (the moment: end of trial, subtask, or some other interval of task performance), what (the objective function: e.g., performance time or performance accuracy), and how much (the magnitude: with binary, categorical, or continuous values). In this (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   6 citations  
  32.  9
    Orthogonal Learning Firefly Algorithm.Tomas Kadavy, Roman Senkerik, Michal Pluhacek & Adam Viktorin - 2021 - Logic Journal of the IGPL 29 (2):167-179.
    The primary aim of this original work is to provide a more in-depth insight into the relations between control parameters adjustments, learning techniques, inner swarm dynamics and possible hybridization strategies for popular swarm metaheuristic Firefly Algorithm. In this paper, a proven method, orthogonal learning, is fused with FA, specifically with its hybrid modification Firefly Particle Swarm Optimization. The parameters of the proposed Orthogonal Learning Firefly Algorithm are also initially thoroughly explored and tuned. The performance of the developed (...) is examined and compared with canonical FA and above-mentioned FFPSO. Comparisons have been conducted on well-known CEC 2017 benchmark functions, and the results have been evaluated for statistical significance using the Friedman rank test. (shrink)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark  
  33.  23
    Κ-確実探査法と動的計画法を用いた mdps 環境の効率的探索法.Kawada Seiichi Tateyama Takeshi - 2001 - Transactions of the Japanese Society for Artificial Intelligence 16:11-19.
    One most common problem in reinforcement learning systems (e.g. Q-learning) is to reduce the number of trials to converge to an optimal policy. As one of the solution to the problem, k-certainty exploration method was proposed. Miyazaki reported that this method could determine an optimal policy faster than Q-learning in Markov decision processes (MDPs). This method is very efficient learning method. But, we propose an improvement plan that makes this method more efficient. In k-certainty exploration method, in (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  34.  10
    Reinforcement Learning for Production‐Based Cognitive Models.Adrian Brasoveanu & Jakub Dotlačil - 2021 - Topics in Cognitive Science 13 (3):467-487.
    We investigate how Reinforcement Learning methods can be used to solve the production selection and production ordering problem in ACT‐R. We focus on four algorithms from the Q learning family, tabular Q and three versions of Deep Q Networks, as well as the ACT‐R utility learning algorithm, which provides a baseline for the Q algorithms. We compare the performance of these five algorithms in a range of lexical decision tasks framed as sequential decision problems.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  35. The role of constrained self-organization in genome structural evolution.Richard Sternberg - 1996 - Acta Biotheoretica 44 (2).
    A hypothesis of genome structural evolution is explored. Rapid and cohesive alterations in genome organization are viewed as resulting from the dynamic and constrained interactions of chromosomal subsystem components. A combination of macromolecular boundary conditions and DNA element involvement in far-from-equilibrium reactions is proposed to increase the complexity of genomic subsystems via the channelling of genome turnover; interactions between subsystems create higher-order subsystems expanding the phase space for further genetic evolution. The operation of generic constraints on structuration in (...)
     
    Export citation  
     
    Bookmark  
  36.  11
    Stochasticity, Nonlinear Value Functions, and Update Rules in Learning Aesthetic Biases.Norberto M. Grzywacz - 2021 - Frontiers in Human Neuroscience 15:639081.
    A theoretical framework for the reinforcement learning of aesthetic biases was recently proposed based on brain circuitries revealed by neuroimaging. A model grounded on that framework accounted for interesting features of human aesthetic biases. These features included individuality, cultural predispositions, stochastic dynamics of learning and aesthetic biases, and the peak-shift effect. However, despite the success in explaining these features, a potential weakness was the linearity of the value function used to predict reward. This linearity meant that the learning process (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  37.  28
    Evolutionary Schema of Modeling Based on Genetic Algorithms.Paweł Stacewicz - 2015 - Studies in Logic, Grammar and Rhetoric 40 (1):219-239.
    In this paper, I propose a populational schema of modeling that consists of: a linear AFSV schema, and a higher-level schema employing the genetic algorithm. The basic ideas of the proposed solution are as follows: whole populations of models are considered at subsequent stages of the modeling process, successive populations are subjected to the activity of genetic operators and undergo selection procedures, the basis for selection is the evaluation function of the genetic algorithm. The schema (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  38.  14
    Characterizing Motor Control of Mastication With Soft Actor-Critic.Amir H. Abdi, Benedikt Sagl, Venkata P. Srungarapu, Ian Stavness, Eitan Prisman, Purang Abolmaesumi & Sidney Fels - 2020 - Frontiers in Human Neuroscience 14:523954.
    The human masticatory system is a complex functional unit characterized by a multitude of skeletal components, muscles, soft tissues, and teeth. Muscle activation dynamics cannot be directly measured on live human subjects due to ethical, safety, and accessibility limitations. Therefore, estimation of muscle activations and their resultant forces is a longstanding and active area of research. Reinforcement learning (RL) is an adaptive learning strategy which is inspired by the behavioral psychology and enables an agent to learn the dynamics of (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  39. HCI Model with Learning Mechanism for Cooperative Design in Pervasive Computing Environment.Hong Liu, Bin Hu & Philip Moore - 2015 - Journal of Internet Technology 16.
    This paper presents a human-computer interaction model with a three layers learning mechanism in a pervasive environment. We begin with a discussion around a number of important issues related to human-computer interaction followed by a description of the architecture for a multi-agent cooperative design system for pervasive computing environment. We present our proposed three- layer HCI model and introduce the group formation algorithm, which is predicated on a dynamic sharing niche technology. Finally, we explore the cooperative reinforcement (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  40.  49
    English Grammar Error Correction Algorithm Based on Classification Model.Shanchun Zhou & Wei Liu - 2021 - Complexity 2021:1-11.
    English grammar error correction algorithm refers to the use of computer programming technology to automatically recognize and correct the grammar errors contained in English text written by nonnative language learners. Classification model is the core of machine learning and data mining, which can be applied to extracting information from English text data and constructing a reliable grammar correction method. On the basis of summarizing and analyzing previous research works, this paper expounded the research status and significance of English grammar (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  41.  59
    Algorithmic management in a work context.Will Sutherland, Eliscia Kinder, Christine T. Wolf, Min Kyung Lee, Gemma Newlands & Mohammad Hossein Jarrahi - 2021 - Big Data and Society 8 (2).
    The rapid development of machine-learning algorithms, which underpin contemporary artificial intelligence systems, has created new opportunities for the automation of work processes and management functions. While algorithmic management has been observed primarily within the platform-mediated gig economy, its transformative reach and consequences are also spreading to more standard work settings. Exploring algorithmic management as a sociotechnical concept, which reflects both technological infrastructures and organizational choices, we discuss how algorithmic management may influence existing power and social structures within organizations. We identify (...)
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark   8 citations  
  42.  13
    Complying with norms. a neurocomputational exploration.Matteo Colombo - 2012 - Dissertation, University of Edinburgh
    The subject matter of this thesis can be summarized by a triplet of questions and answers. Showing what these questions and answers mean is, in essence, the goal of my project. The triplet goes like this: Q: How can we make progress in our understanding of social norms and norm compliance? A: Adopting a neurocomputational framework is one effective way to make progress in our understanding of social norms and norm compliance. Q: What could the neurocomputational mechanism of social norm (...)
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  43.  34
    Learning Causal Structure from Undersampled Time Series.David Danks & Sergey Plis - unknown
    Even if one can experiment on relevant factors, learning the causal structure of a dynamical system can be quite difficult if the relevant measurement processes occur at a much slower sampling rate than the “true” underlying dynamics. This problem is exacerbated if the degree of mismatch is unknown. This paper gives a formal characterization of this learning problem, and then provides two sets of results. First, we prove a set of theorems characterizing how causal structures change under undersampling. Second, we (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   3 citations  
  44.  15
    Solving a Joint Pricing and Inventory Control Problem for Perishables via Deep Reinforcement Learning.Rui Wang, Xianghua Gan, Qing Li & Xiao Yan - 2021 - Complexity 2021:1-17.
    We study a joint pricing and inventory control problem for perishables with positive lead time in a finite horizon periodic-review system. Unlike most studies considering a continuous density function of demand, in our paper the customer demand depends on the price of current period and arrives according to a homogeneous Poisson process. We consider both backlogging and lost-sales cases, and our goal is to find a simultaneously ordering and pricing policy to maximize the expected discounted profit over the planning horizon. (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  45.  83
    No Switchbacks: Rethinking Aspiration-Based Dynamics in the Ultimatum Game. [REVIEW]Jeffrey Carpenter & Peter Hans Matthews - 2005 - Theory and Decision 58 (4):351-385.
    Aspiration-based evolutionary dynamics have recently been used to model the evolution of fair play in the ultimatum game showing that incredible threats to reject low offers persist in equilibrium. We focus on two extensions of this analysis: we experimentally test whether assumptions about agent motivations (aspiration levels) and the structure of the game (binary strategy space) reflect actual play, and we examine the problematic assumption embedded in the standard replicator dynamic that unhappy agents who switch strategies may return to (...)
    Direct download (5 more)  
     
    Export citation  
     
    Bookmark  
  46.  15
    Determination of Fire Resistance of Eccentrically Loaded Reinforced Concrete Columns Using Fuzzy Neural Networks.Marijana Lazarevska, Ana Trombeva Gavriloska, Mirjana Laban, Milos Knezevic & Meri Cvetkovska - 2018 - Complexity 2018:1-12.
    Artificial neural networks, in interaction with fuzzy logic, genetic algorithms, and fuzzy neural networks, represent an example of a modern interdisciplinary field, especially when it comes to solving certain types of engineering problems that could not be solved using traditional modeling methods and statistical methods. They represent a modern trend in practical developments within the prognostic modeling field and, with acceptable limitations, enjoy a generally recognized perspective for application in construction. Results obtained from numerical analysis, which includes analysis of (...)
    No categories
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  47.  16
    The Zhuangzi: Personal Freedom and/or Incongruity of Names?Paul J. D'Ambrosio - 2023 - Philosophy East and West 73 (2):458-466.
    In lieu of an abstract, here is a brief excerpt of the content:The Zhuangzi:Personal Freedom and/or Incongruity of Names?Paul J. D'Ambrosio (bio)Tao Jiang's Origins of Moral-Political Philosophy in Early China: Contestation of Humaneness, Justice, and Personal Freedom (hereafter Origins) has sparked much scholarly debate. Already numerous presentations, various types of discussions, and reviews have appeared based on Origins. The present review focuses specifically on the Zhuangzi chapter. The entire project actually began, Jiang writes, fifteen years ago as a book on (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  48.  41
    Interrogating Feature Learning Models to Discover Insights Into the Development of Human Expertise in a Real‐Time, Dynamic Decision‐Making Task.Catherine Sibert, Wayne D. Gray & John K. Lindstedt - 2017 - Topics in Cognitive Science 9 (2):374-394.
    Tetris provides a difficult, dynamic task environment within which some people are novices and others, after years of work and practice, become extreme experts. Here we study two core skills; namely, choosing the goal or objective function that will maximize performance and a feature-based analysis of the current game board to determine where to place the currently falling zoid so as to maximize the goal. In Study 1, we build cross-entropy reinforcement learning models to determine whether different goals (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   7 citations  
  49.  11
    A Dynamic Opposite Learning Assisted Grasshopper Optimization Algorithm for the Flexible JobScheduling Problem.Yi Feng, Mengru Liu, Yuqian Zhang & Jinglin Wang - 2020 - Complexity 2020:1-19.
    Job shop scheduling problem is one of the most difficult optimization problems in manufacturing industry, and flexible job shop scheduling problem is an extension of the classical JSP, which further challenges the algorithm performance. In FJSP, a machine should be selected for each process from a given set, which introduces another decision element within the job path, making FJSP be more difficult than traditional JSP. In this paper, a variant of grasshopper optimization algorithm named dynamic opposite learning (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  50.  10
    Averaged Soft Actor-Critic for Deep Reinforcement Learning.Feng Ding, Guanfeng Ma, Zhikui Chen, Jing Gao & Peng Li - 2021 - Complexity 2021:1-16.
    With the advent of the era of artificial intelligence, deep reinforcement learning has achieved unprecedented success in high-dimensional and large-scale artificial intelligence tasks. However, the insecurity and instability of the DRL algorithm have an important impact on its performance. The Soft Actor-Critic algorithm uses advanced functions to update the policy and value network to alleviate some of these problems. However, SAC still has some problems. In order to reduce the error caused by the overestimation of SAC, we (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
1 — 50 / 1000