Learning Representation and Control in Markov Decision Processes

Learning Representation and Control in Markov Decision Processes PDF Author: Sridhar Mahadevan
Publisher: Now Publishers Inc
ISBN: 1601982380
Category : Computers
Languages : en
Pages : 185

Get Book

Book Description
Provides a comprehensive survey of techniques to automatically construct basis functions or features for value function approximation in Markov decision processes and reinforcement learning.

Learning Representation and Control in Markov Decision Processes

Learning Representation and Control in Markov Decision Processes PDF Author: Sridhar Mahadevan
Publisher: Now Publishers Inc
ISBN: 1601982380
Category : Computers
Languages : en
Pages : 185

Get Book

Book Description
Provides a comprehensive survey of techniques to automatically construct basis functions or features for value function approximation in Markov decision processes and reinforcement learning.

Reinforcement Learning

Reinforcement Learning PDF Author: Marco Wiering
Publisher: Springer Science & Business Media
ISBN: 3642276458
Category : Technology & Engineering
Languages : en
Pages : 653

Get Book

Book Description
Reinforcement learning encompasses both a science of adaptive behavior of rational beings in uncertain environments and a computational methodology for finding optimal behaviors for challenging problems in control, optimization and adaptive behavior of intelligent agents. As a field, reinforcement learning has progressed tremendously in the past decade. The main goal of this book is to present an up-to-date series of survey articles on the main contemporary sub-fields of reinforcement learning. This includes surveys on partially observable environments, hierarchical task decompositions, relational knowledge representation and predictive state representations. Furthermore, topics such as transfer, evolutionary methods and continuous spaces in reinforcement learning are surveyed. In addition, several chapters review reinforcement learning methods in robotics, in games, and in computational neuroscience. In total seventeen different subfields are presented by mostly young experts in those areas, and together they truly represent a state-of-the-art of current reinforcement learning research. Marco Wiering works at the artificial intelligence department of the University of Groningen in the Netherlands. He has published extensively on various reinforcement learning topics. Martijn van Otterlo works in the cognitive artificial intelligence group at the Radboud University Nijmegen in The Netherlands. He has mainly focused on expressive knowledge representation in reinforcement learning settings.

Stochastic Model Checking

Stochastic Model Checking PDF Author: Anne Remke
Publisher: Springer
ISBN: 3662454890
Category : Computers
Languages : en
Pages : 290

Get Book

Book Description
The use of stochastic models in computer science is wide spread, for instance in performance modeling, analysis of randomized algorithms and communication protocols which form the structure of the Internet. Stochastic model checking is an important field in stochastic analysis. It has rapidly gained popularity, due to its powerful and systematic methods to model and analyze stochastic systems. This book presents 7 tutorial lectures given by leading scientists at the ROCKS Autumn School on Stochastic Model Checking, held in Vahrn, Italy, in October 2012. The 7 chapters of this tutorial went through two rounds of reviewing and improvement and are summarizing the state-of-the-art in the field, centered around the tree areas of stochastic models, abstraction techniques and stochastic model checking.

Adaptive Representations for Reinforcement Learning

Adaptive Representations for Reinforcement Learning PDF Author: Shimon Whiteson
Publisher: Springer
ISBN: 3642139329
Category : Technology & Engineering
Languages : en
Pages : 127

Get Book

Book Description
This book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seeks a control policy for a sequential decision task. Since current methods typically rely on manually designed solution representations, agents that automatically adapt their own representations have the potential to dramatically improve performance. This book introduces two novel approaches for automatically discovering high-performing representations. The first approach synthesizes temporal difference methods, the traditional approach to reinforcement learning, with evolutionary methods, which can learn representations for a broad class of optimization problems. This synthesis is accomplished by customizing evolutionary methods to the on-line nature of reinforcement learning and using them to evolve representations for value function approximators. The second approach automatically learns representations based on piecewise-constant approximations of value functions. It begins with coarse representations and gradually refines them during learning, analyzing the current policy and value function to deduce the best refinements. This book also introduces a novel method for devising input representations. This method addresses the feature selection problem by extending an algorithm that evolves the topology and weights of neural networks such that it evolves their inputs too. In addition to introducing these new methods, this book presents extensive empirical results in multiple domains demonstrating that these techniques can substantially improve performance over methods with manual representations.

Constrained Markov Decision Processes

Constrained Markov Decision Processes PDF Author: Eitan Altman
Publisher: Routledge
ISBN: 1351458248
Category : Mathematics
Languages : en
Pages : 256

Get Book

Book Description
This book provides a unified approach for the study of constrained Markov decision processes with a finite state space and unbounded costs. Unlike the single controller case considered in many other books, the author considers a single controller with several objectives, such as minimizing delays and loss, probabilities, and maximization of throughputs. It is desirable to design a controller that minimizes one cost objective, subject to inequality constraints on other cost objectives. This framework describes dynamic decision problems arising frequently in many engineering fields. A thorough overview of these applications is presented in the introduction. The book is then divided into three sections that build upon each other.

Markov Decision Processes in Artificial Intelligence

Markov Decision Processes in Artificial Intelligence PDF Author: Olivier Sigaud
Publisher: John Wiley & Sons
ISBN: 1118620100
Category : Technology & Engineering
Languages : en
Pages : 367

Get Book

Book Description
Markov Decision Processes (MDPs) are a mathematical framework for modeling sequential decision problems under uncertainty as well as reinforcement learning problems. Written by experts in the field, this book provides a global view of current research using MDPs in artificial intelligence. It starts with an introductory presentation of the fundamental aspects of MDPs (planning in MDPs, reinforcement learning, partially observable MDPs, Markov games and the use of non-classical criteria). It then presents more advanced research trends in the field and gives some concrete examples using illustrative real life applications.

Algorithms for Reinforcement Learning

Algorithms for Reinforcement Learning PDF Author: Csaba Grossi
Publisher: Springer Nature
ISBN: 3031015517
Category : Computers
Languages : en
Pages : 89

Get Book

Book Description
Reinforcement learning is a learning paradigm concerned with learning to control a system so as to maximize a numerical performance measure that expresses a long-term objective. What distinguishes reinforcement learning from supervised learning is that only partial feedback is given to the learner about the learner's predictions. Further, the predictions may have long term effects through influencing the future state of the controlled system. Thus, time plays a special role. The goal in reinforcement learning is to develop efficient learning algorithms, as well as to understand the algorithms' merits and limitations. Reinforcement learning is of great interest because of the large number of practical applications that it can be used to address, ranging from problems in artificial intelligence to operations research or control engineering. In this book, we focus on those algorithms of reinforcement learning that build on the powerful theory of dynamic programming. We give a fairly comprehensive catalog of learning problems, describe the core ideas, note a large number of state of the art algorithms, followed by the discussion of their theoretical properties and limitations. Table of Contents: Markov Decision Processes / Value Prediction Problems / Control / For Further Exploration

Representation Discovery using Harmonic Analysis

Representation Discovery using Harmonic Analysis PDF Author: Sridhar Mahadevan
Publisher: Morgan & Claypool Publishers
ISBN: 1598296604
Category : Technology & Engineering
Languages : en
Pages : 160

Get Book

Book Description
Representations are at the heart of artificial intelligence (AI). This book is devoted to the problem of representation discovery: how can an intelligent system construct representations from its experience? Representation discovery re-parameterizes the state space - prior to the application of information retrieval, machine learning, or optimization techniques - facilitating later inference processes by constructing new task-specific bases adapted to the state space geometry. This book presents a general approach to representation discovery using the framework of harmonic analysis, in particular Fourier and wavelet analysis. Biometric compression methods, the compact disc, the computerized axial tomography (CAT) scanner in medicine, JPEG compression, and spectral analysis of time-series data are among the many applications of classical Fourier and wavelet analysis. A central goal of this book is to show that these analytical tools can be generalized from their usual setting in (infinite-dimensional) Euclidean spaces to discrete (finite-dimensional) spaces typically studied in many subfields of AI. Generalizing harmonic analysis to discrete spaces poses many challenges: a discrete representation of the space must be adaptively acquired; basis functions are not pre-defined, but rather must be constructed. Algorithms for efficiently computing and representing bases require dealing with the curse of dimensionality. However, the benefits can outweigh the costs, since the extracted basis functions outperform parametric bases as they often reflect the irregular shape of a particular state space. Case studies from computer graphics, information retrieval, machine learning, and state space planning are used to illustrate the benefits of the proposed framework, and the challenges that remain to be addressed. Representation discovery is an actively developing field, and the author hopes this book will encourage other researchers to explore this exciting area of research. Table of Contents: Overview / Vector Spaces / Fourier Bases on Graphs / Multiscale Bases on Graphs / Scaling to Large Spaces / Case Study: State-Space Planning / Case Study: Computer Graphics / Case Study: Natural Language / Future Directions

Handbook of Simulation Optimization

Handbook of Simulation Optimization PDF Author: Michael C Fu
Publisher: Springer
ISBN: 1493913840
Category : Business & Economics
Languages : en
Pages : 400

Get Book

Book Description
The Handbook of Simulation Optimization presents an overview of the state of the art of simulation optimization, providing a survey of the most well-established approaches for optimizing stochastic simulation models and a sampling of recent research advances in theory and methodology. Leading contributors cover such topics as discrete optimization via simulation, ranking and selection, efficient simulation budget allocation, random search methods, response surface methodology, stochastic gradient estimation, stochastic approximation, sample average approximation, stochastic constraints, variance reduction techniques, model-based stochastic search methods and Markov decision processes. This single volume should serve as a reference for those already in the field and as a means for those new to the field for understanding and applying the main approaches. The intended audience includes researchers, practitioners and graduate students in the business/engineering fields of operations research, management science, operations management and stochastic control, as well as in economics/finance and computer science.

Artificial Intelligence XXXV

Artificial Intelligence XXXV PDF Author: Max Bramer
Publisher: Springer
ISBN: 3030041913
Category : Computers
Languages : en
Pages : 454

Get Book

Book Description
This book constitutes the proceedings of the 38th SGAI International Conference on Innovative Techniques and Applications of Artificial Intelligence, AI 2018, held in Cambridge, UK, in December 2018. The 25 full papers and 12 short papers presented in this volume were carefully reviewed and selected from 46 submissions. There are technical and application papers which were organized in topical sections named: Neural Networks; Planning and Scheduling; Machine Learning; Industrial Applications of Artificial Intelligence; Planning and Scheduling in Action; Machine Learning in Action; Applications of Machine Learning; and Applications of Agent Systems and Genetic Algorithms.