Author: Olivier Sigaud
Publisher: John Wiley & Sons
ISBN: 1118620100
Category : Technology & Engineering
Languages : en
Pages : 367
Book Description
Markov Decision Processes (MDPs) are a mathematical framework for modeling sequential decision problems under uncertainty as well as reinforcement learning problems. Written by experts in the field, this book provides a global view of current research using MDPs in artificial intelligence. It starts with an introductory presentation of the fundamental aspects of MDPs (planning in MDPs, reinforcement learning, partially observable MDPs, Markov games and the use of non-classical criteria). It then presents more advanced research trends in the field and gives some concrete examples using illustrative real life applications.
Markov Decision Processes in Artificial Intelligence
Author: Olivier Sigaud
Publisher: John Wiley & Sons
ISBN: 1118620100
Category : Technology & Engineering
Languages : en
Pages : 367
Book Description
Markov Decision Processes (MDPs) are a mathematical framework for modeling sequential decision problems under uncertainty as well as reinforcement learning problems. Written by experts in the field, this book provides a global view of current research using MDPs in artificial intelligence. It starts with an introductory presentation of the fundamental aspects of MDPs (planning in MDPs, reinforcement learning, partially observable MDPs, Markov games and the use of non-classical criteria). It then presents more advanced research trends in the field and gives some concrete examples using illustrative real life applications.
Publisher: John Wiley & Sons
ISBN: 1118620100
Category : Technology & Engineering
Languages : en
Pages : 367
Book Description
Markov Decision Processes (MDPs) are a mathematical framework for modeling sequential decision problems under uncertainty as well as reinforcement learning problems. Written by experts in the field, this book provides a global view of current research using MDPs in artificial intelligence. It starts with an introductory presentation of the fundamental aspects of MDPs (planning in MDPs, reinforcement learning, partially observable MDPs, Markov games and the use of non-classical criteria). It then presents more advanced research trends in the field and gives some concrete examples using illustrative real life applications.
Planning with Markov Decision Processes
Author: Mausam Natarajan
Publisher: Springer Nature
ISBN: 3031015592
Category : Computers
Languages : en
Pages : 204
Book Description
Markov Decision Processes (MDPs) are widely popular in Artificial Intelligence for modeling sequential decision-making scenarios with probabilistic dynamics. They are the framework of choice when designing an intelligent agent that needs to act for long periods of time in an environment where its actions could have uncertain outcomes. MDPs are actively researched in two related subareas of AI, probabilistic planning and reinforcement learning. Probabilistic planning assumes known models for the agent's goals and domain dynamics, and focuses on determining how the agent should behave to achieve its objectives. On the other hand, reinforcement learning additionally learns these models based on the feedback the agent gets from the environment. This book provides a concise introduction to the use of MDPs for solving probabilistic planning problems, with an emphasis on the algorithmic perspective. It covers the whole spectrum of the field, from the basics to state-of-the-art optimal and approximation algorithms. We first describe the theoretical foundations of MDPs and the fundamental solution techniques for them. We then discuss modern optimal algorithms based on heuristic search and the use of structured representations. A major focus of the book is on the numerous approximation schemes for MDPs that have been developed in the AI literature. These include determinization-based approaches, sampling techniques, heuristic functions, dimensionality reduction, and hierarchical representations. Finally, we briefly introduce several extensions of the standard MDP classes that model and solve even more complex planning problems. Table of Contents: Introduction / MDPs / Fundamental Algorithms / Heuristic Search Algorithms / Symbolic Algorithms / Approximation Algorithms / Advanced Notes
Publisher: Springer Nature
ISBN: 3031015592
Category : Computers
Languages : en
Pages : 204
Book Description
Markov Decision Processes (MDPs) are widely popular in Artificial Intelligence for modeling sequential decision-making scenarios with probabilistic dynamics. They are the framework of choice when designing an intelligent agent that needs to act for long periods of time in an environment where its actions could have uncertain outcomes. MDPs are actively researched in two related subareas of AI, probabilistic planning and reinforcement learning. Probabilistic planning assumes known models for the agent's goals and domain dynamics, and focuses on determining how the agent should behave to achieve its objectives. On the other hand, reinforcement learning additionally learns these models based on the feedback the agent gets from the environment. This book provides a concise introduction to the use of MDPs for solving probabilistic planning problems, with an emphasis on the algorithmic perspective. It covers the whole spectrum of the field, from the basics to state-of-the-art optimal and approximation algorithms. We first describe the theoretical foundations of MDPs and the fundamental solution techniques for them. We then discuss modern optimal algorithms based on heuristic search and the use of structured representations. A major focus of the book is on the numerous approximation schemes for MDPs that have been developed in the AI literature. These include determinization-based approaches, sampling techniques, heuristic functions, dimensionality reduction, and hierarchical representations. Finally, we briefly introduce several extensions of the standard MDP classes that model and solve even more complex planning problems. Table of Contents: Introduction / MDPs / Fundamental Algorithms / Heuristic Search Algorithms / Symbolic Algorithms / Approximation Algorithms / Advanced Notes
Markov Decision Processes
Author: Martin L. Puterman
Publisher: John Wiley & Sons
ISBN: 1118625870
Category : Mathematics
Languages : en
Pages : 544
Book Description
The Wiley-Interscience Paperback Series consists of selected books that have been made more accessible to consumers in an effort to increase global appeal and general circulation. With these new unabridged softcover volumes, Wiley hopes to extend the lives of these works by making them available to future generations of statisticians, mathematicians, and scientists. "This text is unique in bringing together so many results hitherto found only in part in other texts and papers. . . . The text is fairly self-contained, inclusive of some basic mathematical results needed, and provides a rich diet of examples, applications, and exercises. The bibliographical material at the end of each chapter is excellent, not only from a historical perspective, but because it is valuable for researchers in acquiring a good perspective of the MDP research potential." —Zentralblatt fur Mathematik ". . . it is of great value to advanced-level students, researchers, and professional practitioners of this field to have now a complete volume (with more than 600 pages) devoted to this topic. . . . Markov Decision Processes: Discrete Stochastic Dynamic Programming represents an up-to-date, unified, and rigorous treatment of theoretical and computational aspects of discrete-time Markov decision processes." —Journal of the American Statistical Association
Publisher: John Wiley & Sons
ISBN: 1118625870
Category : Mathematics
Languages : en
Pages : 544
Book Description
The Wiley-Interscience Paperback Series consists of selected books that have been made more accessible to consumers in an effort to increase global appeal and general circulation. With these new unabridged softcover volumes, Wiley hopes to extend the lives of these works by making them available to future generations of statisticians, mathematicians, and scientists. "This text is unique in bringing together so many results hitherto found only in part in other texts and papers. . . . The text is fairly self-contained, inclusive of some basic mathematical results needed, and provides a rich diet of examples, applications, and exercises. The bibliographical material at the end of each chapter is excellent, not only from a historical perspective, but because it is valuable for researchers in acquiring a good perspective of the MDP research potential." —Zentralblatt fur Mathematik ". . . it is of great value to advanced-level students, researchers, and professional practitioners of this field to have now a complete volume (with more than 600 pages) devoted to this topic. . . . Markov Decision Processes: Discrete Stochastic Dynamic Programming represents an up-to-date, unified, and rigorous treatment of theoretical and computational aspects of discrete-time Markov decision processes." —Journal of the American Statistical Association
Reinforcement Learning
Author: Marco Wiering
Publisher: Springer Science & Business Media
ISBN: 3642276458
Category : Technology & Engineering
Languages : en
Pages : 653
Book Description
Reinforcement learning encompasses both a science of adaptive behavior of rational beings in uncertain environments and a computational methodology for finding optimal behaviors for challenging problems in control, optimization and adaptive behavior of intelligent agents. As a field, reinforcement learning has progressed tremendously in the past decade. The main goal of this book is to present an up-to-date series of survey articles on the main contemporary sub-fields of reinforcement learning. This includes surveys on partially observable environments, hierarchical task decompositions, relational knowledge representation and predictive state representations. Furthermore, topics such as transfer, evolutionary methods and continuous spaces in reinforcement learning are surveyed. In addition, several chapters review reinforcement learning methods in robotics, in games, and in computational neuroscience. In total seventeen different subfields are presented by mostly young experts in those areas, and together they truly represent a state-of-the-art of current reinforcement learning research. Marco Wiering works at the artificial intelligence department of the University of Groningen in the Netherlands. He has published extensively on various reinforcement learning topics. Martijn van Otterlo works in the cognitive artificial intelligence group at the Radboud University Nijmegen in The Netherlands. He has mainly focused on expressive knowledge representation in reinforcement learning settings.
Publisher: Springer Science & Business Media
ISBN: 3642276458
Category : Technology & Engineering
Languages : en
Pages : 653
Book Description
Reinforcement learning encompasses both a science of adaptive behavior of rational beings in uncertain environments and a computational methodology for finding optimal behaviors for challenging problems in control, optimization and adaptive behavior of intelligent agents. As a field, reinforcement learning has progressed tremendously in the past decade. The main goal of this book is to present an up-to-date series of survey articles on the main contemporary sub-fields of reinforcement learning. This includes surveys on partially observable environments, hierarchical task decompositions, relational knowledge representation and predictive state representations. Furthermore, topics such as transfer, evolutionary methods and continuous spaces in reinforcement learning are surveyed. In addition, several chapters review reinforcement learning methods in robotics, in games, and in computational neuroscience. In total seventeen different subfields are presented by mostly young experts in those areas, and together they truly represent a state-of-the-art of current reinforcement learning research. Marco Wiering works at the artificial intelligence department of the University of Groningen in the Netherlands. He has published extensively on various reinforcement learning topics. Martijn van Otterlo works in the cognitive artificial intelligence group at the Radboud University Nijmegen in The Netherlands. He has mainly focused on expressive knowledge representation in reinforcement learning settings.
Reinforcement Learning, second edition
Author: Richard S. Sutton
Publisher: MIT Press
ISBN: 0262352702
Category : Computers
Languages : en
Pages : 549
Book Description
The significantly expanded and updated new edition of a widely used text on reinforcement learning, one of the most active research areas in artificial intelligence. Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives while interacting with a complex, uncertain environment. In Reinforcement Learning, Richard Sutton and Andrew Barto provide a clear and simple account of the field's key ideas and algorithms. This second edition has been significantly expanded and updated, presenting new topics and updating coverage of other topics. Like the first edition, this second edition focuses on core online learning algorithms, with the more mathematical material set off in shaded boxes. Part I covers as much of reinforcement learning as possible without going beyond the tabular case for which exact solutions can be found. Many algorithms presented in this part are new to the second edition, including UCB, Expected Sarsa, and Double Learning. Part II extends these ideas to function approximation, with new sections on such topics as artificial neural networks and the Fourier basis, and offers expanded treatment of off-policy learning and policy-gradient methods. Part III has new chapters on reinforcement learning's relationships to psychology and neuroscience, as well as an updated case-studies chapter including AlphaGo and AlphaGo Zero, Atari game playing, and IBM Watson's wagering strategy. The final chapter discusses the future societal impacts of reinforcement learning.
Publisher: MIT Press
ISBN: 0262352702
Category : Computers
Languages : en
Pages : 549
Book Description
The significantly expanded and updated new edition of a widely used text on reinforcement learning, one of the most active research areas in artificial intelligence. Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives while interacting with a complex, uncertain environment. In Reinforcement Learning, Richard Sutton and Andrew Barto provide a clear and simple account of the field's key ideas and algorithms. This second edition has been significantly expanded and updated, presenting new topics and updating coverage of other topics. Like the first edition, this second edition focuses on core online learning algorithms, with the more mathematical material set off in shaded boxes. Part I covers as much of reinforcement learning as possible without going beyond the tabular case for which exact solutions can be found. Many algorithms presented in this part are new to the second edition, including UCB, Expected Sarsa, and Double Learning. Part II extends these ideas to function approximation, with new sections on such topics as artificial neural networks and the Fourier basis, and offers expanded treatment of off-policy learning and policy-gradient methods. Part III has new chapters on reinforcement learning's relationships to psychology and neuroscience, as well as an updated case-studies chapter including AlphaGo and AlphaGo Zero, Atari game playing, and IBM Watson's wagering strategy. The final chapter discusses the future societal impacts of reinforcement learning.
Constrained Markov Decision Processes
Author: Eitan Altman
Publisher: Routledge
ISBN: 1351458248
Category : Mathematics
Languages : en
Pages : 257
Book Description
This book provides a unified approach for the study of constrained Markov decision processes with a finite state space and unbounded costs. Unlike the single controller case considered in many other books, the author considers a single controller with several objectives, such as minimizing delays and loss, probabilities, and maximization of throughputs. It is desirable to design a controller that minimizes one cost objective, subject to inequality constraints on other cost objectives. This framework describes dynamic decision problems arising frequently in many engineering fields. A thorough overview of these applications is presented in the introduction. The book is then divided into three sections that build upon each other.
Publisher: Routledge
ISBN: 1351458248
Category : Mathematics
Languages : en
Pages : 257
Book Description
This book provides a unified approach for the study of constrained Markov decision processes with a finite state space and unbounded costs. Unlike the single controller case considered in many other books, the author considers a single controller with several objectives, such as minimizing delays and loss, probabilities, and maximization of throughputs. It is desirable to design a controller that minimizes one cost objective, subject to inequality constraints on other cost objectives. This framework describes dynamic decision problems arising frequently in many engineering fields. A thorough overview of these applications is presented in the introduction. The book is then divided into three sections that build upon each other.
Operations Research and Health Care
Author: Margaret L. Brandeau
Publisher: Springer Science & Business Media
ISBN: 1402080662
Category : Medical
Languages : en
Pages : 870
Book Description
In both rich and poor nations, public resources for health care are inadequate to meet demand. Policy makers and health care providers must determine how to provide the most effective health care to citizens using the limited resources that are available. This chapter describes current and future challenges in the delivery of health care, and outlines the role that operations research (OR) models can play in helping to solve those problems. The chapter concludes with an overview of this book – its intended audience, the areas covered, and a description of the subsequent chapters. KEY WORDS Health care delivery, Health care planning HEALTH CARE DELIVERY: PROBLEMS AND CHALLENGES 3 1.1 WORLDWIDE HEALTH: THE PAST 50 YEARS Human health has improved significantly in the last 50 years. In 1950, global life expectancy was 46 years [1]. That figure rose to 61 years by 1980 and to 67 years by 1998 [2]. Much of these gains occurred in low- and middle-income countries, and were due in large part to improved nutrition and sanitation, medical innovations, and improvements in public health infrastructure.
Publisher: Springer Science & Business Media
ISBN: 1402080662
Category : Medical
Languages : en
Pages : 870
Book Description
In both rich and poor nations, public resources for health care are inadequate to meet demand. Policy makers and health care providers must determine how to provide the most effective health care to citizens using the limited resources that are available. This chapter describes current and future challenges in the delivery of health care, and outlines the role that operations research (OR) models can play in helping to solve those problems. The chapter concludes with an overview of this book – its intended audience, the areas covered, and a description of the subsequent chapters. KEY WORDS Health care delivery, Health care planning HEALTH CARE DELIVERY: PROBLEMS AND CHALLENGES 3 1.1 WORLDWIDE HEALTH: THE PAST 50 YEARS Human health has improved significantly in the last 50 years. In 1950, global life expectancy was 46 years [1]. That figure rose to 61 years by 1980 and to 67 years by 1998 [2]. Much of these gains occurred in low- and middle-income countries, and were due in large part to improved nutrition and sanitation, medical innovations, and improvements in public health infrastructure.
Handbook of Healthcare Analytics
Author: Tinglong Dai
Publisher: John Wiley & Sons
ISBN: 1119300940
Category : Business & Economics
Languages : en
Pages : 482
Book Description
How can analytics scholars and healthcare professionals access the most exciting and important healthcare topics and tools for the 21st century? Editors Tinglong Dai and Sridhar Tayur, aided by a team of internationally acclaimed experts, have curated this timely volume to help newcomers and seasoned researchers alike to rapidly comprehend a diverse set of thrusts and tools in this rapidly growing cross-disciplinary field. The Handbook covers a wide range of macro-, meso- and micro-level thrusts—such as market design, competing interests, global health, personalized medicine, residential care and concierge medicine, among others—and structures what has been a highly fragmented research area into a coherent scientific discipline. The handbook also provides an easy-to-comprehend introduction to five essential research tools—Markov decision process, game theory and information economics, queueing games, econometric methods, and data science—by illustrating their uses and applicability on examples from diverse healthcare settings, thus connecting tools with thrusts. The primary audience of the Handbook includes analytics scholars interested in healthcare and healthcare practitioners interested in analytics. This Handbook: Instills analytics scholars with a way of thinking that incorporates behavioral, incentive, and policy considerations in various healthcare settings. This change in perspective—a shift in gaze away from narrow, local and one-off operational improvement efforts that do not replicate, scale or remain sustainable—can lead to new knowledge and innovative solutions that healthcare has been seeking so desperately. Facilitates collaboration between healthcare experts and analytics scholar to frame and tackle their pressing concerns through appropriate modern mathematical tools designed for this very purpose. The handbook is designed to be accessible to the independent reader, and it may be used in a variety of settings, from a short lecture series on specific topics to a semester-long course.
Publisher: John Wiley & Sons
ISBN: 1119300940
Category : Business & Economics
Languages : en
Pages : 482
Book Description
How can analytics scholars and healthcare professionals access the most exciting and important healthcare topics and tools for the 21st century? Editors Tinglong Dai and Sridhar Tayur, aided by a team of internationally acclaimed experts, have curated this timely volume to help newcomers and seasoned researchers alike to rapidly comprehend a diverse set of thrusts and tools in this rapidly growing cross-disciplinary field. The Handbook covers a wide range of macro-, meso- and micro-level thrusts—such as market design, competing interests, global health, personalized medicine, residential care and concierge medicine, among others—and structures what has been a highly fragmented research area into a coherent scientific discipline. The handbook also provides an easy-to-comprehend introduction to five essential research tools—Markov decision process, game theory and information economics, queueing games, econometric methods, and data science—by illustrating their uses and applicability on examples from diverse healthcare settings, thus connecting tools with thrusts. The primary audience of the Handbook includes analytics scholars interested in healthcare and healthcare practitioners interested in analytics. This Handbook: Instills analytics scholars with a way of thinking that incorporates behavioral, incentive, and policy considerations in various healthcare settings. This change in perspective—a shift in gaze away from narrow, local and one-off operational improvement efforts that do not replicate, scale or remain sustainable—can lead to new knowledge and innovative solutions that healthcare has been seeking so desperately. Facilitates collaboration between healthcare experts and analytics scholar to frame and tackle their pressing concerns through appropriate modern mathematical tools designed for this very purpose. The handbook is designed to be accessible to the independent reader, and it may be used in a variety of settings, from a short lecture series on specific topics to a semester-long course.
Partially Observed Markov Decision Processes
Author: Vikram Krishnamurthy
Publisher: Cambridge University Press
ISBN: 1107134609
Category : Mathematics
Languages : en
Pages : 491
Book Description
This book covers formulation, algorithms, and structural results of partially observed Markov decision processes, whilst linking theory to real-world applications in controlled sensing. Computations are kept to a minimum, enabling students and researchers in engineering, operations research, and economics to understand the methods and determine the structure of their optimal solution.
Publisher: Cambridge University Press
ISBN: 1107134609
Category : Mathematics
Languages : en
Pages : 491
Book Description
This book covers formulation, algorithms, and structural results of partially observed Markov decision processes, whilst linking theory to real-world applications in controlled sensing. Computations are kept to a minimum, enabling students and researchers in engineering, operations research, and economics to understand the methods and determine the structure of their optimal solution.
The Logic of Adaptive Behavior
Author: Martijn van Otterlo
Publisher: IOS Press
ISBN: 1586039695
Category : Business & Economics
Languages : en
Pages : 508
Book Description
Markov decision processes have become the de facto standard in modeling and solving sequential decision making problems under uncertainty. This book studies lifting Markov decision processes, reinforcement learning and dynamic programming to the first-order (or, relational) setting.
Publisher: IOS Press
ISBN: 1586039695
Category : Business & Economics
Languages : en
Pages : 508
Book Description
Markov decision processes have become the de facto standard in modeling and solving sequential decision making problems under uncertainty. This book studies lifting Markov decision processes, reinforcement learning and dynamic programming to the first-order (or, relational) setting.