Author: Jacques Janssen
Publisher: Springer Science & Business Media
ISBN: 148990574X
Category : Mathematics
Languages : en
Pages : 572
Book Description
This book is the result of the International Symposium on Semi Markov Processes and their Applications held on June 4-7, 1984 at the Universite Libre de Bruxelles with the help of the FNRS (Fonds National de la Recherche Scientifique, Belgium), the Ministere de l'Education Nationale (Belgium) and the Bernoulli Society for Mathe matical Statistics and Probability. This international meeting was planned to make a state of the art for the area of semi-Markov theory and its applications, to bring together researchers in this field and to create a platform for open and thorough discussion. Main themes of the Symposium are the first ten sections of this book. The last section presented here gives an exhaustive biblio graphy on semi-Markov processes for the last ten years. Papers selected for this book are all invited papers and in addition some contributed papers retained after strong refereeing. Sections are I. Markov additive processes and regenerative systems II. Semi-Markov decision processes III. Algorithmic and computer-oriented approach IV. Semi-Markov models in economy and insurance V. Semi-Markov processes and reliability theory VI. Simulation and statistics for semi-Markov processes VII. Semi-Markov processes and queueing theory VIII. Branching IX. Applications in medicine X. Applications in other fields v PREFACE XI. A second bibliography on semi-Markov processes It is interesting to quote that sections IV to X represent a good sample of the main applications of semi-Markov processes i. e.
Semi-Markov Models
Author: Jacques Janssen
Publisher: Springer Science & Business Media
ISBN: 148990574X
Category : Mathematics
Languages : en
Pages : 572
Book Description
This book is the result of the International Symposium on Semi Markov Processes and their Applications held on June 4-7, 1984 at the Universite Libre de Bruxelles with the help of the FNRS (Fonds National de la Recherche Scientifique, Belgium), the Ministere de l'Education Nationale (Belgium) and the Bernoulli Society for Mathe matical Statistics and Probability. This international meeting was planned to make a state of the art for the area of semi-Markov theory and its applications, to bring together researchers in this field and to create a platform for open and thorough discussion. Main themes of the Symposium are the first ten sections of this book. The last section presented here gives an exhaustive biblio graphy on semi-Markov processes for the last ten years. Papers selected for this book are all invited papers and in addition some contributed papers retained after strong refereeing. Sections are I. Markov additive processes and regenerative systems II. Semi-Markov decision processes III. Algorithmic and computer-oriented approach IV. Semi-Markov models in economy and insurance V. Semi-Markov processes and reliability theory VI. Simulation and statistics for semi-Markov processes VII. Semi-Markov processes and queueing theory VIII. Branching IX. Applications in medicine X. Applications in other fields v PREFACE XI. A second bibliography on semi-Markov processes It is interesting to quote that sections IV to X represent a good sample of the main applications of semi-Markov processes i. e.
Publisher: Springer Science & Business Media
ISBN: 148990574X
Category : Mathematics
Languages : en
Pages : 572
Book Description
This book is the result of the International Symposium on Semi Markov Processes and their Applications held on June 4-7, 1984 at the Universite Libre de Bruxelles with the help of the FNRS (Fonds National de la Recherche Scientifique, Belgium), the Ministere de l'Education Nationale (Belgium) and the Bernoulli Society for Mathe matical Statistics and Probability. This international meeting was planned to make a state of the art for the area of semi-Markov theory and its applications, to bring together researchers in this field and to create a platform for open and thorough discussion. Main themes of the Symposium are the first ten sections of this book. The last section presented here gives an exhaustive biblio graphy on semi-Markov processes for the last ten years. Papers selected for this book are all invited papers and in addition some contributed papers retained after strong refereeing. Sections are I. Markov additive processes and regenerative systems II. Semi-Markov decision processes III. Algorithmic and computer-oriented approach IV. Semi-Markov models in economy and insurance V. Semi-Markov processes and reliability theory VI. Simulation and statistics for semi-Markov processes VII. Semi-Markov processes and queueing theory VIII. Branching IX. Applications in medicine X. Applications in other fields v PREFACE XI. A second bibliography on semi-Markov processes It is interesting to quote that sections IV to X represent a good sample of the main applications of semi-Markov processes i. e.
Markov Decision Processes with Their Applications
Author: Qiying Hu
Publisher: Springer Science & Business Media
ISBN: 0387369511
Category : Business & Economics
Languages : en
Pages : 305
Book Description
Put together by two top researchers in the Far East, this text examines Markov Decision Processes - also called stochastic dynamic programming - and their applications in the optimal control of discrete event systems, optimal replacement, and optimal allocations in sequential online auctions. This dynamic new book offers fresh applications of MDPs in areas such as the control of discrete event systems and the optimal allocations in sequential online auctions.
Publisher: Springer Science & Business Media
ISBN: 0387369511
Category : Business & Economics
Languages : en
Pages : 305
Book Description
Put together by two top researchers in the Far East, this text examines Markov Decision Processes - also called stochastic dynamic programming - and their applications in the optimal control of discrete event systems, optimal replacement, and optimal allocations in sequential online auctions. This dynamic new book offers fresh applications of MDPs in areas such as the control of discrete event systems and the optimal allocations in sequential online auctions.
Markov Decision Processes
Author: Martin L. Puterman
Publisher: John Wiley & Sons
ISBN: 1118625870
Category : Mathematics
Languages : en
Pages : 544
Book Description
The Wiley-Interscience Paperback Series consists of selected books that have been made more accessible to consumers in an effort to increase global appeal and general circulation. With these new unabridged softcover volumes, Wiley hopes to extend the lives of these works by making them available to future generations of statisticians, mathematicians, and scientists. "This text is unique in bringing together so many results hitherto found only in part in other texts and papers. . . . The text is fairly self-contained, inclusive of some basic mathematical results needed, and provides a rich diet of examples, applications, and exercises. The bibliographical material at the end of each chapter is excellent, not only from a historical perspective, but because it is valuable for researchers in acquiring a good perspective of the MDP research potential." —Zentralblatt fur Mathematik ". . . it is of great value to advanced-level students, researchers, and professional practitioners of this field to have now a complete volume (with more than 600 pages) devoted to this topic. . . . Markov Decision Processes: Discrete Stochastic Dynamic Programming represents an up-to-date, unified, and rigorous treatment of theoretical and computational aspects of discrete-time Markov decision processes." —Journal of the American Statistical Association
Publisher: John Wiley & Sons
ISBN: 1118625870
Category : Mathematics
Languages : en
Pages : 544
Book Description
The Wiley-Interscience Paperback Series consists of selected books that have been made more accessible to consumers in an effort to increase global appeal and general circulation. With these new unabridged softcover volumes, Wiley hopes to extend the lives of these works by making them available to future generations of statisticians, mathematicians, and scientists. "This text is unique in bringing together so many results hitherto found only in part in other texts and papers. . . . The text is fairly self-contained, inclusive of some basic mathematical results needed, and provides a rich diet of examples, applications, and exercises. The bibliographical material at the end of each chapter is excellent, not only from a historical perspective, but because it is valuable for researchers in acquiring a good perspective of the MDP research potential." —Zentralblatt fur Mathematik ". . . it is of great value to advanced-level students, researchers, and professional practitioners of this field to have now a complete volume (with more than 600 pages) devoted to this topic. . . . Markov Decision Processes: Discrete Stochastic Dynamic Programming represents an up-to-date, unified, and rigorous treatment of theoretical and computational aspects of discrete-time Markov decision processes." —Journal of the American Statistical Association
Handbook of Markov Decision Processes
Author: Eugene A. Feinberg
Publisher: Springer Science & Business Media
ISBN: 1461508053
Category : Business & Economics
Languages : en
Pages : 560
Book Description
Eugene A. Feinberg Adam Shwartz This volume deals with the theory of Markov Decision Processes (MDPs) and their applications. Each chapter was written by a leading expert in the re spective area. The papers cover major research areas and methodologies, and discuss open questions and future research directions. The papers can be read independently, with the basic notation and concepts ofSection 1.2. Most chap ters should be accessible by graduate or advanced undergraduate students in fields of operations research, electrical engineering, and computer science. 1.1 AN OVERVIEW OF MARKOV DECISION PROCESSES The theory of Markov Decision Processes-also known under several other names including sequential stochastic optimization, discrete-time stochastic control, and stochastic dynamic programming-studiessequential optimization ofdiscrete time stochastic systems. The basic object is a discrete-time stochas tic system whose transition mechanism can be controlled over time. Each control policy defines the stochastic process and values of objective functions associated with this process. The goal is to select a "good" control policy. In real life, decisions that humans and computers make on all levels usually have two types ofimpacts: (i) they cost orsavetime, money, or other resources, or they bring revenues, as well as (ii) they have an impact on the future, by influencing the dynamics. In many situations, decisions with the largest immediate profit may not be good in view offuture events. MDPs model this paradigm and provide results on the structure and existence of good policies and on methods for their calculation.
Publisher: Springer Science & Business Media
ISBN: 1461508053
Category : Business & Economics
Languages : en
Pages : 560
Book Description
Eugene A. Feinberg Adam Shwartz This volume deals with the theory of Markov Decision Processes (MDPs) and their applications. Each chapter was written by a leading expert in the re spective area. The papers cover major research areas and methodologies, and discuss open questions and future research directions. The papers can be read independently, with the basic notation and concepts ofSection 1.2. Most chap ters should be accessible by graduate or advanced undergraduate students in fields of operations research, electrical engineering, and computer science. 1.1 AN OVERVIEW OF MARKOV DECISION PROCESSES The theory of Markov Decision Processes-also known under several other names including sequential stochastic optimization, discrete-time stochastic control, and stochastic dynamic programming-studiessequential optimization ofdiscrete time stochastic systems. The basic object is a discrete-time stochas tic system whose transition mechanism can be controlled over time. Each control policy defines the stochastic process and values of objective functions associated with this process. The goal is to select a "good" control policy. In real life, decisions that humans and computers make on all levels usually have two types ofimpacts: (i) they cost orsavetime, money, or other resources, or they bring revenues, as well as (ii) they have an impact on the future, by influencing the dynamics. In many situations, decisions with the largest immediate profit may not be good in view offuture events. MDPs model this paradigm and provide results on the structure and existence of good policies and on methods for their calculation.
Markov Decision Processes
Decision Processes in Dynamic Probabilistic Systems
Author: A.V. Gheorghe
Publisher: Springer Science & Business Media
ISBN: 9400904932
Category : Mathematics
Languages : en
Pages : 370
Book Description
'Et moi - ... - si j'avait su comment en revenir. One service mathematics has rendered the je n'y serais point aile: human race. It has put common sense back where it belongs. on the topmost shelf next Jules Verne (0 the dusty canister labelled 'discarded non sense'. The series is divergent; therefore we may be able to do something with it. Eric T. Bell O. Heaviside Mathematics is a tool for thought. A highly necessary tool in a world where both feedback and non linearities abound. Similarly, all kinds of parts of mathematics serve as tools for other parts and for other sciences. Applying a simple rewriting rule to the quote on the right above one finds such statements as: 'One service topology has rendered mathematical physics .. .'; 'One service logic has rendered com puter science .. .'; 'One service category theory has rendered mathematics .. .'. All arguably true. And all statements obtainable this way form part of the raison d'etre of this series.
Publisher: Springer Science & Business Media
ISBN: 9400904932
Category : Mathematics
Languages : en
Pages : 370
Book Description
'Et moi - ... - si j'avait su comment en revenir. One service mathematics has rendered the je n'y serais point aile: human race. It has put common sense back where it belongs. on the topmost shelf next Jules Verne (0 the dusty canister labelled 'discarded non sense'. The series is divergent; therefore we may be able to do something with it. Eric T. Bell O. Heaviside Mathematics is a tool for thought. A highly necessary tool in a world where both feedback and non linearities abound. Similarly, all kinds of parts of mathematics serve as tools for other parts and for other sciences. Applying a simple rewriting rule to the quote on the right above one finds such statements as: 'One service topology has rendered mathematical physics .. .'; 'One service logic has rendered com puter science .. .'; 'One service category theory has rendered mathematics .. .'. All arguably true. And all statements obtainable this way form part of the raison d'etre of this series.
Naval Research Logistics Quarterly
Discrete-Time Markov Control Processes
Author: Onesimo Hernandez-Lerma
Publisher: Springer Science & Business Media
ISBN: 1461207290
Category : Mathematics
Languages : en
Pages : 223
Book Description
This book presents the first part of a planned two-volume series devoted to a systematic exposition of some recent developments in the theory of discrete-time Markov control processes (MCPs). Interest is mainly confined to MCPs with Borel state and control (or action) spaces, and possibly unbounded costs and noncompact control constraint sets. MCPs are a class of stochastic control problems, also known as Markov decision processes, controlled Markov processes, or stochastic dynamic pro grams; sometimes, particularly when the state space is a countable set, they are also called Markov decision (or controlled Markov) chains. Regardless of the name used, MCPs appear in many fields, for example, engineering, economics, operations research, statistics, renewable and nonrenewable re source management, (control of) epidemics, etc. However, most of the lit erature (say, at least 90%) is concentrated on MCPs for which (a) the state space is a countable set, and/or (b) the costs-per-stage are bounded, and/or (c) the control constraint sets are compact. But curiously enough, the most widely used control model in engineering and economics--namely the LQ (Linear system/Quadratic cost) model-satisfies none of these conditions. Moreover, when dealing with "partially observable" systems) a standard approach is to transform them into equivalent "completely observable" sys tems in a larger state space (in fact, a space of probability measures), which is uncountable even if the original state process is finite-valued.
Publisher: Springer Science & Business Media
ISBN: 1461207290
Category : Mathematics
Languages : en
Pages : 223
Book Description
This book presents the first part of a planned two-volume series devoted to a systematic exposition of some recent developments in the theory of discrete-time Markov control processes (MCPs). Interest is mainly confined to MCPs with Borel state and control (or action) spaces, and possibly unbounded costs and noncompact control constraint sets. MCPs are a class of stochastic control problems, also known as Markov decision processes, controlled Markov processes, or stochastic dynamic pro grams; sometimes, particularly when the state space is a countable set, they are also called Markov decision (or controlled Markov) chains. Regardless of the name used, MCPs appear in many fields, for example, engineering, economics, operations research, statistics, renewable and nonrenewable re source management, (control of) epidemics, etc. However, most of the lit erature (say, at least 90%) is concentrated on MCPs for which (a) the state space is a countable set, and/or (b) the costs-per-stage are bounded, and/or (c) the control constraint sets are compact. But curiously enough, the most widely used control model in engineering and economics--namely the LQ (Linear system/Quadratic cost) model-satisfies none of these conditions. Moreover, when dealing with "partially observable" systems) a standard approach is to transform them into equivalent "completely observable" sys tems in a larger state space (in fact, a space of probability measures), which is uncountable even if the original state process is finite-valued.
Markov Decision Processes in Practice
Author: Richard J. Boucherie
Publisher: Springer
ISBN: 3319477668
Category : Business & Economics
Languages : en
Pages : 563
Book Description
This book presents classical Markov Decision Processes (MDP) for real-life applications and optimization. MDP allows users to develop and formally support approximate and simple decision rules, and this book showcases state-of-the-art applications in which MDP was key to the solution approach. The book is divided into six parts. Part 1 is devoted to the state-of-the-art theoretical foundation of MDP, including approximate methods such as policy improvement, successive approximation and infinite state spaces as well as an instructive chapter on Approximate Dynamic Programming. It then continues with five parts of specific and non-exhaustive application areas. Part 2 covers MDP healthcare applications, which includes different screening procedures, appointment scheduling, ambulance scheduling and blood management. Part 3 explores MDP modeling within transportation. This ranges from public to private transportation, from airports and traffic lights to car parking or charging your electric car . Part 4 contains three chapters that illustrates the structure of approximate policies for production or manufacturing structures. In Part 5, communications is highlighted as an important application area for MDP. It includes Gittins indices, down-to-earth call centers and wireless sensor networks. Finally Part 6 is dedicated to financial modeling, offering an instructive review to account for financial portfolios and derivatives under proportional transactional costs. The MDP applications in this book illustrate a variety of both standard and non-standard aspects of MDP modeling and its practical use. This book should appeal to readers for practitioning, academic research and educational purposes, with a background in, among others, operations research, mathematics, computer science, and industrial engineering.
Publisher: Springer
ISBN: 3319477668
Category : Business & Economics
Languages : en
Pages : 563
Book Description
This book presents classical Markov Decision Processes (MDP) for real-life applications and optimization. MDP allows users to develop and formally support approximate and simple decision rules, and this book showcases state-of-the-art applications in which MDP was key to the solution approach. The book is divided into six parts. Part 1 is devoted to the state-of-the-art theoretical foundation of MDP, including approximate methods such as policy improvement, successive approximation and infinite state spaces as well as an instructive chapter on Approximate Dynamic Programming. It then continues with five parts of specific and non-exhaustive application areas. Part 2 covers MDP healthcare applications, which includes different screening procedures, appointment scheduling, ambulance scheduling and blood management. Part 3 explores MDP modeling within transportation. This ranges from public to private transportation, from airports and traffic lights to car parking or charging your electric car . Part 4 contains three chapters that illustrates the structure of approximate policies for production or manufacturing structures. In Part 5, communications is highlighted as an important application area for MDP. It includes Gittins indices, down-to-earth call centers and wireless sensor networks. Finally Part 6 is dedicated to financial modeling, offering an instructive review to account for financial portfolios and derivatives under proportional transactional costs. The MDP applications in this book illustrate a variety of both standard and non-standard aspects of MDP modeling and its practical use. This book should appeal to readers for practitioning, academic research and educational purposes, with a background in, among others, operations research, mathematics, computer science, and industrial engineering.
Control of Spatially Structured Random Processes and Random Fields with Applications
Author: Ruslan K. Chornei
Publisher: Springer Science & Business Media
ISBN: 038731279X
Category : Mathematics
Languages : en
Pages : 269
Book Description
This book is devoted to the study and optimization of spatiotemporal stochastic processes - processes which develop simultaneously in space and time under random influences. These processes are seen to occur almost everywhere when studying the global behavior of complex systems. The book presents problems and content not considered in other books on controlled Markov processes, especially regarding controlled Markov fields on graphs.
Publisher: Springer Science & Business Media
ISBN: 038731279X
Category : Mathematics
Languages : en
Pages : 269
Book Description
This book is devoted to the study and optimization of spatiotemporal stochastic processes - processes which develop simultaneously in space and time under random influences. These processes are seen to occur almost everywhere when studying the global behavior of complex systems. The book presents problems and content not considered in other books on controlled Markov processes, especially regarding controlled Markov fields on graphs.