Cantitate/Preț
Produs

Optimization and Games for Controllable Markov Chains: Numerical Methods with Application to Finance and Engineering: Studies in Systems, Decision and Control, cartea 504

Autor Julio B. Clempner, Alexander Poznyak
en Limba Engleză Hardback – 14 dec 2023
This book considers a class of ergodic finite controllable Markov's chains. The main idea behind the method, described in this book, is to develop the original discrete optimization problems (or game models) in the space of randomized formulations, where the variables stand in for the distributions (mixed strategies or preferences) of the original discrete (pure) strategies in the use. The following suppositions are made: a finite state space, a limited action space, continuity of the probabilities and rewards associated with the actions, and a necessity for accessibility. These hypotheses lead to the existence of an optimal policy. The best course of action is always stationary. It is either simple (i.e., nonrandomized stationary) or composed of two nonrandomized policies, which is equivalent to randomly selecting one of two simple policies throughout each epoch by tossing a biased coin. As a bonus, the optimization procedure just has to repeatedly solve the time-average dynamic programming equation, making it theoretically feasible to choose the optimum course of action under the global restriction. In the ergodic cases the state distributions, generated by the corresponding transition equations, exponentially quickly converge to their stationary (final) values. This makes it possible to employ all widely used optimization methods (such as Gradient-like procedures, Extra-proximal method, Lagrange's multipliers, Tikhonov's regularization), including the related numerical techniques. In the book we tackle different problems and theoretical Markov models like controllable and ergodic Markov chains, multi-objective Pareto front solutions, partially observable Markov chains, continuous-time Markov chains, Nash equilibrium and Stackelberg equilibrium, Lyapunov-like function in Markov chains, Best-reply strategy, Bayesian incentive-compatible mechanisms, Bayesian Partially Observable Markov Games, bargaining solutions for Nash and Kalai-Smorodinsky formulations, multi-traffic signal-control synchronization problem, Rubinstein's non-cooperative bargaining solutions, the transfer pricing problem as bargaining.

Citește tot Restrânge

Din seria Studies in Systems, Decision and Control

Preț: 88054 lei

Preț vechi: 107384 lei
-18% Nou

Puncte Express: 1321

Preț estimativ în valută:
16850 17768$ 14028£

Carte tipărită la comandă

Livrare economică 10-24 ianuarie 25

Preluare comenzi: 021 569.72.76

Specificații

ISBN-13: 9783031435744
ISBN-10: 3031435745
Pagini: 332
Ilustrații: XVIII, 332 p. 99 illus., 94 illus. in color.
Dimensiuni: 155 x 235 mm
Greutate: 0.67 kg
Ediția:1st ed. 2024
Editura: Springer Nature Switzerland
Colecția Springer
Seria Studies in Systems, Decision and Control

Locul publicării:Cham, Switzerland

Cuprins

Controllable Markov Chains.- Multiobjective Control.- Partially Observable Markov Chains.- Continuous-Time Markov Chains.- Nash and Stackelberg Equilibrium.- Best-Reply Strategies in Repeated Games.- Mechanism design.- Joint Observer and Mechanism Design.- Bargaining Games or How to Negotiate.- Multi-Traffic Signal-Control Synchronization.- Non-cooperative bargaining with unsophisticated agents.- Transfer Pricing as Bargaining.- Index.

Textul de pe ultima copertă

This book considers a class of ergodic finite controllable Markov's chains. The main idea behind the method, described in this book, is to develop the original discrete optimization problems (or game models) in the space of randomized formulations, where the variables stand in for the distributions (mixed strategies or preferences) of the original discrete (pure) strategies in the use. The following suppositions are made: a finite state space, a limited action space, continuity of the probabilities and rewards associated with the actions, and a necessity for accessibility. These hypotheses lead to the existence of an optimal policy. The best course of action is always stationary. It is either simple (i.e., nonrandomized stationary) or composed of two nonrandomized policies, which is equivalent to randomly selecting one of two simple policies throughout each epoch by tossing a biased coin. As a bonus, the optimization procedure just has to repeatedly solve the time-average dynamic programming equation, making it theoretically feasible to choose the optimum course of action under the global restriction. In the ergodic cases the state distributions, generated by the corresponding transition equations, exponentially quickly converge to their stationary (final) values. This makes it possible to employ all widely used optimization methods (such as Gradient-like procedures, Extra-proximal method, Lagrange's multipliers, Tikhonov's regularization), including the related numerical techniques. In the book we tackle different problems and theoretical Markov models like controllable and ergodic Markov chains, multi-objective Pareto front solutions, partially observable Markov chains, continuous-time Markov chains, Nash equilibrium and Stackelberg equilibrium, Lyapunov-like function in Markov chains, Best-reply strategy, Bayesian incentive-compatible mechanisms, Bayesian Partially Observable Markov Games, bargaining solutions for Nash and Kalai-Smorodinsky formulations, multi-traffic signal-control synchronization problem, Rubinstein's non-cooperative bargaining solutions, the transfer pricing problem as bargaining.

Caracteristici

Considers a class of ergodic finite controllable Markov's chains Develops the original discrete optimization problems (or game models) in the space of randomized formulations Focuses on how Markov chain theory solves different problems that arise in the real world