Decision Processes in Dynamic Probabilistic Systems: Mathematics and its Applications, cartea 42
Autor A. V. Gheorgheen Limba Engleză Hardback – 31 iul 1990
Toate formatele și edițiile | Preț | Express |
---|---|---|
Paperback (1) | 631.27 lei 6-8 săpt. | |
SPRINGER NETHERLANDS – 22 sep 2011 | 631.27 lei 6-8 săpt. | |
Hardback (1) | 637.54 lei 6-8 săpt. | |
SPRINGER NETHERLANDS – 31 iul 1990 | 637.54 lei 6-8 săpt. |
Din seria Mathematics and its Applications
- Preț: 473.55 lei
- 18% Preț: 883.19 lei
- Preț: 381.89 lei
- Preț: 426.42 lei
- 23% Preț: 320.53 lei
- 20% Preț: 363.78 lei
- Preț: 423.80 lei
- 22% Preț: 320.55 lei
- Preț: 426.63 lei
- 15% Preț: 449.89 lei
- Preț: 419.10 lei
- 13% Preț: 350.83 lei
- 22% Preț: 332.60 lei
- Preț: 356.63 lei
- 18% Preț: 1102.61 lei
- Preț: 385.07 lei
- 15% Preț: 638.48 lei
- 15% Preț: 629.83 lei
- 15% Preț: 634.50 lei
- 15% Preț: 634.50 lei
- Preț: 381.68 lei
- Preț: 386.20 lei
- 15% Preț: 638.99 lei
- 20% Preț: 638.38 lei
- 15% Preț: 637.04 lei
- Preț: 374.69 lei
- Preț: 385.07 lei
- 15% Preț: 644.11 lei
- 15% Preț: 645.07 lei
- Preț: 386.36 lei
Preț: 637.54 lei
Preț vechi: 750.05 lei
-15% Nou
Puncte Express: 956
Preț estimativ în valută:
122.03€ • 127.96$ • 100.83£
122.03€ • 127.96$ • 100.83£
Carte tipărită la comandă
Livrare economică 29 ianuarie-12 februarie 25
Preluare comenzi: 021 569.72.76
Specificații
ISBN-13: 9780792305446
ISBN-10: 0792305442
Pagini: 376
Ilustrații: 376 p.
Greutate: 0.7 kg
Ediția:1990
Editura: SPRINGER NETHERLANDS
Colecția Springer
Seria Mathematics and its Applications
Locul publicării:Dordrecht, Netherlands
ISBN-10: 0792305442
Pagini: 376
Ilustrații: 376 p.
Greutate: 0.7 kg
Ediția:1990
Editura: SPRINGER NETHERLANDS
Colecția Springer
Seria Mathematics and its Applications
Locul publicării:Dordrecht, Netherlands
Public țintă
ResearchCuprins
1 Semi-Markov and Markov Chains.- 1.1 Definitions and basic properties.- 1.2 Algebraic and analytical methods in the study of Markovian systems.- 1.3 Transient and recurrent processes.- 1.4 Markovian populations.- 1.5 Partially observable Markov chains.- 1.6 Rewards and discounting.- 1.7 Models and applications.- 1.8 Dynamic-decision models for clinical diagnosis.- 2 Dynamic and Linear Programming.- 2.1 Discrete dynamic programming.- 2.2 A linear programming formulation and an algorithm for computation.- 3 Utility Functions and Decisions under Risk.- 3.1 Informational lotteries and axioms for utility functions.- 3.2 Exponential utility functions.- 3.3 Decisions under risk and uncertainty; event trees.- 3.4 Probability encoding.- 4 Markovian Decision Processes (Semi-Markov and Markov) with Complete Information (Completely Observable).- 4.1 Value iteration algorithm (the finite horizon case).- 4.2 Policy iteration algorithm (the finite horizon optimization).- 4.3 Policy iteration with discounting.- 4.4 Optimization algorithm using linear programming.- 4.5 Risk-sensitive decision processes.- 4.6 On eliminating sub-optimal decision alternatives in Markov and semi-Markov decision processes.- 5 Partially Observable Markovian Decision Processes.- 5.1 Finite horizon partially observable Markov decision processes.- 5.2 The infinite horizon with discounting for partially observable Markov decision processes.- 5.3 A useful policy iteration algorithm, for discounted (? < 1) partially observable Markov decision processes.- 5.4 The infinite horizon without discounting for partially observable Markov processes.- 5.5 Partially observable semi-Markov decision processes.- 5.6 Risk-sensitive partially observable Markov decision processes.- 6 Policy Constraints in Markov DecisionProcesses.- 6.1 Methods of investigating policy costraints in Markov decision processes.- 6.2 Markov decision processes with policy constraints.- 6.3 Risk-sensitive Markov decision process with policy constraints.- 7 Applications.- 7.1 The emergency repair control for electrical power systems.- 7.2 Stochastic models for evaluation of inspection and repair schedules [2].- 7.3 A Markovian dicision model for clinical diagnosis and treatment applied to the respiratory system.