图书信息:

丛 书 名:Stochastic Modelling and Applied Probability
书  名:Stochastic Approximation and Recursive Algorithms and Applications (Second Edition)
作  者:Harold J. Kushner, G. George Yin
出 版 社:Springer
出版日期:2003年7月
语  种:英文
I S B N:978-0-387-00894-3
页  数:478

内容简介:   

  This revised and expanded second edition presents a thorough development of the modern theory of stochastic approximation or recursive stochastic algorithms for both constrained and unconstrained problems. There is a complete development of both probability one and weak convergence methods for very general noise processes. The proofs of convergence use the ODE method, the most powerful to date. The assumptions and proof methods are designed to cover the needs of recent applications. The development proceeds from simple to complex problems, allowing the underlying ideas to be more easily understood. Rate of convergence, iterate averaging, high-dimensional problems, stability-ODE methods, two time scale, asynchronous and decentralized algorithms, state-dependent noise, stability methods for correlated noise, perturbed test function methods, and large deviations methods are covered. Many motivating examples from learning theory, ergodic cost problems for discrete event systems, wireless communications, adaptive control, signal processing, and elsewhere illustrate the applications of the theory.

英文目录:

Preface and Introduction
1 Introduction: Applications and Issues
  1.0 Outline of Chapter
  1.1 The Robbins–Monro Algorithm
    1.1.1 Introduction
    1.1.2 Finding the Zeros of an Unknown Function
    1.1.3 Best Linear Least Squares Fit
    1.1.4 Minimization by Recursive Monte Carlo
  1.2 The Kiefer–Wolfowitz Procedure
    1.2.1 The Basic Procedure
    1.2.2 Random Directions
  1.3 Extensions of the Algorithms
    1.3.1 A Variance Reduction Method
    1.3.2 Constraints
    1.3.3 Averaging of the Iterates: “Polyak Averaging”
    1.3.4 Averaging the Observations
    1.3.5 Robust Algorithms
    1.3.6 Nonexistence of the Derivative at Some θ
    1.3.7 Convex Optimization and Subgradients
  1.4 A Lagrangian Algorithm for Constrained Function Minimization
2 Applications
  2.0 Outline of Chapter
  2.1 An Animal Learning Model
  2.2 A Neural Network
  2.3 State-Dependent Noise
  2.4 Learning Optimal Controls
    2.4.1 Q-Learning
    2.4.2 Approximating a Value Function
    2.4.3 Parametric Optimization of a Markov Chain Control Problem
  2.5 Optimization of a GI/G/1 Queue
    2.5.1 Derivative Estimation and Infinitesimal Perturbation Analysis: A Brief Review
    2.5.2 The Derivative Estimate for the Queueing Problem
  2.6 Passive Stochastic Approximation
  2.7 Learning in Repeated Stochastic Games
3 Signal Processing, Communications, and Control
  3.0 Outline of Chapter
  3.1 Parameter Identification and Tracking
    3.1.1 The Classical Model
    3.1.2 ARMA and ARMAX Models
  3.2 Tracking Time Varying Systems
    3.2.1 The Algorithm
    3.2.2 Some Data
  3.3 Feedback and Averaging
  3.4 Applications in Communications Theory
    3.4.1 Adaptive Noise Cancellation and Disturbance Rejection
    3.4.2 Adaptive Equalizers
    3.4.3 An ARMA Model, with a Training Sequence
  3.5 Adaptive Antennas and Mobile Communications
  3.6 Proportional Fair Sharing
4 Mathematical Background
  4.0 Outline of Chapter
  4.1 Martingales and Inequalities
  4.2 Ordinary Differential Equations
    4.2.1 Limits of a Sequence of Continuous Functions
    4.2.2 Stability of Ordinary Differential Equations
  4.3 Projected ODE
  4.4 Cooperative Systems and Chain Recurrence
    4.4.1 Cooperative Systems
    4.4.2 Chain Recurrence
  4.5 Stochastic Stability
5 Convergence w.p.1: Martingale Difference Noise
  5.0 Outline of Chapter
  5.1 Truncated Algorithms: Introduction
  5.2 The ODE Method
    5.2.1 Assumptions and the Main Convergence Theorem
    5.2.2 Convergence to Chain Recurrent Points
  5.3 A General Compactness Method
    5.3.1 The Basic Convergence Theorem
    5.3.2 Suffcient Conditions for the Rate of Change Condition
    5.3.3 The Kiefer–Wolfowitz Algorithm
  5.4 Stability and Combined Stability–ODE Methods
    5.4.1 A Liapunov Function Method for Convergence
    5.4.2 Combined Stability–ODE Methods
  5.5 Soft Constraints
  5.6 Random Directions, Subgradients, and Differential Inclusions
  5.7 Animal Learning and Pattern Classification
    5.7.1 The Animal Learning Problem
    5.7.2 The Pattern Classification Problem
  5.8 Non-Convergence to Unstable Points
6 Convergence w.p.1: Correlated Noise
  6.0 Outline of Chapter
  6.1 A General Compactness Method
    6.1.1 Introduction and General Assumptions
    6.1.2 The Basic Convergence Theorem
    6.1.3 Local Convergence Results
  6.2 Suffcient Conditions
  6.3 Perturbed State Criteria
    6.3.1 Perturbed Iterates
    6.3.2 General Conditions for the Asymptotic Rate of Change
    6.3.3 Alternative Perturbations
  6.4 Examples of State Perturbation
  6.5 Kiefer–Wolfowitz Algorithms
  6.6 State-Dependent Noise
  6.7 Stability-ODE Methods
  6.8 Differential Inclusions
  6.9 Bounds on Escape Probabilities
  6.10 Large Deviations
    6.10.1 Two-Sided Estimates
    6.10.2 Upper Bounds
    6.10.3 Bounds on Escape Times
7 Weak Convergence: Introduction
  7.0 Outline of Chapter
  7.1 Introduction
  7.2 Martingale Difference Noise
  7.3 Weak Convergence
    7.3.1 Definitions
    7.3.2 Basic Convergence Theorems
  7.4 Martingale Limits
    7.4.1 Verifying that a Process Is a Martingale
    7.4.2 The Wiener Process
    7.4.3 Perturbed Test Functions
8 Weak Convergence Methods for General Algorithms
  8.0 Outline of Chapter
  8.1 Exogenous Noise
  8.2 Convergence: Exogenous Noise
    8.2.1 Constant Step Size: Martingale Difference Noise
    8.2.2 Correlated Noise
    8.2.3 Step Size
    8.2.4 Random
    8.2.5 Differential Inclusions
    8.2.6 Time-Dependent ODEs
  8.3 The Kiefer–Wolfowitz Algorithm
    8.3.1 Martingale Difference Noise
    8.3.2 Correlated Noise
  8.4 State-Dependent Noise
    8.4.1 Constant Step Size
    8.4.2 Decreasing Step Size
    8.4.3 The Invariant Measure Method
    8.4.4 General Forms of the Conditions
    8.4.5 Observations Depending on the Past of the Iterate Sequence or Working Directly with
  8.5 Unconstrained Algorithms and the ODE-Stability Method
  8.6 Two-Time-Scale Problems
    8.6.1 The Constrained Algorithm
    8.6.2 Unconstrained Algorithms: Stability
9 Applications: Proofs of Convergence
  9.0 Outline of Chapter
  9.1 Introduction
    9.1.1 General Comments
    9.1.2 A Simple Illustrative SDE Example
  9.2 A SDE Example
  9.3 A Discrete Example: A GI/G/1 Queue
  9.4 Signal Processing Problems
  9.5 Proportional Fair Sharing
10 Rate of Convergence
  10.0 Outline of Chapter
  10.1 Exogenous Noise: Constant Step Size
    10.1.1 Martingale Difference Noise
    10.1.2 Correlated Noise
  10.2 Exogenous Noise: Decreasing Step Size
    10.2.1 Martingale Difference Noise
    10.2.2 Optimal Step Size Sequence
    10.2.3 Correlated Noise
  10.3 Kiefer–Wolfowitz Algorithm
    10.3.1 Martingale Difference Noise
    10.3.2 Correlated Noise
  10.4 Tightness: W.P.1 Convergence
    10.4.1 Martingale Difference Noise: Robbins–Monro Algorithm
    10.4.2 Correlated Noise
    10.4.3 Kiefer–Wolfowitz Algorithm
  10.5 Tightness: Weak Convergence
    10.5.1 Unconstrained Algorithm
    10.5.2 Local Methods for Proving Tightness
  10.6 Weak Convergence to a Wiener Process
  10.7 Random Directions
    10.7.1 Comparison of Algorithms
  10.8 State-Dependent Noise
  10.9 Limit Point on the Boundary
11 Averaging of the Iterates
  11.0 Outline of Chapter
  11.1 Minimal Window of Averaging
    11.1.1 Robbins–Monro Algorithm: Decreasing Step Size
    11.1.2 Constant Step Size
    11.1.3 Averaging with Feedback and Constant Step Size
    11.1.4 Kiefer–Wolfowitz Algorithm
  11.2 A Two-Time-Scale Interpretation
  11.3 Maximal Window of Averaging
  11.4 The Parameter Identification Problem
12 Distributed/Decentralized and Asynchronous Algorithms
  12.0 Outline of Chapter
  12.1 Examples
    12.1.1 Introductory Comments
    12.1.2 Pipelined Computations
    12.1.3 A Distributed and Decentralized Network Model
    12.1.4 Multiaccess Communications
  12.2 Real-Time Scale: Introduction
  12.3 The Basic Algorithms
    12.3.1 Constant Step Size: Introduction
    12.3.2 Martingale Difference Noise
    12.3.3 Correlated Noise
    12.3.4 Analysis for and
  12.4 Decreasing Step Size
  12.5 State-Dependent Noise
  12.6 Rate of Convergence
  12.7 Stability and Tightness of the Normalized Iterates
    12.7.1 Unconstrained Algorithms
  12.8 Convergence for Q-Learning: Discounted Cost
References
Symbol Index
Subject Index


  控制理论专业委员会 ©2011-2015 版权所有

中国自动化学会 控制理论专业委员会
电话:86-10-82541403;Email:tcct@iss.ac.cn