Mathematical Theory of Control Systems Design

Author:   V.N. Afanasiev ,  V. Kolmanovskii ,  V.R. Nosov
Publisher:   Springer
Edition:   1996 ed.
Volume:   341
ISBN:  

9780792337249


Pages:   672
Publication Date:   31 January 1996
Format:   Hardback
Availability:   In Print   Availability explained
This item will be ordered in for you from one of our suppliers. Upon receipt, we will promptly dispatch it out to you. For in store availability, please contact us.

Our Price $366.96 Quantity:  
Add to Cart

Share |

Mathematical Theory of Control Systems Design


Add your own review!

Overview

The topics covered in this work are spread over an introduction and four parts. Each chapter concludes with a brief review of the main results and formulae, and each part ends with an exercise section. Part One treats the fundamentals of modern stability theory. Part Two is devoted to the optimal control of deterministic systems. Part Three is concerned with problems of the control of systems under random disturbances of their parameters, and Part Four provides an outline of modern numerical methods of control theory. The many examples included illustrate the main assertions, teaching the reader the skills needed to construct models of relevant phenomena, to design nonlinear control systems, to explain the qualitative differences between various classes of control systems, and to apply what they have learned to the investigation of particular systems.

Full Product Details

Author:   V.N. Afanasiev ,  V. Kolmanovskii ,  V.R. Nosov
Publisher:   Springer
Imprint:   Springer
Edition:   1996 ed.
Volume:   341
Dimensions:   Width: 15.60cm , Height: 3.80cm , Length: 23.40cm
Weight:   2.540kg
ISBN:  

9780792337249


ISBN 10:   0792337247
Pages:   672
Publication Date:   31 January 1996
Audience:   College/higher education ,  Professional and scholarly ,  Postgraduate, Research & Scholarly ,  Professional & Vocational
Format:   Hardback
Publisher's Status:   Active
Availability:   In Print   Availability explained
This item will be ordered in for you from one of our suppliers. Upon receipt, we will promptly dispatch it out to you. For in store availability, please contact us.

Table of Contents

I. Continuous and Discrete Deterministic Systems.- II. Stability of Stochastic Systems.- III. Description of Control Problems.- IV. The Classical Calculus of Variations and Optimal Control.- V. The Maximum Principle.- VI. Linear Control Systems.- VII. Dynamic Programming Approach. Sufficient Conditions for Optimal Control.- VIII. Some Additional Topics of Optimal Control Theory.- IX. Control of Stochastic Systems. Problem Statements and Investigation Techniques.- X. Optimal Control on a Time Interval of Random Duration.- XI. Optimal Estimation of the State of the System.- XII. Optimal Control of the Observation Process.- XIII. Linear Time-Invariant Control Systems.- XIV. Numerical Methods for the Investigation of Nonlinear Control Systems.- XV. Numerical Design of Optimal Control Systems.- General References.

Reviews

Author Information

Tab Content 6

Author Website:  

Customer Reviews

Recent Reviews

No review item found!

Add your own review!

Countries Available

All regions
Latest Reading Guide

MRG2025CC

 

Shopping Cart
Your cart is empty
Shopping cart
Mailing List