Model Predictive Control : Fundamentals and Practice

個数:
  • 予約

Model Predictive Control : Fundamentals and Practice

  • 現在予約受付中です。出版後の入荷・発送となります。
    重要:表示されている発売日は予定となり、発売が延期、中止、生産限定品で商品確保ができないなどの理由により、ご注文をお取消しさせていただく場合がございます。予めご了承ください。

    ●3Dセキュア導入とクレジットカードによるお支払いについて
  • 【入荷遅延について】
    世界情勢の影響により、海外からお取り寄せとなる洋書・洋古書の入荷が、表示している標準的な納期よりも遅延する場合がございます。
    おそれいりますが、あらかじめご了承くださいますようお願い申し上げます。
  • ◆画像の表紙や帯等は実物とは異なる場合があります。
  • ◆ウェブストアでの洋書販売価格は、弊社店舗等での販売価格とは異なります。
    また、洋書販売価格は、ご注文確定時点での日本円価格となります。
    ご注文確定後に、同じ洋書の販売価格が変動しても、それは反映されません。
  • 製本 Hardcover:ハードカバー版/ページ数 592 p.
  • 言語 ENG
  • 商品コード 9781394333295

Full Description

Master advanced control methods bridging academic theory and industrial practice

Model Predictive Control: Fundamentals and Practice walks engineers through the transition from academic study to industrial application of advanced process control. This comprehensive text connects current model predictive control (MPC) theory to its industrial origins and classical linear control methods, providing the foundations necessary for effective real-world application.

This book's three-part structure guides readers from basic industrial algorithms through linear systems fundamentals to advanced MPC topics. It clarifies equivalences between MPC and Linear-Quadratic optimal control, and between Moving Horizon Estimation and Kalman filtering. It also includes practical coverage of system identification. The book balances up-to-date theory with hands-on applications and maintains accessibility without sacrificing mathematical rigor.

Readers will learn to:

Effectively transition theoretical knowledge into practical control applications for complex processes
Understand connections between MPC and classical optimal control methods through clear detailed explanations
Master system identification techniques essential for developing accurate process models
Explore nonlinear MPC and the innovative Repetitive MPC for advanced real-world control challenges
Apply concepts through curated sample problems designed to enhance practical understanding and implementation skills

This is an ideal graduate-level textbook and essential reference for practicing engineers seeking to master advanced control strategies. It balances authoritative theoretical explanation with practical application, preparing readers to solve real-world control problems.

Contents

Contents

Acknowledgments

Preface

1 Introduction
1.1 What's MPC?
1.2 Why MPC?
1.2.1 Economic Drivers of APC/MPC
1.2.2 Economic Advantages of MPC vs. Other Tools
1.3 Historical Overview
1.3.1 Early Computer Control
1.3.2 The Pioneers
1.3.3 Adoption Growth
1.4 Impact of MPC on Control Research
1.4.1 Early Theoretical Developments
1.4.2 State-Space Model Formulation and Stability Results
1.4.3 Other Theoretical Developments
1.4.4 Lessons Learned Along the MPC Journey
1.5 A Typical Industrial Control Problem
1.6 Organization of This Book
Exercises

2 Step Response Modeling and Identification
2.1 Linear Time Invariant Systems
2.2 Impulse / Step Response Models
2.2.1 Impulse Response Models
2.2.2 Step Response Models
2.3 Multi-Step Prediction
2.3.1 Recursive Multi-Step Prediction Based on FIR Model
2.3.2 Recursive Multi-Step Prediction Based on Step-Response Model
2.3.3 Multivariable Generalization
2.4 Examples
2.5 Identification
2.5.1 Settling Time
2.5.2 Sampling Time
2.5.3 Choice of the Input Signal for Experimental Identification
2.5.4 The Linear Least Squares Problem
2.5.5 Linear Least Squares Identification
Exercises

3 Dynamic Matrix Control - The Basic Algorithm
3.1 The Concept of Moving Horizon Control
3.2 Multi-Step Prediction
3.3 Objective Function
3.4 Constraints
3.4.1 Manipulated Variable Constraints
3.4.2 Manipulated Variable Rate Constraints
3.4.3 Output Variable Constraints
3.4.4 Combined Constraints
3.5 Quadratic Programming Solution of the Control Problem
3.5.1 Quadratic Programs
3.5.2 Formulation as a Quadratic Program
3.6 Implementation
3.6.1 Moving Horizon Algorithm
3.6.2 DMC Examples
3.6.3 Efficient Solutions to the QP
3.6.4 Proper Constraint Formulation
3.6.5 Choice of Horizon Length
3.6.6 Input Blocking
3.6.7 Filtering of the Feedback Signal
3.7 Examples: Analysis and Guidelines
3.7.1 Unconstrained SISO Systems
3.7.2 Constrained SISO Systems
3.7.3 MIMO System with Strong Gain Directionality
3.7.4 Constrained MIMO Systems
3.7.5 Conclusions and General Tuning Guidelines
3.8 Case Study: Control of "Shell Heavy Oil Fractionator" using Dynamic Matrix Control
3.8.1 Heavy Oil Fractionator: Background
3.8.2 Control Structure Description
Exercises

4 Dynamic Matrix Control - Extensions and Variations

4.1 Features Found in Other Industrial Algorithms
4.1.1 Reference Trajectories
4.1.2 Coincidence Points
4.1.3 The Funnel Approach
4.1.4 Use of Other Norms
4.1.5 Input Parameterization
4.1.6 Model Conditioning
4.1.7 Prioritization of CVs and MVs

4.2 Connection with Internal Model Control

4.3 Some Possible Enhancements to DMC
4.3.1 Closed-Loop Update of Model State
4.3.2 Integrating Dynamics
4.3.3 Noise Filter
4.3.4 Bi-Level Optimization
4.3.5 Product Property Estimation

Exercises

5 Linear Time Invariant System Models

5.1 Sampling and Reconstruction
5.1.1 Introduction to Digital Control
5.1.2 Sampling
5.1.3 Aliasing
5.1.4 Reconstruction

5.2 Introduction to z-transform

5.3 Transfer Function Models
5.3.1 Continuous Time
5.3.2 Discrete Time
5.3.3 Transfer Matrix
5.3.4 Converting Continuous Transfer Function to Discrete Transfer Function
5.3.5 Stability and Implications of Poles
5.3.6 Gain, Frequency Response

5.4 State-Space Model
5.4.1 Continuous Time
5.4.2 Discrete Time
5.4.3 Converting Continuous- to Discrete-Time System

5.5 Conversion Between Discrete-Time Models
5.5.1 Representing State-Space System as Transfer Function
5.5.2 Realization of Transfer Function as State-Space System
5.5.3 Impulse and Step Responses of State-Space System
5.5.4 Derivation of Transfer Matrix from Impulse Response
5.5.5 From Impulse / Step Response to State-Space Model

Exercises

6 Discrete-Time State Space Models

6.1 State-Coordinate Transformation

6.2 Stability
6.2.1 System Poles and Characteristic Equation
6.2.2 Stability
6.2.3 Lyapunov Equation

6.3 Controllability, Reachability, and Stabilizability
6.3.1 Definitions
6.3.2 Conditions for Reachability
6.3.3 Coordinate Transformation

6.4 Observability, Reconstructability, and Detectability
6.4.1 Definitions
6.4.2 Conditions for Observability
6.4.3 Coordinate Transformation

6.5 Kalman's Decomposition and Minimal Realization
6.5.1 Kalman's Decomposition
6.5.2 Minimal Realization

6.6 Disturbance Modeling
6.6.1 Linear Stochastic System Model for Stationary Processes
6.6.2 Stochastic System Models for Processes with Nonstationary Behavior
6.6.3 Models for Estimation and Control

Exercises

7 State Estimation

7.1 Linear Estimator Structure

7.2 Observer Pole Placement

7.3 Kalman Filter
7.3.1 Derivation of the Optimal Filter Gain Matrix
7.3.2 Correlated Noise Case
7.3.3 Stability of Kalman Filter

7.4 Extensions
7.4.1 Inferential Estimation
7.4.2 Non-stationary (Integrating) Noise
7.4.3 Time-Varying System
7.4.4 Periodically Time-Varying System
7.4.5 Measurement Delays

7.5 Least Squares Formulation of State Estimation
7.5.1 Batch Least Squares Formulation
7.5.2 Recursive Solution and Equivalence with Kalman Filter
7.5.3 Use of Moving Estimation Window

Exercises

8 Unconstrained Quadratic Optimal Control

8.1 Linear State Feedback Controller Design

8.2 Finite Horizon Quadratic Optimal Control
8.2.1 Open‑Loop Optimal Solution via Least Squares
8.2.2 State Feedback Solution via Dynamic Programming
8.2.3 Comparison of the Two Approaches

8.3 Infinite Horizon Quadratic Optimal Control
8.3.1 Optimal State Feedback Law: Asymptotic Solution of the Finite Horizon Problem
8.3.2 Receding Horizon Implementation of the Finite Horizon Solution
8.3.3 Equivalence Between Finite and Infinite Horizon Problems

8.4 Analysis
8.4.1 State Feedback Case
8.4.2 Output Feedback Case
8.4.3 Setpoint Tracking and Disturbance Rejection

8.5 Stochastic LQ Control
8.5.1 Finite Horizon Problem
8.5.2 Output Feedback LQ Control

Exercises

9 Constrained Quadratic Optimal Control

9.1 Finite Horizon Problem

9.2 Infinite Horizon Problem
9.2.1 Options for Re‑formulation as an Equivalent Finite‑Horizon Problem
9.2.2 Comparison of Various Options

9.3 Constraint Softening

9.4 Derivation of an Explicit Form of the Optimal Control Law via Multi‑Parametric Programming

9.5 Analysis
9.5.1 Stability Concepts and Lyapunov's Direct Method
9.5.2 State Feedback Case
9.5.3 Output Feedback Case

9.6 Stochastic Case (*)

Exercises

10 System Identification

10.1 Problem Overview

10.2 Model Structures
10.2.1 Finite Impulse Response Model
10.2.2 Structures for Parametric Identification
10.2.3 Key Issues in Parametric Models

10.3 Parametric Identification Methods
10.3.1 Prediction Error Method
10.3.2 Properties of Linear Least Squares Identification
10.3.3 Persistency of Excitation
10.3.4 Frequency‑Domain Bias Distribution Under PEM
10.3.5 Parameter Estimation via Statistical Methods ()
10.3.6 Other Methods ()

10.4 Nonparametric Identification
10.4.1 Impulse Response Identification
10.4.2 Frequency Response Identification (*)

10.5 Subspace Identification
10.5.1 The Basic Method
10.5.2 Analysis and Discussion

10.6 Practice of System Identification: A User's Perspective
10.6.1 Experiment Design
10.6.2 PRBS Signals
10.6.3 Data Pre‑Processing
10.6.4 Model Fitting and Validation
10.6.5 Model Quality Assessment and an Integrated Framework

Exercises

11 Linear MPC: State Space Formulation

11.1 Motivation

11.2 Model Construction
11.2.1 Model Structure for State‑Space MPC
11.2.2 Stochastic System Model with Output Disturbance Only
11.2.3 Stochastic System Model with State and Output Disturbances
11.2.4 Summary

11.3 Deterministic State Space MPC
11.3.1 State Regulation Problem
11.3.2 Constraints
11.3.3 Offset‑Free Output Tracking and Regulation Problem

11.4 MPC with State Estimation
11.4.1 State Estimation Using Kalman Filter
11.4.2 Control Calculation Using State Estimate
11.4.3 MPC with Output Disturbance Only
11.4.4 MPC with State Disturbance Model
11.4.5 MPC with Full Disturbance Model
11.4.6 Tracking a Setpoint Trajectory
11.4.7 Constraint Softening

11.5 Inferential Control
11.5.1 Problem Formulation
11.5.2 Infrequent Primary Measurements
11.5.3 Handling Measurement Delays in Primary Measurements

11.6 Sequential Linearization‑Based MPC (for Nonlinear Systems)
11.6.1 Model Construction
11.6.2 Extended Kalman Filter
11.6.3 Multi‑Step Prediction
11.6.4 Objective Function and Constraints
11.6.5 Implementation of Sequential Linearization‑Based MPC

Exercises

12 Nonlinear MPC

12.1 Introduction

12.2 NMPC Formulation

12.3 Solution via Nonlinear Programming
12.3.1 Elements of NLP Formulations
12.3.2 Nonlinear Programming Solvers

12.4 Stability and Other Properties
12.4.1 Invariant Set and Output Admissible Set
12.4.2 Cost‑To‑Go and Terminal Penalty
12.4.3 Establishing Closed‑Loop Stability
12.4.4 Implementation: Quasi‑Infinite Horizon MPC

12.5 Nonlinear State Estimation
12.5.1 Extended Kalman Filter
12.5.2 Moving Horizon Estimation for Nonlinear State Estimation

12.6 Case Study

12.7 Conclusions and Future Directions

Exercises

13 Repetitive MPC for Batch and Periodic Systems

13.1 Introduction
13.1.1 Historical Background

13.2 General Framework
13.2.1 Problem Formulation
13.2.2 Limitations of Conventional Feedback Control for Periodic Processes

13.3 Iterative Learning Model Predictive Control for Batch Systems
13.3.1 Notations
13.3.2 "Run‑To‑Run" IL‑MPC Method for an Unconstrained System
13.3.3 "Run‑To‑Run" IL‑MPC Method for a Constrained System
13.3.4 Real‑Time‑Feedback IL‑MPC Method for an Unconstrained System
13.3.5 Real‑Time‑Feedback IL‑MPC Method for the Constrained System

13.4 Repetitive Model Predictive Control for Continuous Systems with Periodic Operations
13.4.1 Notations
13.4.2 "Run‑To‑Run" R‑MPC Method for an Unconstrained System
13.4.3 "Run‑To‑Run" R‑MPC Methods for the Constrained System
13.4.4 Real‑Time‑Feedback‑Based R‑MPC Methods for the Unconstrained System
13.4.5 Real‑Time‑Feedback‑Based R‑MPC Methods for the Constrained System

13.5 Future Outlook

Exercises

Appendix A: Review of Linear Transformation

A.1 Vector Space
A.1.1 Definition
A.1.2 Dimension of a Vector Space
A.1.3 Linear Independence of Vectors
A.1.4 Basis
A.1.5 Subspace
A.1.6 Union, Intersection, Independence, and Internal Sum
A.1.7 Change of Basis

A.2 Linear Operator
A.2.1 Definition
A.2.2 Matrix Representation

A.2.3 Change of Basis for Linear Operators
A.2.4 Null Space and Image Space
A.2.5 Inverse Operator
A.2.6 Injection, Surjection, Bijection
A.2.7 Inner Product Space
A.2.8 Orthogonal Vectors and Orthonormal Basis
A.2.9 Change of Basis to Orthonormal Basis
A.2.10 Orthogonal Matrix
A.2.11 Projection, Orthogonal Projection

A.3 Matrix Algebra

A.3.1 Eigenvalues, Eigenvectors
A.3.2 Computing Eigenvalues and Eigenvectors
A.3.3 Jordan Decomposition and Its Applications
A.3.4 Singular Value Decomposition
A.3.5 Cayley‑Hamilton Theorem
A.3.6 Matrix Function
A.3.7 Vector Norms
A.3.8 Matrix Norm
A.3.9 Positive (Negative) Definiteness and Semi‑Definiteness

A.4 Exercises

B.1 Random Variables

B.1.1 Introduction
B.1.2 Basic Probability Concepts
B.1.3 Statistics

B.2 Stochastic Process

B.2.1 Basic Probability Concepts

C Model Reduction

C.1 Model Reduction Problem
C.2 Hankel Matrix and Hankel Singular Values
C.3 Balanced Realization and Truncation
C.4 Application to FIR Models

D.1 Kalman Filter as the Bayesian Estimator for Gaussian Systems

D.2 Moving Horizon Estimation: Recursive Solution to the Unconstrained Linear Problem
D.2.1 Dynamic Programming and Arrival Cost
D.2.2 Recursive Calculation of the Arrival Cost and One‑Step‑Ahead Prediction
D.2.3 Equivalence with the Kalman Filter

D.3 Stochastic State Feedback Problems
D.3.1 Open‑Loop Optimal Solution via Least Squares
D.3.2 Optimal Feedback Policy via Dynamic Programming
D.3.3 Open‑Loop Optimal Feedback Control vs. Optimal Feedback Control

D.4 Stochastic Output Feedback Problems
D.4.1 Optimal Output Feedback Controller
D.4.2 Derivation via Dynamic Programming
D.4.3 Extension to the Infinite Horizon Case: LQG Controller and Separation Principle
D.4.4 Analysis

E.1 Discrete Time Systems

E.2 The IMC Loop Structure and Properties
E.2.1 Stability
E.2.2 The Perfect Controller
E.2.3 Zero Offset
E.2.4 Robustness
E.2.5 IMC Feedforward Compensator Design
E.2.6 Saturation Constraints
E.2.7 IMC Tuning Method for PID Controllers
E.2.8 Analysis Tools

Exercises

F MPC Toolbox Tutorial: Shell Oil Fractionator

F.1 Problem Description
F.1.1 Background
F.1.2 Model Definition
F.1.3 Overview of Our Approach

F.2 Tutorial: Using the MPC GUI
F.2.1 Simplified Problem Definition
F.2.2 Using the MPC GUI

F.3 Solving the Shell Oil Control Problem Using the MPC Toolbox
F.3.1 Comparison of Control Structures
F.3.2 Specifying Target for MV

G A Brief Tutorial on Simulink

G.1 A MIMO System Example
G.2 Simulink Model for a CSTR