Sparse Optimization Theory and Methods

個数:1
紙書籍版価格
¥53,123
  • 電子書籍

Sparse Optimization Theory and Methods

  • 著者名:Zhao, Yun-Bin
  • 価格 ¥10,910 (本体¥9,919)
  • CRC Press(2018/07/04発売)
  • ポイント 99pt (実際に付与されるポイントはご注文内容確認画面でご確認下さい)
  • 言語:ENG
  • ISBN:9781138080942
  • eISBN:9781351624145

ファイル: /

Description

Seeking sparse solutions of underdetermined linear systems is required in many areas of engineering and science such as signal and image processing. The efficient sparse representation becomes central in various big or high-dimensional data processing, yielding fruitful theoretical and realistic results in these fields. The mathematical optimization plays a fundamentally important role in the development of these results and acts as the mainstream numerical algorithms for the sparsity-seeking problems arising from big-data processing, compressed sensing, statistical learning, computer vision, and so on. This has attracted the interest of many researchers at the interface of engineering, mathematics and computer science.

Sparse Optimization Theory and Methods presents the state of the art in theory and algorithms for signal recovery under the sparsity assumption. The up-to-date uniqueness conditions for the sparsest solution of underdertemined linear systems are described. The results for sparse signal recovery under the matrix property called range space property (RSP) are introduced, which is a deep and mild condition for the sparse signal to be recovered by convex optimization methods. This framework is generalized to 1-bit compressed sensing, leading to a novel sign recovery theory in this area. Two efficient sparsity-seeking algorithms, reweighted l1-minimization in primal space and the algorithm based on complementary slackness property, are presented. The theoretical efficiency of these algorithms is rigorously analysed in this book. Under the RSP assumption, the author also provides a novel and unified stability analysis for several popular optimization methods for sparse signal recovery, including l1-mininization, Dantzig selector and LASSO. This book incorporates recent development and the author’s latest research in the field that have not appeared in other books.

Table of Contents

Preface

Uniqueness of the Sparsest Solution of Linear Systems

Introduction

Spark

Uniqueness via Mutual Coherence

Improved Uniqueness Criteria via Coherence Rank

Babel Function and Sub-Babel Function

Notes

Uniqueness of Solutions to `1-Minimization Problems

Strict Complementary Slackness Property (SCSP)

Least `1-Norm Nonnegative Solution

Least `1-Norm Points in Polyhedra

Notes

Equivalence of `0- and `1-Minimization

Equivalence and Strong Equivalence

Standard `0- and `1-Minimization Problems

Problems with Nonnegativity Constraints

Application to Linear Programming

Equivalence of `0-Problem and Weighted `1-Problem

Sparse Vector Recovery

Sparse Nonnegative Vector Recovery

Notes

Bit Compressed Sensing

Introduction

Sign Measurements and Recovery Criteria

Relaxation Models

Consistency Condition

Reformulation of 1-Bit Compressed Sensing

Nonuniform Sign Recovery

Uniform Sign Recovery

Notes

Stability of Linear Sparse Optimization Methods

Introduction

Hoffman’s Error Bound for Linear Systems

Weak RSP of Order k of AT

Stability of Standard `1-Minimization

Linear Dantzig Selector

Special Cases

Notes

Stability of Nonlinear Sparse Optimization Methods

Introduction

Orthogonal Projection Operator

Polytope Approximation of Unit Balls

A Necessary Condition for Stability

`1-Minimization with `2-Norm Constraints

Nonlinear Dantzig Selector

The LASSO Problem

Summary

Notes

Reweighted `1-Algorithms

Merit Function for Sparsity

Reweighted `1-Methods

Numerical Experiments

Theoretical Analysis

Notes

Sparsity via Dual Density

Introduction

`0-Minimization with Nonnegativity Constraints

DDRW for Standard `0-Minimization

Sparsity Enhancement for Weighted `1-Minimizers

Notes

References