Skip to main content

Numerical Optimization with Computational Errors

  • Book
  • © 2016

Overview

  • Examines approximate solutions of optimization problems in the presence of computational errors
  • Reinforces basic principles with an introductory chapter
  • Analyzes the gradient projection algorithm for minimization of convex and smooth functions
  • Includes supplementary material: sn.pub/extras

Part of the book series: Springer Optimization and Its Applications (SOIA, volume 108)

This is a preview of subscription content, log in via an institution to check access.

Access this book

eBook USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book USD 109.00
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book USD 109.00
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Other ways to access

Licence this eBook for your library

Institutional subscriptions

Table of contents (16 chapters)

Keywords

About this book

This book studies the approximate solutions of optimization problems in the presence of computational errors. A number of results are presented on the convergence behavior of algorithms in a Hilbert space; these algorithms are examined taking into account computational errors. The author illustrates that algorithms generate a good approximate solution, if computational errors are bounded from above by a small positive constant. Known computational errors  are examined with the aim of determining an approximate solution. Researchers and students interested in the optimization theory and its applications will find this book instructive and informative.

 

This monograph contains 16 chapters; including a chapters devoted to the subgradient projection algorithm, the mirror descent algorithm, gradient projection algorithm, the Weiszfelds method, constrained convex minimization problems, the convergence of a proximal point method in a Hilbert space, the continuous subgradient method, penalty methods and Newton’s method.

  


Reviews

“The author studies the approximate solutions of optimization problems in the presence of computational errors. A number of results are presented on the convergence behavior of algorithms in a Hilbert space. Researchers and students will find this book instructive and informative. The book has contains 16 chapters … .” (Hans Benker, zbMATH 1347.65112, 2016)

Authors and Affiliations

  • Department of Mathematics, The Technion – Israel Institute of Techn, Haifa, Israel

    Alexander J. Zaslavski

Bibliographic Information

Publish with us