Skip to main content
  • Book
  • © 2021

A Derivative-free Two Level Random Search Method for Unconstrained Optimization

Authors:

  • Clarity of presentation as well as discussion of open problems are an attractive feature for instructors and potential practitioners in derivative-free methods for optimization
  • Highlights a new and simple derivative-free optimization algorithm which proves to be efficient and robust for solving unconstrained optimization problems
  • Underscores the two distinct phases of the algorithm has two distinct phases

Part of the book series: SpringerBriefs in Optimization (BRIEFSOPTI)

  • 1794 Accesses

Buy it now

Buying options

eBook USD 49.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book USD 64.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Other ways to access

This is a preview of subscription content, log in via an institution to check for access.

Table of contents (5 chapters)

  1. Front Matter

    Pages i-xi
  2. Introduction

    • Neculai Andrei
    Pages 1-17
  3. Convergence of the Algorithm

    • Neculai Andrei
    Pages 37-42
  4. Numerical Results

    • Neculai Andrei
    Pages 43-60
  5. Conclusion

    • Neculai Andrei
    Pages 61-63
  6. Back Matter

    Pages 65-118

About this book

The book is intended for graduate students and researchers in mathematics, computer science, and operational research. The book presents a new derivative-free optimization method/algorithm based on randomly generated trial points in specified domains and where the best ones are selected at each iteration by using a number of rules. This method is different from many other well established methods presented in the literature and proves to be competitive for solving many unconstrained optimization problems with different structures and complexities, with a relative large number of variables. Intensive numerical experiments with 140 unconstrained optimization problems, with up to 500 variables, have shown that this approach is efficient and robust.

Structured into 4 chapters, Chapter 1 is introductory. Chapter 2 is dedicated to presenting a two level derivative-free random search method for unconstrained optimization. It is assumed that the minimizing function is continuous, lower bounded and its minimum value is known. Chapter 3 proves the convergence of the algorithm. In Chapter 4, the numerical performances of the algorithm are shown for solving 140 unconstrained optimization problems, out of which 16 are real applications. This shows that the optimization process has two phases: the reduction phase and the stalling one. Finally, the performances of the algorithm for solving a number of 30 large-scale unconstrained optimization problems up to 500 variables are presented. These numerical results show that this approach based on the two level random search method for unconstrained optimization is able to solve a large diversity of problems with different structures and complexities.

There are a number of open problems which refer to the following aspects: the selection of the number of trial or the number of the local trial points, the selection of the bounds of the domains where the trial points and the local trial points are randomly generated and a criterion for initiating the line search.

Authors and Affiliations

  • Center for Advanced Modeling and Optimization, Academy of Romanian Scientists, Bucharest, Romania

    Neculai Andrei

About the author

Neculai Andrei holds a position at the Center for Advanced Modeling and Optimization at the Academy of Romanian Scientists in Bucharest, Romania. Dr. Andrei’s areas of interest include mathematical modeling, linear programming, nonlinear optimization, high performance computing, and numerical methods in mathematical programming. In addition to this present volume, Neculai Andrei has published several books with Springer including Nonlinear Conjugate Gradient Methods for Unconstrained Optimization (2020), Continuous Nonlinear Optimization for Engineering Applications in GAMS Technology (2017), and Nonlinear Optimization Applications Using the GAMS Technology (2013).

Bibliographic Information

  • Book Title: A Derivative-free Two Level Random Search Method for Unconstrained Optimization

  • Authors: Neculai Andrei

  • Series Title: SpringerBriefs in Optimization

  • DOI: https://doi.org/10.1007/978-3-030-68517-1

  • Publisher: Springer Cham

  • eBook Packages: Mathematics and Statistics, Mathematics and Statistics (R0)

  • Copyright Information: The Author(s), under exclusive license to Springer Nature Switzerland AG 2021

  • Softcover ISBN: 978-3-030-68516-4Published: 01 April 2021

  • eBook ISBN: 978-3-030-68517-1Published: 31 March 2021

  • Series ISSN: 2190-8354

  • Series E-ISSN: 2191-575X

  • Edition Number: 1

  • Number of Pages: XI, 118

  • Number of Illustrations: 1 b/w illustrations, 13 illustrations in colour

  • Topics: Optimization, Operations Research, Management Science

Buy it now

Buying options

eBook USD 49.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book USD 64.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Other ways to access