Marknadens största urval
Snabb leverans

Böcker i SpringerBriefs in Optimization-serien

Filter
Filter
Sortera efterSortera Serieföljd
  • av Remigijus Paulavi¿ius
    590,-

    After providing an in-depth introduction to derivative-free global optimization with various constraints, this book presents new original results from well-known experts on the subject. A primary focus of this book is the well-known class of deterministic DIRECT (DIviding RECTangle)-type algorithms. This book describes a new set of algorithms derived from newly developed partitioning, sampling, and selection approaches in the box- and generally-constrained global optimization, including extensions to multi-objective optimization. DIRECT-type optimization algorithms are discussed in terms of fundamental principles, potential, and boundaries of their applicability. The algorithms are analyzed from various perspectives to offer insight into their main features. This explains how and why they are effective at solving optimization problems. As part of this book, the authors also present several techniques for accelerating the DIRECT-type algorithms through parallelization and implementing efficient data structures by revealing the pros and cons of the design challenges involved. A collection of DIRECT-type algorithms described and analyzed in this book is available in DIRECTGO, a MATLAB toolbox on GitHub. Lastly, the authors demonstrate the performance of the algorithms for solving a wide range of global optimization problems with various constraints ranging from a few to hundreds of variables.Additionally, well-known practical problems from the literature are used to demonstrate the effectiveness of the developed algorithms. It is evident from these numerical results that the newly developed approaches are capable of solving problems with a wide variety of structures and complexity levels.Since implementations of the algorithms are publicly available, this monograph is full of examples showing how to use them and how to choose the most efficient ones, depending on the nature of the problem being solved. Therefore, many specialists, students, researchers, engineers, economists, computer scientists, operations researchers, and others will find this book interesting and helpful.

  • av Alexander J. Zaslavski
    650,-

    The book is devoted to the study of constrained minimization  problems on closed and convex sets in Banach spaces with a Frechet differentiable objective function. Such  problems are well studied in a  finite-dimensional space and in an infinite-dimensional Hilbert space. When the space is Hilbert there are many algorithms for solving optimization problems including the gradient projection algorithm which  is one of the most important tools in the optimization theory, nonlinear analysis and their applications. An optimization problem is described by an  objective function  and a set of feasible points. For the gradient projection algorithm each iteration consists of two steps. The first step is a calculation of a gradient of the objective function while in the second one  we calculate a projection on the feasible  set. In each of these two steps there is a computational error. In our recent research we show that the gradient projection algorithm generates a good approximate solution, if all the computational errors are bounded from above by a small positive constant. It should be mentioned that  the properties of a Hilbert space play an important role. When we consider an optimization problem in a general Banach space the situation becomes more difficult and less understood. On the other hand such problems arise in the approximation theory. The book is of interest for mathematicians working in  optimization. It also can be useful in preparation courses for graduate students.  The main feature of the book which appeals specifically to this audience is the study of algorithms for convex and nonconvex minimization problems in a general Banach space. The book is of interest for experts in applications of optimization to the approximation theory.In this book the goal is to obtain a good approximate solution of the constrained optimization problem in a general Banach space under  the presence of computational errors.  It is shown that the algorithm generates a good approximate solution, if the sequence of computational errors is bounded from above by a small constant. The book consists of four chapters. In the first we discuss several algorithms which are studied in the book and  prove a convergence result for an unconstrained problem which is a prototype of our results for the constrained problem. In Chapter 2 we analyze convex optimization problems. Nonconvex optimization problems  are studied in Chapter 3. In Chapter 4 we study  continuous   algorithms for minimization problems under the presence of computational errors. The algorithm generates a good approximate solution, if the sequence of computational errors is bounded from above by a small constant. The book consists of four chapters. In the first we discuss several algorithms which are studied in the book and  prove a convergence result for an unconstrained problem which is a prototype of our results for the constrained problem. In Chapter 2 we analyze convex optimization problems. Nonconvex optimization problems  are studied in Chapter 3. In Chapter 4 we study  continuous   algorithms for minimization problems under the presence of computational errors.

  • - Advances, Applications, and Challenges
    av Dimitris Souravlias
    730,-

    This book covers algorithm portfolios, multi-method schemes that harness optimization algorithms into a joint framework to solve optimization problems.

  • av Anatoly Zhigljavsky
    750,-

    Accessible to a variety of readers, this book is of interest to specialists, graduate students and researchers in mathematics, optimization, computer science, operations research, management science, engineering and other applied areas interested in solving optimization problems.

  • - A Metaheuristic Approach
    av Maude Josee Blondin
    816,-

    Students and researchers in engineering and optimization interested in optimization methods for controller tuning will utilize this book to apply optimization algorithms to controller tuning, to choose the most suitable optimization algorithm for a specific application, and to develop new optimization techniques for controller tuning.

  • - On Multiparametric Linear Complementarity Problems with Parameters in General Locations
    av Nathan Adelgren
    816,-

  • av Alexander J. Zaslavski
    680,-

    The discussion takes into consideration the fact that for every algorithm its iteration consists of several steps and that computational errors for different steps are different, in general.

  • av Wen Xu
    680,-

    The main topics covered in this book are: chrematistics of social networks, modeling of social influence propagation, popular research problems in social influence analysis such as influence maximization, rumor blocking, rumor source detection, and multiple social influence competing.

  • av Alexander J. Zaslavski
    680,-

    This book is devoted to the study of optimal control problems arising in forest management, an important and fascinating topic in mathematical economics studied by many researchers over the years.

  • av Xin-She Yang
    750,-

    Specific nature-inspired algorithms include: swarm intelligence, ant colony optimization, particle swarm optimization, bee-inspired algorithms, bat algorithm, firefly algorithm, and cuckoo search.

  • - Iteration Mappings, Attractors, and Basins of Attraction
    av Adam B. Levy
    796,-

    Numerical minimization of an objective function is analyzed in this book to understand solution algorithms for optimization problems. Multiset-mappings are introduced to engineer numerical minimization as a repeated application of an iteration mapping. Ideas from numerical variational analysis are extended to define and explore notions of continuity and differentiability of multiset-mappings, and prove a fixed-point theorem for iteration mappings. Concepts from dynamical systems are utilized to develop notions of basin size and basin entropy. Simulations to estimate basins of attraction, to measure and classify basin size, and to compute basin are included to shed new light on convergence behavior in numerical minimization.Graduate students, researchers, and practitioners in optimization and mathematics who work theoretically to develop solution algorithms will find this book a useful resource.

  • av Slawomir Koziel & Stanislav Ogurtsov
    796,-

    This Brief reviews a number of techniques exploiting the surrogate-based optimization concept and variable-fidelity EM simulations for efficient optimization of antenna structures.

Gör som tusentals andra bokälskare

Prenumerera på vårt nyhetsbrev för att få fantastiska erbjudanden och inspiration för din nästa läsning.