Gjør som tusenvis av andre bokelskere
Abonner på vårt nyhetsbrev og få rabatter og inspirasjon til din neste leseopplevelse.
Ved å abonnere godtar du vår personvernerklæring.Du kan når som helst melde deg av våre nyhetsbrev.
This Brief reviews a number of techniques exploiting the surrogate-based optimization concept and variable-fidelity EM simulations for efficient optimization of antenna structures.
Specific nature-inspired algorithms include: swarm intelligence, ant colony optimization, particle swarm optimization, bee-inspired algorithms, bat algorithm, firefly algorithm, and cuckoo search.
Numerical minimization of an objective function is analyzed in this book to understand solution algorithms for optimization problems. Multiset-mappings are introduced to engineer numerical minimization as a repeated application of an iteration mapping. Ideas from numerical variational analysis are extended to define and explore notions of continuity and differentiability of multiset-mappings, and prove a fixed-point theorem for iteration mappings. Concepts from dynamical systems are utilized to develop notions of basin size and basin entropy. Simulations to estimate basins of attraction, to measure and classify basin size, and to compute basin are included to shed new light on convergence behavior in numerical minimization.Graduate students, researchers, and practitioners in optimization and mathematics who work theoretically to develop solution algorithms will find this book a useful resource.
This book is devoted to the study of optimal control problems arising in forest management, an important and fascinating topic in mathematical economics studied by many researchers over the years.
The main topics covered in this book are: chrematistics of social networks, modeling of social influence propagation, popular research problems in social influence analysis such as influence maximization, rumor blocking, rumor source detection, and multiple social influence competing.
The discussion takes into consideration the fact that for every algorithm its iteration consists of several steps and that computational errors for different steps are different, in general.
Students and researchers in engineering and optimization interested in optimization methods for controller tuning will utilize this book to apply optimization algorithms to controller tuning, to choose the most suitable optimization algorithm for a specific application, and to develop new optimization techniques for controller tuning.
Accessible to a variety of readers, this book is of interest to specialists, graduate students and researchers in mathematics, optimization, computer science, operations research, management science, engineering and other applied areas interested in solving optimization problems.
This book covers algorithm portfolios, multi-method schemes that harness optimization algorithms into a joint framework to solve optimization problems.
The book is devoted to the study of constrained minimization problems on closed and convex sets in Banach spaces with a Frechet differentiable objective function. Such problems are well studied in a finite-dimensional space and in an infinite-dimensional Hilbert space. When the space is Hilbert there are many algorithms for solving optimization problems including the gradient projection algorithm which is one of the most important tools in the optimization theory, nonlinear analysis and their applications. An optimization problem is described by an objective function and a set of feasible points. For the gradient projection algorithm each iteration consists of two steps. The first step is a calculation of a gradient of the objective function while in the second one we calculate a projection on the feasible set. In each of these two steps there is a computational error. In our recent research we show that the gradient projection algorithm generates a good approximate solution, if all the computational errors are bounded from above by a small positive constant. It should be mentioned that the properties of a Hilbert space play an important role. When we consider an optimization problem in a general Banach space the situation becomes more difficult and less understood. On the other hand such problems arise in the approximation theory. The book is of interest for mathematicians working in optimization. It also can be useful in preparation courses for graduate students. The main feature of the book which appeals specifically to this audience is the study of algorithms for convex and nonconvex minimization problems in a general Banach space. The book is of interest for experts in applications of optimization to the approximation theory.In this book the goal is to obtain a good approximate solution of the constrained optimization problem in a general Banach space under the presence of computational errors. It is shown that the algorithm generates a good approximate solution, if the sequence of computational errors is bounded from above by a small constant. The book consists of four chapters. In the first we discuss several algorithms which are studied in the book and prove a convergence result for an unconstrained problem which is a prototype of our results for the constrained problem. In Chapter 2 we analyze convex optimization problems. Nonconvex optimization problems are studied in Chapter 3. In Chapter 4 we study continuous algorithms for minimization problems under the presence of computational errors. The algorithm generates a good approximate solution, if the sequence of computational errors is bounded from above by a small constant. The book consists of four chapters. In the first we discuss several algorithms which are studied in the book and prove a convergence result for an unconstrained problem which is a prototype of our results for the constrained problem. In Chapter 2 we analyze convex optimization problems. Nonconvex optimization problems are studied in Chapter 3. In Chapter 4 we study continuous algorithms for minimization problems under the presence of computational errors.
After providing an in-depth introduction to derivative-free global optimization with various constraints, this book presents new original results from well-known experts on the subject. A primary focus of this book is the well-known class of deterministic DIRECT (DIviding RECTangle)-type algorithms. This book describes a new set of algorithms derived from newly developed partitioning, sampling, and selection approaches in the box- and generally-constrained global optimization, including extensions to multi-objective optimization. DIRECT-type optimization algorithms are discussed in terms of fundamental principles, potential, and boundaries of their applicability. The algorithms are analyzed from various perspectives to offer insight into their main features. This explains how and why they are effective at solving optimization problems. As part of this book, the authors also present several techniques for accelerating the DIRECT-type algorithms through parallelization and implementing efficient data structures by revealing the pros and cons of the design challenges involved. A collection of DIRECT-type algorithms described and analyzed in this book is available in DIRECTGO, a MATLAB toolbox on GitHub. Lastly, the authors demonstrate the performance of the algorithms for solving a wide range of global optimization problems with various constraints ranging from a few to hundreds of variables.Additionally, well-known practical problems from the literature are used to demonstrate the effectiveness of the developed algorithms. It is evident from these numerical results that the newly developed approaches are capable of solving problems with a wide variety of structures and complexity levels.Since implementations of the algorithms are publicly available, this monograph is full of examples showing how to use them and how to choose the most efficient ones, depending on the nature of the problem being solved. Therefore, many specialists, students, researchers, engineers, economists, computer scientists, operations researchers, and others will find this book interesting and helpful.
Abonner på vårt nyhetsbrev og få rabatter og inspirasjon til din neste leseopplevelse.
Ved å abonnere godtar du vår personvernerklæring.