Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Assessment of the Uniform Wear Bending Strength of Large Modulus Rack and Pinion Pair: Theoretical vs. Experimental Results
Previous Article in Journal
A Study on the Running of a Joystick-Type Six-Wheeled Electric Wheelchair When Curb Climbing
Previous Article in Special Issue
Pipeline Landmark Classification of Miniature Pipeline Robot π-II Based on Residual Network ResNet18
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Optimization Techniques in the Localization Problem: A Survey on Recent Advances

1
Doctoral School of Applied Informatics and Applied Mathematics, Obuda University, 1034 Budapest, Hungary
2
Department of Mechatronics and Automation, Faculty of Engineering, University of Szeged, 6725 Szeged, Hungary
*
Author to whom correspondence should be addressed.
Machines 2024, 12(8), 569; https://doi.org/10.3390/machines12080569
Submission received: 9 July 2024 / Revised: 9 August 2024 / Accepted: 13 August 2024 / Published: 19 August 2024

Abstract

:
Optimization is a mathematical discipline or tool suitable for minimizing or maximizing a function. It has been largely used in every scientific field to solve problems where it is necessary to find a local or global optimum. In the engineering field of localization, optimization has been adopted too, and in the literature, there are several proposals and applications that have been presented. In the first part of this article, the optimization problem is presented by considering the subject from a purely theoretical point of view and both single objective (SO) optimization and multi-objective (MO) optimization problems are defined. Additionally, it is reported how local and global optimization problems can be tackled differently, and the main characteristics of the related algorithms are outlined. In the second part of the article, extensive research about local and global localization algorithms is reported and some optimization methods for local and global optimum algorithms, such as the Gauss–Newton method, Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Differential Evolution (DE), and so on, are presented; for each of them, the main concept on which the algorithm is based, the mathematical model, and an example of the application proposed in the literature for localization purposes are reported. Among all investigated methods, the metaheuristic algorithms, which do not exploit gradient information, are the most suitable to solve localization problems due to their flexibility and capability in solving non-convex and non-linear optimization functions.

1. Introduction

Optimization is a mathematical discipline which provides tools to select the best element from a set of possible alternatives based on some given criteria. In general, an optimization problem is one of maximizing or minimizing a real function by methodically selecting input values from a permitted set and figuring out the output values of the function itself. Optimization problems can be classified as continuous or discrete, and to solve them, several optimization techniques and approaches have been developed. These techniques can be divided into exact methods, approximate algorithms, metaheuristics, and greedy algorithms [1]. Each of them has pros and cons, and the user must choose the proper and most appropriate technique to solve the specific problem.
Nowadays, optimization, sometimes also referred to as mathematical programming, is used in every scientific field where it is necessary to deal with mathematical problems characterized by high complexity, constraints, interdependencies among variables, and a large space of solution [2]. Finance [3], chemical engineering [4], mechanical engineering [5], electrical engineering [6], and system and database design [7] are only a few examples where optimization has been applied in the last decades. In the field of localization, optimization methods have been widely used to solve localization problems pertaining to robots [8], smartphones [9], permanent magnets [10], vehicles [11], underground structures [12], urban utilities [13], and so on. By referring to this field, this paper aims to investigate and collect the main used optimization methods in the localization literature, such as the Gradient Descent method, Levenberg–Marquardt method, Genetic Algorithm (GA), Particle Swarm Optimization (PSO), etc., by reporting the concept of the used technique and an example of its application in the considered field. Furthermore, the paper also aims to provide in the initial part a general introduction to the classic optimization problem, providing the mathematical formulation, and, in the second part, a clear mathematical formulation of each investigated technique.
The rest of the paper is organized as follows. In Section 2, firstly, the single objective (SO) optimization problem is mathematically defined; then, the local vs global cases are presented, and at the end of the section, the multi-objective optimization (MO) problem with the related approaches is reported. In Section 3, some of the most interesting and important applications of the optimization methods in the field of localization are summarized and classified; for each of them, a brief explanation of the method and the inspired idea are reported. In Section 4, the conclusions are reported.

2. A Brief Overview of the Optimization Problem

2.1. Definition

Optimization methods are used to find a set of inputs to an objective function that results in the evaluation of the minimum or maximum of the function. Mathematically, the optimization problem is defined as follows [14]:
m i n i m i z e : f 0 x s u c h   t h a t : g j x 0 j = 1 , , m h k x = 0 k = 1 , , p
where f 0 x is a function, called an objective function, and defined as f 0 : R n R , x = x 1 ; x 2 ; ; x n T is the optimization vector variable (or the design variables) of the problem, and g j x and h k x are the inequality and equality functions, respectively, that are constraint functions.
If for any z with g 1 z 0 , , g m z 0 and h 1 z = 0 , , h k z = 0 , the relation f 0 x * f 0 ( z ) is verified, then the vector x * is the optimal solution of the problem and it has the smallest objective value among all vectors that satisfy the constraint functions. The maximum problem of a function f ( x ) can be solved considering the minimum problem of − f ( x ) .
The objective function and the constraint functions can be linear or non-linear and an optimization problem can be called a linear program if the objective ( f 0 , , f m ) and constraint functions satisfy the linear condition that can be expressed, for example, as follows:
f s α x + β y = α f s x ) + β f s ( y
where the variable s x , y   R n , and the coefficients α , β R . If the equality in (2) is not verified, the problem is a non-linear program [14]. The problems can also be classified as convex optimization problems if the following condition is satisfied:
f s α x + β y α f s x ) + β f s ( y
where α + β = 1 , α 0 , and β 0 . The convex optimization can be considered a generalization of a linear programming because any linear program is also a convex optimization problem [14].
The solution to the problem is found inside the searchable design space that is subjected to the side constraints defined as x i L x i x i U with i = 1 , , n , where variables are limited between the upper and lower bounds [15].
Optimization problems can have some or all of the design variables restricted to integer or discrete values, and they are referred to as integer or discrete optimization problems, respectively.
To find a solution for a class of optimization problems in relation to a given accuracy, a specific algorithm is needed; most of them are able to deal with the side constraints and the constraints function separately because the side constraints are managed directly by the algorithms themselves; unconstrained problems exist, but they can still have side constraints [15].
One or more equality and/or inequality constraints, with or without side restrictions, characterize a constrained optimization problem. For a generic problem, an equality constraint can only either be violated or satisfied, while an inequality constraint can be violated, active, or satisfied. An active inequality constraint, defined as g j x 0 , is said to be active (or tight) when at a design point x * , it is satisfied as an equality, i.e., g j x = 0 [15].
Optimization techniques, or algorithms, use a specific procedure that is able to find the solution among design variable values that results in the best objective function value, and at the same time, it satisfies all the constraint functions (equality, inequality) and the side constraints; more than one optimum can be found, and they are referred to as local or relative optima.

2.2. A Comparison between Local and Global Extrema

The extreme minimum or extreme maximum of the objective function over the whole input search space is known as a global optimum. In general, finding a local optimum may be rather simple, whereas finding the global optimum may be more challenging, in particular for non-linear functions [16]. As an example, Figure 1 shows a generic function f x which has three minima: two of them are of the local type, and one is the global minimum of the problem.
It follows that optimization problems and algorithms can be classified into two types: local optimization and global localization algorithms. The former type of algorithm looks for local optimum solutions in a particular area of the search space, whereas the latter looks for the best solution for problems with multiple local optimums. Furthermore, there could be multiple local optimal points for an objective function, and if there is only one, that local optimum point is referred also to as the global optimum point. In case a global optimum does not exist for an objective function, the problem cannot be classified as an optimization problem.
In general, optimization problems are typically classified as multimodal or unimodal in terms of their modality. When a problem has more local optima solutions in addition to the global optimum, it is said to be multimodal; when it has only the global optimum as its local optima solution, it is said to be unimodal. Multimodal problems can sometimes be trickier and therefore more complex to solve [17].
The authors in reference [18] classify optimization problems as reported in Figure 2. It is highlighted that non-convex functions do not need to be multimodal because a non-convex function can have only one local optimum, and it follows that this is also the global optimum.
Local and global optimization algorithms deal with different problems; a local optimization technique is utilized when the objective function includes a single optimum or when the region of the global optimum is known; when the objective function response surface’s structure is unknown or the function has local optimum points, a global optimization procedure is employed [19].
In order to discover the extrema of the function in a certain region of the search space or to come as close to them as possible, a local optimization method is developed to cross the considered region looking for the minima (or maxima) [14].
Global optimization techniques are suitable for situations where there are few variables and no necessity for fast computation time. They work with a single candidate solution or a population of them, from which new candidate solutions are generated iteratively and assessed using a predetermined function to see whether the newly generated variables are getting better or worse [14].

2.3. Local Optimization Algorithms

Most local optimization algorithms exploit the gradient information of the function to find the optimum solution. They work well for solving problems with lots of design variables, and fine-tuning their settings is rather simple. Furthermore, they are considered efficient compared to the number of function evaluations (or iterations) needed to find the solution. Among the drawbacks are that they can only locate a local optimum, they have issues solving discrete optimization problems, they require algorithms that are difficult to implement efficiently, and they suffer from numerical noise [15].
There are many types of gradient-based algorithms, and each of them is characterized by a different logic or idea to determine the search direction. In general, gradient-based algorithms typically use a two-step iterative process to reach the optimum: the first one is to exploit the gradient information to find a search direction S in which to move, and the second step is moving in the direction d ; the last one is called one-dimensional or a line search and also provides the optimum step size, α k . The two steps are repeated until no more progress is obtained and mathematically, they can be expressed as follows:
x k = x k 1 + α k d
where k identifies the number of the k -th-step [20]. The gradient information is not always easily accessible in complex problems, and techniques such as Finite Difference Gradient Calculations, Automatic Differentiation, or Analytic/Semi-Analytic Gradient Calculations can be used to compute the gradient. Moreover, the first one provides gradient information that is accurate at the working precision; the second one provides only an approximation of the gradient, giving an accuracy that is related to the selected step size; the last one can be applied to linear finite element codes, and it is inexpensive [15].

2.4. Global Optimization Algorithms

Many problems can have multiple local optima (like the problem represented in Figure 1), and local globalization algorithms are able to find one of them without guaranteeing or saying if it is a local or global minimum point because all minima satisfy the Karush–Kuhn–Tucker (KKT) condition, which is a mathematical tool to verify if a solution is a minimum or not, but without specifying if it is a global one.
The found minima depend on the starting point, giving the first met minima as a result. To find other minima in the design space, a multi-start approach can be adopted, which starts from different points [21].
Global optimization algorithms have a far higher possibility of locating the global or near-global optimum. It is preferable to speak of these algorithms as having global properties because, in general, no algorithm can ensure convergence on a global optimum.
They fall into one of two categories: metaheuristic algorithms and deterministic algorithms [22].

2.4.1. Metaheuristic Algorithms

Metaheuristic (or stochastic) algorithms can be defined as computational intelligence paradigms which are especially used for solving complex optimization problems [23]. When precise optimization methods are unable to provide outcomes, they can provide satisfactory results. Metaheuristic algorithms can be classified into evolution-based, swarm intelligence-based, physics-based, and human behavior-related algorithms [24].
This class of optimization techniques is suitable for complex non-linear and discontinuous problems where classical optimization techniques might fail [25].
These algorithms, in contrast to the local methods, use a certain number of design points, which are often referred to as a population of individuals; they allow for determining the optimum without adopting any gradient information. These techniques frequently draw inspiration from various natural events, and some of their advantages include a higher likelihood of locating a global or nearly global optimum, robustness, and simplicity in implementation; moreover, they are also suitable for solving discrete optimization problems. On the other hand, some disadvantages include high computing costs, the limited size of the problem, weak constraint-handling capabilities, and difficulties in tuning problem-specific parameters [15].
These days, PSO and the GA are the two most often used metaheuristic algorithms. Additionally, evolutionary algorithms include genetic programming, Ant Colony Optimization (ACO), Simulated Annealing (SA), Differential Evolution (DE), Evolutionary Programming, Harmony Search, and others [22].

2.4.2. Deterministic Algorithms

Deterministic algorithms aim to find the global solution of an optimization problem by guaranteeing that the found solution is the real global minimum (or maximum) of the problem with a predefined tolerance. These algorithms usually deal with specific classes of problems, such as linear programming (LP), mixed-integer linear programming (MILP), non-linear programming (NLP), and mixed-integer non-linear programming (MINLP). An example of a deterministic technique is the DIRECT algorithm [26], which can find the global minimum of a multivariate function with simple bounds; it is a modification of the conventional Lipschitzian method where the need to specify a Lipschitz constant is eliminated.
In general, deterministic approaches can obtain advantages by exploiting the analytical properties of a problem to find the global solution; however, over the years, heuristic methods have shown to be more efficient and flexible than deterministic ones [18].

2.5. From Single to Multi-Objective Optimization

The case described by Equation (1) is referred to as a single objective optimization problem because the function to be optimized is only one. However, in many applications, the optimization of more than one objective function at the same time is requested, and this case, it is referred to as a multi-objective (MO) optimization problem. It is mathematically described as follows:
m i n i m i z e : f m x m = 1 , , M s u c h   t h a t : g j x 0     j = 1 , , J h k x = 0    k = 1 , , K
where m is the m -th function among the M functions to be optimized.
In MO optimization problems, two spaces are defined: the first is the decision variable space of the solution vector that contains all possible solutions x of the problem; the second is the multi-dimensional space of the objective function vector that contains the evaluations of each objective function; each x solution belonging to the decision variable space corresponds to a point in the objective function space. The spaces for a problem with a vector solution x with three components and two membership functions are represented in Figure 3 [27].
If every objective function and solution region in a MO problem are convex, the problem is said to be convex too, and the convexity is crucial to solving the optimization problem.
The methods to find the solution to MO problems can be classified into two different types: the Pareto method and the scalarization method [28]. The former is used when performance indicators and desired solutions are distinct, and in such a way, the Pareto approach is applied to generate a compromise solution, or trade-off, that can be represented as a Pareto optimal front (POF). Unlike in the latter, a performance indicator component method is exploited to create a scalar function that is integrated into the fitness function [29].
The process of finding the optimal solution can be assisted by using metaheuristic algorithms such as GA, PSO, ACO, and so on.

2.5.1. Pareto Method

From a mathematical point of view, a MO problem with n objective functions solved by using the Pareto method is expressed as follows:
f 1 , o p t = min   f 1 ( x ) f 2 , o p t = min   f 2 ( x ) f n , o p t = min   f n ( x ) .
The method keeps the elements of the solution vectors independent and exploits the concept of dominance with which a solution can be considered a Pareto optimal solution (POS) or a non-dominated solution if around the considered solution there is no way of improving any objective without degrading at least one other objective; on the contrary, a solution is said to be a Pareto-dominated solution (PDS) if it can be improved without worsening at least one other objective. As an example, Figure 4 shows the solutions in the space of the objective functions where POSs are represented by red dots, whilst the PDSs are represented by red squares. Furthermore, in agreement with the representation in Figure 4, it is possible to give the following definitions [29]:
  • Anchor point: the best value provided by an objective function.
  • Utopia point: a point which is obtained by intersecting the best values provided by the objective functions.
  • Pareto optimal front (POF): the line where the POSs lay.
Once the POF has been obtained, the MO problem is solved by computing which of the solutions on the POF has the shortest Euclidean distance. The following equation is used:
d E = m i n Q 1 Q 1 * Q 1 n o r m 2 + Q 2 Q 2 * Q 2 n o r m 2
where ( Q 1 * , Q 2 * ) are the coordinates for the utopia point related to f 1 and f 2 , respectively, Q 1 , Q 2 are the coordinates of the solution point on the POF, and Q 1 n o r m and Q 2 n o r m are normalization factors based on the minimum values of each function.
In general, by considering a MO problem with two objective functions, four types of problems can be obtained and are given as follows:
Problem   1 :   f 1 , o p t = min   f 1 ( x ) f 2 , o p t = min   f 2 ( x ) ,
Problem   2 :   f 1 , o p t = min   f 1 x f 2 , o p t = max   f 2 x ,
Problem   3 :   f 1 , o p t = max   f 1 x f 2 , o p t = min   f 2 x ,
Problem   4 :   f 1 , o p t = max   f 1 ( x ) f 2 , o p t = max   f 2 ( x ) .
Each of the four problems gives rise to a different shape of the POF and they are shown in Figure 5.
The objective function space for a problem with three functions can be represented in a three-dimensional space, but for a problem with a number of objective functions greater than three, the space cannot be represented.
To find non-dominated solutions, specific algorithms are used such as the continuously updated method reported in [29].

2.5.2. Scalarization Method

The scalarization method includes each function of the MO problem into a single scalar fitness function, practically transforming the MO problem into a single objective problem. Different types of scalarization can be defined and some basic approaches are the Weighted Global Criterion Method (WGCM), Weighted Sum Method (WSM), Exponential Weighted Criterion (EWC), or the Weighted Product Method (WPM) [30].
From a conceptual point of view, the WSM is the easiest way to implement the scalarization method. Each function of the MO problem is summed by adopting weights; it gives rise to a new total objective function F x which is given as follows:
F x = w 1 f 1 x + w 2 f 2 x + + w i f i ( x ) + + w n f n ( x )
where i is the i -th function of the original MO problem, w i represents the weight of the i -th function, and f i is the i -th function of the original MO problem composed of n objective functions [29]. The values of the weights are chosen in agreement with the performance priority: a large weight is given to functions with higher priority and vice versa for functions with lower priority. The following are three examples of how weights can be set:
  • Equal weights: w i = 1 / n .
  • Rank Order Centroid (ROC) weights: w i = 1 n k = i n 1 k .
  • Rank Sum (RS) weights: w i = 2 ( n + 1 i ) n ( n + 1 ) .
To provide a balance among functions, normalization is needed, and it can be obtained by dividing each objective function by its own RMS value f i R M S [31]. The final expression is given as follows:
F x = + w 1 f 1 x f 1 R M S + w 2 f 2 x f 2 R M S + + w n f n x f n R M S
In Equation (13), the sign of each objective function must agree with the request of minimization (minus) or maximization (plus).
The WSM can be considered as a particular case of the WGCM approaches, in which the scalarization can be obtained by adopting one of the following definitions [30]:
F x = i = 1 n ω i f i x p
F x = i = 1 n ω i f i x p
where p is a parameter which can enhance the minimization of the function, and weights must satisfy ω i = 1 .
The EWC method is defined as follows [32]:
F x = i = 1 n e p ω i 1 e p f i x ,
whilst the WPM is given as follows [33]:
F x = i n f i x ω i .

2.5.3. Summary of the Optimization Problems

In summary, from the point of view of the number of objectives, optimization problems can be divided into two types: SO and MO problems. The main difference in the formulation of the problem is that in the former, only one objective is considered, whereas in the latter, there are at least two objectives. The MO case can be considered an extension of the SO case, and allows the optimization of more functions at the same time that belong to the same problem and have the same design variables.
Finally, the MO problem can be solved by using either the Pareto method or one of the scalarization approaches. The first one deals with the optimization problem by solving each fitness function separately, whereas the second ones provide a function that is a single figure of merit that englobes all the objectives of the problem in one function. The main advantage of the Pareto approach is that each fitness function is solved independently without any regard for the other objectives, and the final solution is extracted by considering the solution that has the shortest distance from the utopia point. On the other hand, scalarization transforms the MO problem into the SO type, but has the drawback of setting the weights of each objective properly in order to correctly set their priority.

3. Optimization Algorithms Used in the Field of Localization

In this section, all optimization methods in the field of localization are reported. Generally, a description of the algorithm, its concept, and the mathematical formulation are reported; then, a real and meaningful application for localization purposes is described.

3.1. Newton’s Method, Gradient Descent, and Gauss–Newton Method

Newton’s Method and the Gradient Descent (GD) algorithm are the first methods which were studied in the field of optimization to deal with local problems, and they have a similar structure but with the difference that the GD algorithm is a first-order method and Newton’s method is a second-order type. The latter has the disadvantage of a higher computational cost because it requires the calculation of the Hessian matrix. At the end of the paragraph, an extension of Newton’s method called the Gauss–Newton method is presented.

3.1.1. Newton’s Method

Newton’s algorithm is a classical method and is an unconstrained algorithm that is derived from a second-order Taylor series expansion of the objective function starting from an initial design point x 0 . The function can be expressed as follows [15]:
f x f x 0 + f x 0 T x x 0 + 1 2 x x 0 T H ( x 0 ) ( x x 0 )
where x is the variable of the problem, and H x 0 is the Hessian matrix that contains the second-order gradient information of the objective function. By differentiating Equation (18) with respect to x and setting the result equal to zero according to the KKT conditions, it gives rise to the following iterative formula for the current design point [34]:
x n + 1 = x n f ( x n ) f ( x n )
where f and f are the first and the second derivative, respectively, which can be expressed as f x n = f x n and f x n = H x n 1 . The step size cannot be changed, and no one-dimensional search is required because it is provided by H x n 1 f x n .
The method has a quadratic rate of convergence, requiring only a single step to obtain the optimum; however, the estimation of the second-order derivative requires a high computational cost and often the method can result in being unpractical. To overcome this drawback, some variation of gradient-based methods can be adopted, which makes use only of the first-order gradient information.

3.1.2. Gradient Descent Algorithm

The GD algorithm is a first-order optimization algorithm that finds a local minimum (or maximum) by using an iterative process. The function to be optimized must be differentiable and convex. However, it is also possible to apply the algorithm to quasi-convex functions, but with the risk of being stuck at saddle points.
The GD algorithm is based on this iterative equation [35]:
x k + 1 = x k α f ( x k )
where f is the function to be optimized, x k is the input variable at the k -th step, and α is a parameter which controls the step size. It is highlighted that the value of α controls the behavior of the method because, for example, the smaller the learning rate, the longer the GD convergence time, and a too-small α could make the algorithm reach the maximum number of iterations too early without finding the right solution; on the other hand, a too-large α could make the process diverge or just jump around to the minimum (or maximum) of the problem. The value α f ( x k ) in (20) is subtracted when the minimum is looked for, whereas it is summed when the maximum is the desired output.
The GD algorithm deals with unconstrained problems, and it can be extended to also handle constraints by adopting a specific approach called the penalty function method, which is able to transform an unconstrained problem into a constrained one [20]. The method adds a penalty function to the original objective function that is composed of a penalty parameter multiplied by a measure of violation of the constraints. The latter is zero when there is no violation and greater than zero when violations are obtained. Considering a generic problem m i n = f ( x ) subjected to c i ( x ) 0 , it is transformed into a series of unconstrained minimization problems, given as follows [36]:
min   Φ k x = f x + σ k i g ( c i x )
where g c i x = max   0 , c i x 2 is the exterior penalty function, σ k is the penalty coefficient, and k represents the iteration. In each step of the algorithm, the penalty coefficient is increased, the unconstrained problem is solved, and the provided result is used in the next iteration as the initial guess. In this manner, the solution of the original constrained problem is obtained by the convergence of the unconstrained one.
In general, this method has been used in many different scientific fields, and, in particular, one of them where it has been widely adopted is Machine Learning (ML), in which the GD algorithm has three variations: the Batch GD, Stochastic GD, and Mini-Batch GD algorithms. These three different approaches manage the training process of a neural network in different ways: in the first one, the algorithm calculates the gradient to correct the weights by considering the complete training dataset in each iteration, in the second one, the gradient is calculated by considering only one sample of the dataset in each iteration, and in the last case, which is an intermediate approach between the two aforementioned approaches, the dataset is split into small subsets (batches) and the gradients are computed for each sub-batch in each iteration [37].

3.1.3. Gauss–Newton Method

The Gauss–Newton method is adopted to solve convex non-linear least-squares problems, i.e., a class of problems where it requested to minimize a sum of squared function values. Given an objective function which is twice differentiable, it has a form given as follows:
f x = 1 2 i = 1 m f i x 2
and its gradient is defined as follows:
f x = i = 1 m f i x f i ( x ) .
The search direction of the Gauss–Newton method is as follows [14]:
Δ x g n = i = 1 m f i x f i x T 1 i = 1 m f i x f i ( x ) .
where it is assumed that the inverse exists. The difference with the classic Newton’s method is that in this case, the direction is obtained by neglecting the second term from the Hessian. This result can be obtained by considering the first order of Taylor’s series f i x + v = f i x + f i x T v , which gives rise to the following:
f x + v = 1 2 i = 1 m f i x + f i x T v 2
where v = Δ x g n is the needed exact value to minimize this approximation; furthermore, it is highlighted that the Gauss–Newton step direction can be estimated by solving a linear least-squares problem.

3.1.4. The Integer Least-Squares Problem

A particular case among least-squares problems is when the design variables are integers and these types of problems are named integer-least-squares (ILS) problems; they arise, for example, to solve the integer frequency ambiguity in Global Navigation Satellite Systems (GNSSs) for real-time applications, or in other fields such as communications, cryptography, and lattice design [38].
The pseudorange and carrier phase are extracted from GNSS measurements, and only the latter allows pose estimation with high precision in the order of the millimeter and centimeter level. However, the unknown integer part of the carrier phase, i.e., the frequency integer ambiguity, needs to be solved, and it can be performed by adopting the geometry-based model, which relies on the resolution of an ILS problem [39]. Such a problem can be defined as follows [38]:
min a Z n   a a ^ T Q a ^ 1 a a ^
where a is the double-differenced integer ambiguity vector, a ^ is the integer estimation of a , and Q a ^ is the symmetric positive n × n variance–covariance matrix.
To solve the problem represented by Equation (22), specific algorithms, such as the least-squares AMBiguity Decorrelation Adjustment (LAMBDA) algorithm or its variations, have been developed to reduce the computational resolution complexity of the problem [38]. They are composed of two parts: the reduction process, and the discrete search process. In this review, they are not investigated further because these algorithms are considered numerical techniques adopted to simplify the resolution of an optimization problem.

3.1.5. Integer Frequency Ambiguities Resolution

The authors in reference [39] propose a single-epoch triple-frequency real-time kinematic positioning method that exploits the information provided by the GNSS to obtain high-precision positioning in urban environments.
In their article, they propose the Random Sample Consensus (RANSAC) that is adopted after the ILS is solved to detect and exclude incorrectly fixed ambiguities of the Extra-Wide-Lane (EWL), Wide-Lane (WL), and original frequencies. The method deals with a triple-frequency integer ambiguity, and the double-differenced pseudorange P i and carrier phase L i observations for the i -th frequency are given as follows:
P i = ρ + M i + ε i L i = ρ + λ i N i + m i + ϵ i  
where ρ is the double-differenced geometric range; M i and m i are the multipath errors in the pseudorange and carrier phase, respectively, λ 1 is the wavelength of the carrier; N i is the integer ambiguity, and ε i and ϵ i are random noises of the pseudorange and carrier phase, respectively. By considering three frequencies, i.e., i = 1 , 2 , 3 , as presented in reference [39], it is possible to define two robust double-differenced geometric ranges, ρ ~ 1 and ρ ~ 2 , which allows for the computation of the EWL and WL ambiguities that are given, respectively, as follows:
N ( 0 , 1,1 ) = r o u n d L 3 λ 3 λ 2 λ 2 ρ ~ 1 λ 3 ρ ~ 1 λ 2
N ( 1 , 1,0 ) = r o u n d L 1 λ 1 λ 2 λ 2 ρ ~ 2 λ 1 ρ ~ 2 λ 2 .
Equations (28) and (29) allow for the definition of the following system of equations:
N ( 1 , 1,0 ) = N 1 N 2 N ( 0 , 1,1 ) = N 3 N 2 L 1 = ρ + λ 1 N 1 + m 1 + ϵ 1 L 2 = ρ + λ 2 N 2 + m 2 + ϵ 2 L 3 = ρ + λ 3 N 3 + m 3 + ϵ 3
and by adopting an ILS method, it is possible to estimate ρ , N 1 , N 2 , and N 3 of the problems.

3.2. Levenberg–Marquardt

The Levenberg–Marquardt algorithm (LMA), or the Damped least-squares (DLS) method, is used to solve non-linear optimization problems. The LMA exploits a parameter λ such that it can switch between the GD algorithm and the Gauss–Newton algorithm; practically, the LMA interpolates between the Gauss–Newton algorithm and the method of GD according to the current prediction accuracy, but at the expense of a certain convergence rate [40]. The algorithm needs the initial guess and, if it is too far, then the error might be large, and the algorithm might not be able to give a correct global solution due to the presence of many local minima, and it might provide a local minimum which is not necessarily the global minimum. To obtain a better correct initial guess for the LMA, it is possible to use EAs in the initial phase because they do not suffer from the initial value problem.
Considering a non-linear least-squares problem, it can be expressed mathematically as follows:
f x = 1 2 j = 1 m r j 2 x = 1 2 r x 2
where x is the vector input, f is the objective function, r i is the residual and it is defined as a function such that R n R with the assumption of m n , and r x = ( r 1 x , r 2 x , , r m x ) T is the residual vector and is defined such that R m R n .
The Levenberg–Marquardt algorithm is a modification of the Gauss–Newton method, and it is based on a local linearization of the residuals as follows [40,41,42]:
r m x + δ x r m x + J m μ δ x
where J m μ = J x = r m x μ is the Jacobian Matrix, and δ x is an infinitesimal of x . Next, the Gauss–Newton technique recursively updates in accordance with the following:
δ x = J T J 1 f = J T J 1 J T r .
If the initial guess of the Gauss–Newton method is near the optimum of f , it will usually converge quickly; however, this is not always the case, and the method can take very large and uncontrolled steps, failing to converge. To deal with this issue and control the steps in a useful way, Levenberg [41] and Marquardt [42] each proposed the damping of the J T J matrix by a diagonal cut-off, modifying (33) as follows:
δ x = J T J + λ D T D 1 f
where D T D is a positive–definite, diagonal matrix that is the relative scale of the parameters and λ is a damping parameter adjusted by the algorithm. If λ is large, the steps are small along the gradient direction, and vice versa when λ is small. As the method is reaching a solution, λ is chosen to be small and the method converges quickly via the Gauss–Newton method.
The implementation of the Levenberg–Marquardt algorithm follows these iterative steps [40]:
  • Based on the current parameter values, update the function and Jacobian values (if necessary);
  • Update the scaling matrix D T D and damping parameter λ ;
  • Propose a parameter step δ x and evaluate the function at the new parameter values x + δ x ;
  • Accept or reject the parameter step depending on whether the cost has decreased at the new parameters;
  • Stop if any of the desired convergence criteria of the algorithm are met or if the limit of function evaluations or iterations has been exceeded.
The methods for the choice strategy of the damping parameter λ and the scaling matrix D T D require specific methods (as described in [40]) which are not further investigated in this article.

Magnetic Localization Technique Based on the Levenberg–Marquardt Algorithm

The authors in reference [10] propose a method for the localization of a permanent magnet based on the LMA. The permanent magnet is modelled as a dipole, and the magnetic field is measured by an array of 4 × 4 magnetometers (16 in total). The general optimization problem can be expressed as follows:
f a , b , c , θ , φ = i = 1 N B ^ i B i 2
where f is the objective function, a , b , and c are the coordinates of the permanent magnet position, θ and φ identify the orientation of the permanent magnet, N is the number of the i -th magnetometer, B ^ i is the measured magnetic field, and B i is a function representing the model of the system based on the magnetic dipole theory, and it includes the coordinates and the orientation angles.
The problem can be solved by using the LMA, but the issue is the need for a good initial guess to obtain a solution because when a too-far initial point is provided, divergence can occur. To deal with this problem, a solution could be to combine the PSO and LMA methods: the former is used to calculate an initial guess, and the second is used to calculate the local minima because it results in a faster method. The drawback of this approach is that it is not suitable for real-time applications, and the authors of paper [10] deal with it by proposing to use only the LMA and a properly modified model of Equation (35). In the first step, the new model neglects the orientation to find only the coordinates of the permanent magnet; then, in the second step, the orientation is also obtained. The modified model is expressed as follows:
f a , b , c = i = 1 N B ^ i B T R i 3 A i m ^ a v g 2
where B T is a scalar quantity that represents the magnetic strength, R i is the vector from the center of the permanent magnet to the i -th sensor, A i is a matrix that is a function of the magnet position, and m ^ a v g is the averaged orientation vector, expressed as follows:
m ^ a v g = i = 1 N m ~ i m ~ i i = 1 N m ~ i m ~ i
where m ~ i = R i 3 2 B T ( A i I ) B ^ i , and I is the one-diagonal matrix. The orientation vector is obtained by applying the following equation:
m ^ = R i 3 2 B T A i I B i
The authors claim that in such a way, a more suitable optimization problem for the LMA is obtained, and for this application, the method always provides an acceptable solution independently from the distance of the initial guess; moreover, the speed of computing is improved.

3.3. Lagrange Multipliers Method and Lagrange Duality

3.3.1. Lagrange Multipliers Method

The Lagrange multipliers method is an analytic method for finding the local maxima and minima of a function subject to equality constraints. The concept is to reformulate the original problem, obtaining a new function called the Lagrangian function, which allows for solving a new problem without constraints. Given a problem with an objective function f x subjected to the equality constraint h x = 0 , the Lagrangian function is expressed as follows [43]:
L x , λ f x + λ h ( x )
where λ is the Lagrangian multiplier. The solution is obtained by finding the stationary point of L by setting L x , λ x = 0 and L x , λ λ = 0 , which gives rise to the following system:
f x x + λ h x x = 0 h x = 0
The solution corresponding to the original problem is always a saddle point of the Lagrangian function [44]. Then, a possible solution of the system (40) ( x * , λ * ) allows for the finding of an optimal point of the original problem given by f ( x * ) .
The method of Lagrange multipliers is generalized by the KKT condition, also allowing for the consideration of inequality constraints; given a minimization non-linear problem (in which the x variable is a vector) with an objective function f x subject to equality constraints h i x = 0 (with i = 1 , , p ), and inequality constraints g j x 0 (with j = 1 , , m ), the Lagrangian function is set as follows:
L x , λ , μ f x + μ T g x + λ T h ( x )
where μ and λ are the Lagrange multipliers related to inequalities and equalities, respectively, and they are vectors. If ( x * , λ * , μ * ) is a saddle point of L x , λ , μ in the x domain, and the KKT conditions are satisfied, then x * identifies an optimal solution. The KKT conditions can be summarized as follows [45]:
g j x * 0 h i x * = 0 μ j * 0 μ j * g j x * = 0 f x * + j μ j * g j ( x * ) + i λ i * h i x * = 0
In general, once the gradient has been obtained, to determine if a constrained local optimum has been found, the KKT conditions are used to verify the necessary conditions for a local optimum; for unconstrained problems, the KKT conditions only require the gradient of the objective function to vanish at the optimum design point.
It is highlighted that the KKT conditions cannot indicate whether a global optimum has been found, and they are useful only for identifying a local optimum.

3.3.2. Lagrange Duality

The duality principle claims that an initial optimization problem called the primal problem can be transformed into a dual problem by applying a specific procedure. In such a way, it allows for the transformation of a minimization problem into a maximization one and vice versa. In some cases, the method might make the solution of the transformed problem easier and more manageable than the primal problem [14].
Mathematically, the Lagrange duality is defined as follows [46]. A not necessarily convex optimization problem is considered and is represented as:
m i n i m i z e : f 0 x s u b j e c t   t o : f i x 0 i = 1 , , m h i x = 0 i = 1 , , p
where x R n , with a domain D and optimal value p * . Then, by using (41), the Lagrangian function of the problem is obtained and given as:
L x , λ , ν f 0 x + i = 1 m λ i f i x + i = 1 p ν i h i x
where L x , λ , ν is such that R n × R m × R p R with a domain L = D × R m × R p , and λ i and ν i are the Lagrange multipliers related to inequalities and equalities, respectively. Now, it is possible to define the Lagrangian dual function g λ , ν of the problem as the infimum of the Lagrangian function over x and it is given as:
g λ , ν = inf xϵD   L x , λ , ν = inf xϵD   f 0 x + i = 1 m λ i f i x + i = 1 p ν i h i x
where g λ , ν is such that R m × R p R . It is observed that the infimum is unconstrained (unlike the beginning problem) and g λ , ν is concave; moreover, the function g could be for some λ , ν .
At this point, by adopting the lower bound property, if λ 0 , then g λ , ν p * ; this allows us to define the dual problem by maximizing the obtained Lagrangian dual function in (45). It is given as follows:
m a x i m i z e λ , ν   g λ , ν s u b j e c t   t o   λ 0
The solution of (46), denoted with λ * , ν * , provides for all convex and non-convex problems a lower bound for the optimum function such that g λ * , ν * f ( x * ) , like it is represented in Figure 6.
Furthermore, Figure 6 clearly represents the weak and strong duality principles. The former provides a bound for the optimal values of the transformed problem, whereas the latter claims that the optimal values of the two problems are equal. The strong duality is usually true for convex optimization problems.

3.3.3. A Localization Non-Convex Problem

The authors in reference [47] propose a method to obtain localization by exploiting only the range measurement data provided by wireless sensors. The robot localization is modelled as a non-convex optimization problem, and to obtain the solution, as a first step, it is reformulated as a convex problem, and then duality theory is applied to find, in an easier way, the optimum.
The localization problem is formulated as follows. The measured range data from anchors are given as follows:
d ^ k = x y k 2 + η k
where x and y k are two column vectors to determine the positions of the target and the anchors, respectively, η k is the corresponding measurement noise, k = 1 , , m identifies the k -th anchor, which are m in total. The optimal position can be obtained by minimizing the measurement noise between the measured distance d ^ k and the real distance d k = x y k 2 ; then, the objective function can be expressed as follows:
f 0 x = k = 1 m d ^ k x y k 2 2
The function in (48) is non-convex, and to solve the problem, the authors proceeded in this way. In the first step, they reformulated the problem into a convex-constrained problem by exploiting a change in the variable; the new problem is expressed as follows:
min   f z = A z b 2 2 s u b j e c t   t o   ( z T C z + 2 f T z ) = 0
where z = x T x T x = x T t T and coefficients are defined as A = 2 y 1 T 2 y 2 T 2 y k T 1 1 1 , b = d 1 2 y 1 2 2 d 2 2 y 2 2 2 d k 2 y k 2 2 , C = 1 0 0 0 1 0 0 0 0 , and f = 0 0 1 / 2 . Then, duality theory is adopted, and the obtained Lagrangian function is given as follows:
L z , v = A z b 2 2 + v z T C z + 2 f T z = z T A T A + v C z 2 A T b v f T z + b T b
where v is the Lagrange multiplier. Now, by computing the minima of L z , v over z , it is possible to obtain the Lagrangian dual function, and the final Lagrange dual problem is given as follows:
m a x i m i z e G ( v ) = A T b v f T A T A + v C A T b v f + b 2 2 s u b j e c t   t o ( A T A + v C ) 0 ( A T b v f ) ϵ R A T A + v C
where R identifies the domain. The problem in (51) is a convex problem and the solution can be computed.

3.4. Genetic Algorithm

The GA was inspired by Darwin’s principle of survival of the fittest solutions and can solve both constrained and unconstrained optimization problems; it is based on natural selection, the process that drives biological evolution [48]. The algorithm is implemented with the following steps:
  • Creation of the initial population.
  • Calculation of fitness of each individual.
  • New individual generation by adopting the following operators:
    3.1.
    Selection;
    3.2.
    Crossover;
    3.3.
    Mutation;
  • Calculation of fitness of each individual.
  • Test:
    5.1.
    If the convergence criteria are satisfied, stop.
    5.2.
    Otherwise, repeat from step 3.
It is iterative, and after the first cycle, it is repeated from step 3 until the desired criteria are satisfied. In step 1, the process begins with a set of individuals, which is called the population; each individual is a solution to the problem to be solved. An individual is characterized by a set of parameters (variables) known as genes, which are joined into a string to form a chromosome, i.e., a solution. In the second step, the fitness function of the problem determines how an individual (or solution) fits, and it measures the ability of an individual to compete with other individuals. The output is a fitness score for each individual, and the probability that an individual will be selected for reproduction is based on its score. In step 3, the new individual generation process begins, and it is composed of four operators, i.e., (i) selection, (ii) crossover, and (iii) mutation. In the selection, some individuals are chosen to generate new ones and selected individuals pass some of their genes into the next generation; in other words, some individuals are selected as parents to generate children. In general, individuals are selected based on their fitness scores, and the higher the fitness value, the higher the probability of being selected for reproduction. In the literature, there are many proposals for selection operators, and some of them are the following: roulette wheel selection, sequential selection, tournament selection, dominant selection, hybrid selection, kin selection, back controlled selection, and so on [49]. It is important to highlight that, at every cycle, an elitist population is extracted from the population of the current iteration to directly pass some of the best individuals to the next generation. In some cases, such a procedure might give rise to problems of fast convergence, and to tackle such an issue, an adaptive elitist population strategy which changes the number of the elitist population can be implemented, as proposed in reference [50]. After selection, the crossover operator creates new children by mixing the relative genes. For example, Figure 7 shows the concept of the 1-crossover point operator, where two parents generate two offspring by exchanging portions of their genes. In the literature, there are many proposals for crossover operators and the authors in reference [51] classify them in three categories: standard (such as 1-crossover point, k-crossover point, shuffle-crossover, reduced surrogate-crossover, and so on), binary (random respectful crossover, masked crossover, 1-bit adaption crossover, and so on), and application dependent (or real/tree).
Finally, the mutation operators can introduce random variation of the genes, and for this purpose, the mutation probability value is adopted, which is a random value between 0 and 100%; a value equal to zero means no-mutations are adopted, whereas a 100% value modifies all genes of the child. In general, mutations should not occur so often, but they are important because they allow for the avoidance of being stuck in local minima [52]. For example, some mutation operators in the literature are mirror mutation and binary bit-flipping mutation, mutation based on directed variation techniques, directed mutation, and so on [53].
Once a new population is obtained, step 4 computes the fitness values of the offspring and, finally, step 5 verifies whether the convergence criteria are satisfied or not. If not, the cycle is repeated from step 3 until a satisfactory solution is found or the maximum number of iterations is reached; over successive generations, the population evolves toward an optimal solution [15,48,54].
The GA can solve a variety of optimization problems that are not well-suited for standard optimization algorithms, including problems in which the objective function is discontinuous, nondifferentiable, stochastic, or highly non-linear. Furthermore, GAs are also suitable for problems of mixed-integer programming where solutions have some components that are defined as integers.

Evolutionary Optimization Strategy for Indoor Position Estimation Using Smartphones

The authors in reference [9] propose a smartphone-based indoor localization system adopting a sensor fusion algorithm where localization information from an Inertial Measurement Unit (IMU), Wireless Local Area Network (WLAN), Bluetooth, and magnetic field anomalies is fused to provide the estimated localization.
In the fusion step, the probability distributions of each localization information source are superimposed; in the summation, each distribution is weighted with a coefficient in the interval from 0 to 1.
For determining the weights, optimization methods can be applied; they exploit the result of some tests that were run in buildings to obtain the ground truth points (GTPs). In such a way, it is possible to estimate the weights with a Root-Mean-Square Error (RMSE)-based fitness metric, defined as follows:
F i t n e s s = t e s t   r u n = 1 N 10 R M S E t e s t   r u n
Due to the high ambiguity and the unknown (mixed) stochastic distributions of the input variables, using classical search methods (e.g., gradient-based methods), the optimization quickly converges to a local minimum. Then, a global optimization procedure is needed to maximize the quality metric and avoid local extrema. Because of the building model’s non-linearities, no gradients can be easily derived analytically, and gradient-based optimization methods are not applicable. Therefore, the authors proposed the use of a GA for finding the optimal weights, and in their article, they compared the results with four other local and global optimization strategies: Hill Climbing, the Nelder–Mead method, SA, and PSO. The outcome is that the GA outperforms the other algorithms; however, the drawback is that it needs the most computing time, but it is not a particular issue because the optimization process of the parameters must only be calculated once the end of the offline phase is reached.

3.5. Quantum Genetic Algorithm

The Quantum Genetic Algorithm (QGA) exploits the concept of qubits and the superposition of states in quantum mechanics. The algorithm is explained as follows [11,55]. The information is represented by a quantum bit or qubit which may be in the ‘0’ state, in the ‘1’ state, or any superposition of the two states [55]. A representation of the qubit state is given as follows:
Ψ = α 0 + β 1
where α and β are complex numbers that specify the probability amplitudes of the corresponding states, and α 2 and β 2 are the probabilities of the states to be either 0 or 1, respectively, and the condition α 2 + β 2 = 1 must be satisfied.
A system with m qubits represents 2 m states at the same time, which collapse into a single state during the observing process of a quantum state.
In general, the traditional GA can use binary, numeric, or symbolic representation for a solution encoded with a chromosome. The QGA uses the qubits, where one qubit is represented as α β , whilst m qubits are represented as follows:
α 1 β 1 α 2 β 2 α m β m
where α i 2 + β i 2 = 1 for i = 1 , 2 , , m . For example, considering m = 3 and the following chosen amplitudes of probability, a superimposition of states is represented as follows:
1 2 1 1 2 1 2 0 3 2 ,
and the system state is expressed as follows:
1 2 2 000 + 3 2 2 001 + 1 2 2 100 + 3 2 2 101 .
It has four pieces of information at the same time, i.e., the probabilities of 000 , 001 , 100 , and 101 are 1 8 , 3 8 , 1 8 , and 3 8 , respectively. For a better comprehension of how coefficients are obtained, the coefficient 1 2 2 of the binary number 000 is obtained by multiplying each α i in the matrix (55), i.e., α 1 α 2 α 3 = 1 2 * 1 * 1 2 , whereas the coefficient 3 2 2 of the number 101 is obtained by β 1 α 2 β 3 = 1 2 * 1 * 3 2 . The square power of a coefficient gives the probability of the related binary number. Moreover, the second digit is always zero, because α 2 = 1 .
Because QGA computing can represent a superposition of states, it provides a better characteristic of diversity than classical approaches; Equation (56) is one qubit chromosome that represents four states at the same time, whereas, in a classical representation, at least four chromosomes (000, 001, 100, and 101) would be needed.
During the convergence process, the α 2 or β 2 of the qubit chromosomes can approach 1 or 0, giving rise to a single state.
In the QGA, a population of qubit chromosomes Q t = q 1 t , q 2 t , , q n t at generation t is defined, where n is the size of the population, and q j t is a qubit chromosome with m qubits defined as follows:
q j t = α 1 t β 1 t α 2 t β 2 t α m t β m t .
Considering the main step to implement the QGA, in the initialization step of Q t , and all α i t and β i t of each q j t are set to 1 / 2 and it means that the qubit chromosome at t = 0 represents the linear superposition of all possible states with the same probability 1 2 m , and it can be expressed as follows:
Ψ q j 0 = k = 1 2 m 1 2 m S k
where S k is the k -th state represented by the binary string x 1 x 2 x m , where x i can be equal to 0 or 1.
The next step is to generate a set of solutions from Q t , and this is called the observing process, which provides P t = p 1 t , p 2 t , , p n t at the time t .
One binary solution, p j t , for j from 1 to n , is a binary string of length m , and is formed by selecting each bit using the probability of qubit α i 2 (or β i 2 ), which is compared with an r i value that is a randomly generated number between 0 and 1; if r i α i 2 , then the i -th element x i of the j -th binary string p j t is 1, or, if r i < α i 2 , it is 0. Each solution p j t is evaluated to give a measure of its fitness. The initial best solution is then selected and stored among the solutions P t .
After initialization, the iterative part of the algorithm begins, and a set of binary solutions P t is formed by observing the Q t 1 population; each binary solution in P t is evaluated to give the fitness value. In the next step, called “update Q t ”, a set of qubit chromosomes q j t is updated by applying an appropriate quantum gate U t . A quantum gate is a reversible gate represented by a unitary operator U acting on the qubit basis states, and it satisfies the relation U + U = U U + , where U + is the Hermitian adjoint of U . The quantum gate is selected or designed in agreement with the considered practical problem to be solved; for example, the rotation gates can be used for knapsack problems, and it is given as follows:
U θ = cos   ( θ ) s i n   ( θ ) sin   ( θ ) cos   ( θ )
where θ is the rotation angle. For instance, a qubit amplitude probability α i β i T in a chromosome can be updated by a quantum rotation that is given as follows:
α i β i = U θ i α i β i
where θ i = s α i , β i Δ θ i , and s ( α i , β i ) is a function to determine the sign of θ i , and Δ θ i is the variation of θ i 1 . The functions s ( α i , β i ) and Δ θ i are chosen in agreement with the lookup table (as indicated in reference [55]), which is based on experience and experiments; Δ θ i is a key parameter because it influences the computation quality and efficiency of the method; if it is too small, the number of iterations might be too big, whereas when it is too large, the solution might be trapped into divergent locations or give rise to a premature convergence to a local optimum [11].
Thanks to this step, the qubit chromosomes move and can converge to a better solution. In the next step, among the P t solutions, the best one is selected and compared with the previous best one; if the new solution is better than the previous one, the older solution is discarded; otherwise, the old one is retained. By the application of some genetic operators, it is also possible to obtain mutations and crossovers of chromosomes; the mutation gives rise to new individuals characterized by a small change, whereas the crossover allows the generation of new individuals by combining parts of two or more different chromosomes. Genetic operators make the probability of the linear superposition of states change; however, mutation and crossover are not needed in the QGA because the quantum representation provides a diversity in chromosomes and the application of genetic gates with a high probability of variation can result in worse performance.
The diversity of the QGA can provide better convergence than conventional GAs, which can deal only with fixed 0 and 1 information. Moreover, the QGA has a speed of calculation N times faster than the speed of the traditional genetic method. The main steps of the QGA are summarised as follows:
  • Initialize of Q t = 0 .
  • t = t + 1.
  • Observe Q t 1 and create P t .
  • Evaluation of P t .
  • Update Q t using quantum gates U t
  • Store the best solution among P t .
  • If max generation is not reached or convergence criteria are not satisfied, repeat from step 2.

An Adaptive Quantum-Inspired Evolutionary Algorithm for Localization of Vehicles

The authors in reference [11] introduce a method based on an Adaptive Quantum-Inspired Evolutionary Algorithm (AQIEA) to estimate the calibration parameters for a real-size intelligent vehicle that can be driven along a non-predefined road path. They use the Differential Global Positioning System (DGPS) as a ground truth and propose an optimization-based calibration method to reduce the systematic errors of wheeled odometers (the differences between the actual wheel diameters and the nominal diameters) and the gyroscope (initial installation error due to misalignment of the sensor with the longitudinal axis of the vehicle). The diameter of the left and right rear wheels and the heading of the gyroscope are considered as the calibration parameters to be optimized to reduce the localization errors in the dead reckoning.
The dead reckoning model can be expressed as follows:
x t + Δ t = x t + Δ d · cos   ( ϑ t + Δ ϑ / 2 ) y t + Δ t = y t + Δ d · sin   ( ϑ t + Δ ϑ / 2 ) ϑ t + Δ t = ϑ t + Δ ϑ
where x , y , and ϑ are, respectively, the coordinates of the vehicle along the x and y -direction in the Cartesian coordinate system and the heading of the vehicle at time t ; Δ t is the sampling period; Δ d is the incremental travel displacement of the vehicle, and it can be expressed as Δ d = Δ d r + Δ d l 2 , where Δ d r ( l ) is the displacement of the right (left) rear wheel of the vehicle, and Δ ϑ is the heading change during the sampling period.
The wheel displacement can be expressed as follows:
Δ d r = Δ N r l R r l π D r ( l )
where Δ N r l is the increment of the encoder pulses of the right (left) during the sampling period, R r ( l ) is the encoder resolution of the right (left) wheel in the unit of pulses per revolution, and D r ( l ) is the nominal diameter of the right (left) wheel.
In the real application, this model is influenced by systematic errors and random errors; for example, considering the odometers, the former are due to differences between wheel diameters, different nominal diameters, and the misalignment of wheels; the latter is due to the limited resolution of encoders, the limited sampling rate, the condition of the road (rough surface), the wheel slippage, and so on.
The systematic errors can be compensated and by including them in the model, the displacements and heading variation are expressed as follows:
Δ d ~ r = Δ d r + c 1 Δ d ~ l = Δ d l + c 2 ϑ ~ t = ϑ t + c 3
where the symbol ~ denotes the new variables that include systematic errors, and c 1 , c 2 , and c 3 are the parameters to compensate for the right wheel diameter, the left wheel diameter, and the misalignment of the gyroscope, respectively. By considering these errors, the model represented by Equation (61) is rewritten as follows:
x ~ t + Δ t = x ~ t + Δ d ~ r + Δ d ~ l 2 · cos   ( ϑ ~ t + Δ ϑ ~ / 2 ) y ~ t + Δ t = y ~ t + Δ d ~ r + Δ d ~ l 2 · sin   ( ϑ ~ t + Δ ϑ ~ / 2 ) ϑ ~ t + Δ t = ϑ ~ t + Δ t .
The positioning error E i at the i -th sample is defined as the difference between the data provided by DGPS and the dead reckoning method, and it is computed by the following:
E i c 1 , c 2 , c 3 , p i G P S , p i D R = x i G P S x ~ i 2 + y i G P S y ~ i 2
for i = 1 , 2 , , N where N is the total number of samples.
Then, it is possible to formulate the optimization problem, which is expressed as follows:
m i n i m i z e : M c 1 , c 2 , c 3       s u b j e c t   t o : c 1 L c 1 c 1 U c 2 L c 2 c 2 U c 3 L c 3 c 3 U
where M is the objective function to be minimized with respect to c 1 , c 2 , and c 3 , and the superscript L and U denote the low and upper limits, respectively, for the considered variables. Two different optimization objective functions are considered to determine the optimal calibration parameters; the first one is the maximum positioning error, which is defined as follows:
M = m a x   E i ,
and the second one is the mean positioning error, defined as follows:
M = 1 N i = 1 N E i
To solve the optimization problem, the authors use the AQIEA, which is an evolution of the QGA. In the adaptive method, the lookup table is not used because it is substituted by a different approach based on the idea of choosing as input for the quantum gate either a large value for Δ θ i or a small one. The former is used when an individual is away from the optimal solution, and it needs to bring individuals quickly towards the optimal solution; the latter is used to avoid an individual moving away from the optimal solution when it is going to be reached. Then the magnitude of the rotation angle is obtained by the following:
Δ θ i = θ m a x θ m a x θ m i n f f a v e f m a x f a v e     f > f a v e θ m a x f f a v e
where θ m a x and θ m i n are two positive real numbers ( 0.001 π θ m i n < θ m a x < 0.5 π ), f m a x is the maximum fitness in the current population, f a v e is the average fitness of the individuals in the current population, and f is the fitness of the selected individual. In such a way, an adaptive change in the rotation angle magnitude of the quantum rotation gate is obtained.
The sign of the rotation angle is determined by the following:
s α i , β i = s g n o i * c α i β i α i β i 0 0 α i = 0   a n d   o i * = 1   o r   β i = 0   a n d   o i * = 0 ± 1 o t h e r w i s e
where s g n is a function that provide the sign, o i * is the observed state (0 or 1) of the i -th qubit of the best individual, and c is a positive real number between 0 and 1. The proof of (70) is given in reference [11].
Moreover, to improve global searching and to avoid being trapped in local optimal minima, an adaptive quantum mutation operation is introduced (unlike the traditional QGA, where mutations are not used). Given a selected chromosome in the population, the mutation operation is conducted based on a mutation probability p m :
p m = p m a x p m a x p m i n f f a v e f m a x f a v e f > f a v e p m a x f f a v e
where p m a x and p m i n are two small positive real values between 0 and 1. It is observed, for example, that for an individual with f = f m a x , p m is 0 and no mutations are applied. The mutation is applied by choosing two random positions in the selected individual, and two qubits are mutated by a quantum-not gate that exchanges the values of the amplitude probabilities; the gate is defined as follows:
U θ = 0 1 1 0 .

3.6. Differential Evolution Algorithm

The DE algorithm is a power population-based and stochastic evolutionary algorithm to solve continuous global optimization problems that can be constrained or unconstrained; it maintains a population of candidate solutions and creates new candidate solutions by combining existing ones according to its formula and then keeping whichever candidate solution has the best score or fitness on the considered optimization problem. In this way, the optimization problem is treated as a black box that provides a measure of quality given by a candidate solution.
The main steps of the DE algorithm are as follows: (i) generate the initial population of the specified size; the population evolves by using (ii) mutation, (iii) crossover, and (iv) selection operators during each generation; and (v) step (ii) is repeated until a termination criterion is obtained [56].
In the initialization step, a number N of populations is set and each of them has a number M of D -dimensional vectors ( D is a number related to the nature of the problem), which are potential solutions to the problem. The i -th individual, i.e., the i -th D -dimensional vector, is given as follows:
x i , G = { x i , G 1 , x i , G 2 , , x i , G D }
where G identifies the number of the generation. Initial individuals are set with the following formula:
x i , 0 = x m i n + r a n d 0,1 ( x m a x x m i n )
where x m i n and x m a x are vectors which identify the limits of the solution, and r a n d 0,1 is a function to generate random individuals with a normal distribution.
In the mutation step, mutation vectors v i , G are generated for each individual of a population, which is called the target vector. The basic mutation strategies are as follows [57]:
( 1 )   DE / rand / 1 :   v i , G = x r 1 , G + F ( x r 2 , G x r 3 , G )
( 2 )   DE / best / 1 :   v i , G = x b e s t , G + F ( x r 1 , G x r 2 , G )
( 3 )   DE / current   to   best / 1 :   v i , G = x i , G + F ( x b e s t , G x i , G ) + F x r 1 , G x r 2 , G
( 4 )   DE / best / 2 :   v i , G = x b e s t , G + F x r 1 , G x r 2 , G + F ( x r 3 , G x r 4 , G )
( 5 )   DE / rand / 2 :   v i , G = x r 1 , G + F x r 2 , G x r 3 , G + F ( x r 4 , G x r 5 , G )
where r 1 , r 2 , r 3 , r 4 , and r 5 are integers randomly generated in the range between 1 and M , F is a scale factor, and x b e s t , G is the best individual vector in a population during the G -th generation.
As for the crossover step, trial (or test) vectors u i , G are created by crossing each individual x i , G (or target vector) and the corresponding mutation vector v i , G ; it needs to form a set of trial vectors used in the selection step; there are two types of crossover methods largely used in the literature: binomial and exponential. In the former, the test vector is given as follows:
u i , G j = v i , G j i f   r a n d i , j ( 0,1 ) C R x i , G j o t h e r w i s e
where CR is the crossover factor in the range 0 and 1 and controls the probability that a component of a trial vector is obtained from either a mutant or a previous individual, and j is between 1 and D . In this method, the distribution of the number of parameters inherited from the mutant is (almost) binomial. On the other hand, in the latter, a random integer r in the range [ 1 , D ] is chosen and it is the component starting point for the crossover process. The components for the trial vector u i , G j are donated from the mutated vector until a Bernoulli experiment linked to a probability C R is true [58].
u i , G j = v i , G j j = r , r + 1 , u n t i l   B e r n u l l i   e x p e r i m e n t   i s   t r u e   o r   j = D x i , G j o t h e r w i s e .
In the end, in the selection step, the objective function of the problem is evaluated for all generated trial vectors and the corresponding target vectors; the best one is selected, and it is a new individual of the next generation. The selection is given as follows:
x i , G + 1 = u i , G i f   f ( u i , G ) f ( x i , G ) x i , G o t h e r w i s e
There are two main differences between the GA and DE algorithm: (i) the GA is conducted by changing some genes in the chromosomes, while in the DE algorithm, the mutant individuals are obtained by adding the difference between two individuals to a third individual, which has a stronger global search capability; (ii) in the traditional GA, offspring individuals replace their parent individuals with a certain probability, while in the DE algorithm, such a condition only happens when the fitness of the offspring is better than the fitness of their parents; in such a way, a greatly increased convergence speed is provided [59].

Visible Light Communication Using a Modified Differential Evolution Algorithm

Authors in reference [59] have developed a Visible Light Communication (VLC) system where the VLC positioning model is transformed into a global optimization problem and a DE algorithm is used to calculate the z-coordinate of the receiver by adopting a modified DE algorithm that is called the self-adaptive DE algorithm; in this work, the 3D problem is reduced to a 1D problem, improving the speed of localization. Because the individual dimension is one (only the z-coordinate), the crossover process is ignored. The 3D positioning strategy exploits some specific equations to calculate the overlapped area of three circles determined by the position of the sources and the target, and is obtained by the projection on the x -y plan; the DE algorithm is used to solve the specific equation and calculate the z-coordinate by finding the optimal fitness value; the x and y coordinates are calculated by exploiting three equations for trilateration in the projection plan.
The proposed algorithm has three steps: (i) the generation of the population, (ii) calculation of the fitness, and (iii) mutation operator. The fitness function evaluates the quality of each of the candidate solutions in the population. In the VLC-based 3D positioning system, the individual with the smallest overlapped area can be regarded as the estimated position. The fitness equation can be treated as the overlapped area S and is represented as follows:
f i t n e s s z = S
If no overlapped area is formed between the three circles, the fitness is zero; the presumed height z is increased by 0.2 m for each iteration until the overlapped area appears. If the fitness 10 7 , then the currently estimated height jumps out of the iteration loop and it is considered as the optimal solution z e ; otherwise, the iteration process goes on (step (iii)). In the case that the number of iterations reaches the maximum, and the fitness is still higher than 10 7 , the algorithm goes back to step (i).
The authors considered three types of DE algorithms: (i) DE/rand/1/binomial; (ii) DE/current-to-best/1/binomial; and (iii) the proposed self-adaptive DE algorithm. The first and the second are two traditional DE algorithms which adopt the strategy given in (75) and in (77), respectively; furthermore, the DE/current-to-best/1/binomial benefits from its fast convergence because the best solutions in the search space are used and its performance is better than the performance of the DE/rand/1/binomial algorithm. As for the third algorithm, the proposed self-adaptive DE algorithm is derived on the following observation: as the iteration number increases, the difference among individuals is reduced, and therefore a smaller amount of adjustment is needed because an individual has a better fitness and most of its information can be retained; on the contrary, when the number of iterations are low, the solution is still far from the optimal solution and a larger scaling factor might help the exploration process. Then, the authors proposed the following variation:
v i , g + 1 = x b e s t , g + F ( x r 2 , g x r 3 , g )
where F is defined as follows:
F = f i t x i f i t ( x b e s t ) σ + f i t x w o r s t f i t ( x b e s t ) × F m a x F m i n + F m i n
where x b e s t and x w o r s t are the best and worst individuals in the population, namely the vectors with the lowest and highest fitness values in the entire search space, and σ = 10 13 is a constant added to avoid the zero division. In such a way, the scaling factor varies between the maximum F m a x and minima F m i n values according to its difference with the best individual at each generation.

3.7. Mind Evolutionary Algorithm

The Mind Evolutionary Algorithm (MEA) is a method which emulates the learning modes and activities of the human mind. Unlike genetic-based algorithms, which use mutation, crossing, and selection operators, in the MEA, an evolutionary process is also exploited but with the difference that two operations named similartaxis and dissimilation are adopted to overcome the drawback of the bad solutions generated during the mutation process; furthermore, every generation’s evolutionary information is memorized, and it is used to guide the evolution toward the global optimum [60].
The MEA steps are defined as follows:
  • Initialization.
  • Similartaxis and local competition.
  • Dissimilation and global competition.
In the initialization, two main spaces are considered: the superior group space, which is composed of S superior groups, and the temporary group space with T temporary groups. Every group is initialized with m individuals, which are generated by considering an individual as the center and adopting a uniform distribution to generate the other individuals around the center. Furthermore, every individual is scored in agreement with their adaptability to the search space [60,61].
The similartaxis and local competition phase aims to make the subgroups mature by performing a local competition. In every group, individuals compete with each other to find the local optimum. To create a new generation, a winner who has the best score is selected, and a new generation is created around it. Then, individuals are scored again to identify if there is a better new solution or to keep the previous one; in such a way, a new generation can be created, competition starts again, and so on. The information of every new winner is recorded on the local billboard of the group. The generation process in a group continues until the maximal and minimal scores within one group satisfy the mature condition; then, the information related to the mature group is stored in the global billboard. The process to make the group mature is named similartaxis and is performed for both superior and temporary space [60,61].
In the dissimilation and global competition phase, all groups compete against each other by observing the information reported on the global billboard. If, in the temporary space, there is a group with a higher score than any group in the superior space, then this temporary group will replace the bad groups in the superior space. Moreover, if the score of any mature group in the temporary space is lower than the score of any group in the superior one, this temporary group will be discarded, and new ones will be generated in the temporary space with random values. The generation process is called dissimilation, and in such a way, the similartaxis can continue in every group [60,61].
The aforementioned operations are repeated until the scores of the groups in the superior space are so high that they cannot be further improved. At this point, the algorithm is considered convergent, and the winning individual of the group with the best score in the superior space is the global optimum [61]. The architecture of the MEA is reported in Figure 8 [60].

A Mind Evolutionary Algorithm-Based Geomagnetic Positioning System for Indoor Localization

The authors in reference [60] propose a solution to solve the problem in the pedestrian dead reckoning field where a large error is introduced when a walker goes backwards or laterally but the position is still updated forward. To address this problem, they propose a system for the indoor localization of smartphones based on magnetic positioning and the pedestrian dead reckoning (PDR) method, which are fused using an Extended Kalman Filter (EKF).
To improve the performance of magnetic positioning, the MEA-based heuristic searching strategy is adopted to search for the optimal magnetic position.
The geomagnetic features are extracted using a three-axis magnetometer installed on a smartphone. The magnetic vector measured in the device coordinate reference system is transformed into the equivalent vector in the geographic coordinate reference system in order to obtain a measurement that is independent of the attitude of the device. The transformation can be expressed as follows:
B G = C n b T B D
where B G is the magnetic vector in the geographic coordinate system, C n b is the matrix for the transformation, and B D is the magnetic vector in the device coordinate system.
Then, five magnetic features for matching are extracted from B G and they are the three magnetic vector components, the horizontal component of the magnetic field, and the magnitude of the vector.
A fine-grained magnetic database is constructed by linearly interpolating the features and the coordinates. Once the fingerprint has been obtained, the MEA can be performed to estimate the magnetic position. The proposed algorithm exploits the large number of samples that are provided by the measurement process, which can be considered to be at a high frequency; over an interval time of one second, these samples generate many temporary positions that are used to generate a population for the MEA; by considering an epoch every second identified by k , the population M ( k ) can be represented as follows:
M ( k ) = { g k x 1 , y 1 , , g k x i , y i , , g k x n , y n }
where g k x i , y i is the i -th individual (i.e., the estimated position) of M ( k ) , x i and y i are the coordinates values, i = 1 , 2 , , n is the number of the sample, and n is the sampling frequency. The score is calculated considering the following formula:
s c o r e { k , i } = 1 G k 1 g k ( x i , y i ) 2
where G k 1 is the previous estimated magnetic position.
As the first step, the MEA initializes the population (using the data generated by measurements), and the temporary individuals (i.e., the geomagnetic positions) are scored using Equation (88). The positions with higher scores are selected as centers of the superior groups, whereas the positions with lower scores are selected as centers of the temporary groups. Around the centers, individuals of the population are generated with a random distribution, and they are scored using Equation (88).
As the second step, the MEA executes the similartaxis where individuals within the same subgroup compete by comparing their scores. The second step is completed when the mature condition s c o r e m a x s c o r e m i n ε is obtained.
In the end, as the third step, the MEA executes dissimilation for the global competition. The MEA executes similartaxis and dissimilation until convergent conditions are satisfied and the individual with the best score in the mature superior groups is selected as the estimation of the final magnetic position.

3.8. Particle Swarm Optimization

PSO algorithms are based on a simplified social model that tries to emulate the behavior of a swarm of animals such as bees or birds that are searching for a food source. It can be explained as follows [15,62].
PSO finds a solution by exploiting information from each individual of the population (or swarm) and also information from the swarm as a whole. At the beginning of the process, the number of particles in the swarm and the number of iterations are set, and an initial population distributed throughout the design space is randomly generated. In the next step, the position of each particle is updated using the following formula:
x i q + 1 = x i q + v i q Δ t
where i is the i -th individual in the swarm, q is the q -th iteration, v i q is the velocity vector of the i -th individual at the q -th iteration, and Δ t is the time increment. At the beginning, the velocity vector of each particle is initialized randomly and is updated at each iteration using the following formula:
v i q + 1 = ω v i q + c 1 r 1 ( p i x i q ) Δ t + c 2 r 2 ( p g x i q ) Δ t
where ω is the inertial parameter, r 1 and r 2 are random numbers between 0 and 1, c 1 and c 2 are the trust parameters, p i is the best point found so far by the i -th particle, and p g is the best point found by the swarm. The search behavior of the algorithm is controlled by the inertial parameter ω ; the larger the value, the larger the focus in global searching, whereas the smaller the value, the larger the focus in local searching; reference values for these two extremes are 1.4 and 0.5, for either a more focused global search or for more focused local search, respectively [15]. The c 1 and the c 2 are two trust parameters; the former, also referred to as cognitive memory, indicates how much the particle trusts itself, whereas the latter, also referred to as social memory, indicates how much the particle trusts the group. The p g can be defined as either considering some best points among a subset of particles or considering the best point among all particles of the swarm. The values of ω , c 1 , and c 2 parameters must be tuned in agreement with the managed problem.
The main drawbacks of the method are the tuning parameters and the fact that it is an inherently unconstrained optimization algorithm. The last one is a common characteristic of most of the evolutionary algorithms; researchers have tried to deal with this problem by proposing many constraint-handling techniques for evolution, and among these, the penalty function approach is the most widely used because it has the advantage that it is easy to implement [15].

3.8.1. Particle Swarm Optimization-Based Artificial Neural Network (PSO-ANN)

The authors in reference [63] use an Artificial Neural Network (ANN) to solve the fingerprint localization problem by exploiting the Received Signal Strengths Indicator (RSSI) of the Wi-Fi signal in an indoor environment. The ANN provides the localization of the target, and it is trained by a traditional PSO algorithm; this approach provides a strong learning skill which allows one to better deal with complicated indoor environments. Traditionally ANNs are trained by exploiting a backpropagation algorithm that is characterized by a feedforward propagation of signals and a backpropagation of errors; this approach requires significant time for training and can be easily trapped in a local minimum. To solve these issues, authors adopt the PSO algorithm to find the parameters of the trained ANN; they define the following:
(i)
Each fingerprint is composed by F = [ φ i t , P ] , where φ i t is the RSSI value at time t of the i -th Wi-Fi source, and P is the position of the collected fingerprint.
(ii)
The particle of the PSO is given as p a r t i c l e j = [ W 1 , , W n , B 1 , , B n ] , where W j and B j represent the weight and bias of the j -th neuron.
(iii)
The updating equations for the particle position and velocity are defined by the following:
p k + 1 = p k + v k + 1 v k + 1 = ω v k + r 1 c 1 × p b e s t k p k + r 2 c 2 × g b e s t k p k
where v k is the velocity in the k -th iteration, p k is the particle in the k -th iteration, c 1 and c 2 are two acceleration coefficients, and p b e s t k and g b e s t k are the personal best position and global best position in the k -th iteration, respectively, r 1 and r 2 are two random constriction coefficients in the range (0, 1) sampled from a uniform distribution to prevent explosion and induce particles to converge on optima, and ω is the inertia factor which helps to control the ability of particles to jump out of the local optimum and find the global optimum.
Considering the above definitions, the proposed algorithm follows these steps:
  • Set the particle group size M, the max iteration number m , and the expected error e. Then, initialize randomly the initial position and velocity of the particles;
  • For all RSSI training sets [ φ i t , P ] , put [ φ i t ] in the ANN model and compute the ANN output P ~ ;
  • Update the pbest and gbest according to the distance between the predicted position P ~ and the real position P ;
  • Update the particle position and velocity according to Equation (91);
  • Repeat steps 3 and 4 until the distance between P ~ and the real position P is less than the expected error or the max iteration number n has been reached.

3.8.2. Particle Swarm Optimization-Based Minimum Residual Algorithm

The authors in reference [64] adopt an RSSI range-based localization method for a mobile robot working in an indoor environment where the localization is estimated by a Minimum Residual (MinRe) localization algorithm based on PSO.
According to the principle of MinRE localization, the sum of the target’s squared residual between the real and the measured distances is given as follows:
f ( X ) = i N X A N i d i m e s 2
where f ( X ) is the function to be minimized, X is the real vector position to be found on the target, A N i is the vector position of the i -th anchor node, N is the total number of nodes, and d i m e s is the measured distance from the RSSI, which can be expressed as d i m e s = d i + ω i , where d i is the unknown real distance and ω i is the noise.
The term X A N i represents the Euclidean distance between X and A N i and it can be expressed as follows:
X A N i = x x i 2 + y y i 2 z z i 2
where x , y , and z are the coordinates of the position of the target X , and x i , y i , and z i are the coordinates of the position of the A N i . If there is no error, f ( X ) would be zero.
The PSO algorithm is applied to estimate the position of the target and the final problem can be written as follows:
X = arg   m i n   f ( X ) = i N x x i 2 + y y i 2 + z z i 2 d i m e s 2
Setting the number of particles M , the maximum number of iterations W , considering w as the index of the particle updating iteration for this problem, and m as the index of the m -th particle, the solution of the problem follows these steps:
  • Initialization, where the particles and their initial velocity are generated randomly within the target environment: X m 1 = [ x m 1 , y m 1 , z m 1 ] , and V m 1 = [ v x m 1 , v y m 1 , v z m 1 ] . The fitness f ( X m 1 ) of the m -th particle is calculated and the initial individual best value is set: B m 1 = X m 1 . The initial global best fitness of the particle swarm is the particle which achieves the minimum of f ( X ) and it is represented as B g 1 = a r g m i n   f ( X m 1 ) where m = 1 : M .
  • Particle velocities and positions, which are updated in agreement with the individual best and global best values at ω -th iteration for 2 w W .
    V m w = ω V m w 1 + c 1 ξ B m w 1 X m w 1 + c 2 η B g w 1 X m w 1   X m w = X m w 1 + V m w 1
    where ω is the inertia weight that controls the exploration search region, c 1 and c 2 are the acceleration constants between 0 and 2, and ξ and η are random numbers between 0 and 1.
  • The updating of the optimal particle and optimal swarm is initiated:
    B m w = B m w 1 ,             f ( X m w ) > f ( B m w 1 )                           X m w ,                     f ( X m w ) f ( B m w 1 )                         B g w = B g w 1 ,                                                                             m i n m = 1 : M f ( X m w ) > f ( B g w 1 )   arg   ( m i n m = 1 : M f ( X m w ) ) ,               m i n m = 1 : M f ( X m w ) f ( B g w 1 ) .      
  • The stopping criterion is checked: if w < W , then set w = w + 1 and go to step 2; otherwise, the estimated position is X ~ = B g w .
The estimated X ~ is the position at the time t where the localization is estimated.

3.8.3. Mobile Robot Localization Based on PSO and Simultaneous Perturbation

The authors in reference [65] propose a mobile robot localization based on the PSO estimator where the system exploits the dynamic optimization to find the best robot pose estimate, recursively. The proposed solution is an alternative to using EKF or PF. PSO searches recursively and stochastically along the state space for the best robot pose estimation. Unlike the localization based on PF, in the proposed method, the resampling step (where samples with negligible weights are discarded and substituted by resampling) is not required, and no noise distribution is required either. The robot localization x t at the t -th epoch is modelled like a Markov process (a random process in which the future is independent of the past given the present), where the initial distribution is p ( x t ) . Considering a maximum, a posteriori point of view, the localization estimation can be considered as an optimization problem, where the process seeks to estimate the pose that maximizes the a posteriori probability density, defined as follows:
arg   max x   p ( x t | z t )
where x t = x 0 , , x t is the history of the estimated positions, and z t = z 0 , , z t is the history of the observations. By applying Bayes, the posterior over the robot path is given by the following:
p x t | z t = η p ( z t | x t , z t 1 ) p ( x t | x t 1 , z t 1 ) p ( x t 1 | z t 1 )
where η is a normalizing constant. The problem can be reformulated as follows:
arg   max x   i = 1 t p z i | x i , z i 1 i = 1 t p x i | x i 1 , z i 1 p ( x 0 )
where p x 0 = p ( x t 1 | z t 1 ) and i is an index to count the epochs. After some mathematics passages, the objective function can be expressed as follows:
f 0 x t = log   p z t x t + log   p ( x t | x t 1 ) + f 0 ( x t 1 )
where f 0 ( x t 1 ) considers the previous history for all the t 1 epochs; it is suitable for an iterative application.
Now, the following is considered: (i) the motion model that gives the robot pose at time t . Specifically, this is considered as follows:
x t = f x t 1 , u t + ω t 1 ~ p ( x t | x t 1 , u t )
where f is a non-linear function, ω t is the Gaussian noise distribution with covariance matrix Q t , and u t is the control; (ii) the measurement model for the observed Q t feature is defined as follows:
z t = h x t , θ + v t ~ p θ t ( z t | x t )
where h is the function of the model, v t is the measurement Gaussian noise distribution with covariance matrix R t ; (iii) the state transition probability density p ω t ( x t | x t 1 ) derived from the robot motion model and the observation probability density p υ t z t x t derived from the observation model the problem is rewritten in the following form:
min x t   ( log   p υ t z t x t + log   p ω t ( x t | x t 1 ) )
where p υ t z t x t exp   1 2 z t z ^ t R t 1 z t z ^ t T and p ω t ( x t | x t 1 ) exp   1 2 x t x ^ t Q t 1 x t x ^ t T , where z ^ t and x ^ t are the estimated measurement vector and the estimated state vector, respectively; then, the fitness function can be written as follows:
F i t n e s s = z t z ^ t R t 1 z t z ^ t T + x t x ^ t Q t 1 x t x ^ t T .
It is a quadratic and strictly convex function; the problem is solved using PSO, which can be applied to complex problems such as non-linear and non-differentiable problems.
It is now assumed that the current particles of PSO are given as follows:
C t G = x t , 1 G , , x t , N p G
where C t G is an N p dimensional vector, and x t , i G represents each candidate solution i to the optimization problem at iteration G at time t . In standard PSO, the local information of an objective function obtained, for example, from the gradient, is not exploited. This means that even if a search point of PSO around the optimal value exists, the particle could go past the optimum point, since local information is not considered; nevertheless, the searching ability of PSO can be improved when gradient information is considered, but with the risk of being trapped in a local minimum and missing the global solution of the problem. In order to overcome this problem, authors combine PSO with the Simultaneous Perturbation (SP) method and gradient. The SP is a method that is given as follows:
x t + 1 = x t + a Δ g t Δ g t j = f x t + c t f ( x t ) c t j          
where a is a positive constant, c t and c t j are a perturbation vector and its j -th element, which is determined randomly, and Δ g t j represents the j -th element of Δ g t , which estimates the gradient.
As proposed in reference [66], the combination of PSO and SP gives rise to the following equations:
x t + 1 = x t + Δ x t Δ x t + 1 = χ ( a Δ g t + ϕ 1 p t x t + ϕ 2 n t x t )
where x t is the particle, p t is the best value of the particle, n t is the best value of the swarm, ϕ 1 and ϕ 2 are two positive numbers in a predefined range, and χ is a gain coefficient. To provide global and local search capabilities at the same time, and keeping in the meantime population diversity, the authors define a combination of PSO and SP, given as follows:
x t i = x t 1 i + v t i
v t i = v m a x v t i > v m a x w ( γ ψ x t 1 i ) + c 1 r 1 P p e s t x t 1 i + c 2 r 2 ( P g e s t x t 1 i ) v m a x < v t i < v m a x v m a x v t i < v m a x
where c 1 = 2 , c 2 = 2 are positive coefficients, r 1 and r 2 are random numbers in the interval (0, 1), v m a x is the maximum velocity limited to 10% of the dynamic range of the variables on each dimension, w is the inertial weight, γ is a positive constant, and ψ is the gradient of the objective function.

3.9. Differential Evolution and Particle Swarm Optimization: DEPSO

The authors in reference [67] propose a hybrid algorithm that combines the DE algorithm and PSO and is composed of two main steps. In the former, mutation and selection operators are used to produce a new population, while in the latter, PSO is adopted to explore local optimums, followed by crossover and selection operations. In this way, it is possible to increase and decrease the extent of searching regions for the population.
The advantages are that the algorithm overcomes degradation defects, enhances the effectiveness of the particles, and improves accuracy and robustness.

Hybrid Algorithm DEPSO

In reference [67], authors use DEPSO in a Particle Filter (PF) algorithm to reduce the particle number and the computational time complexity. The DE algorithm aims to create a new candidate, and to select the best ones among the current individuals and the candidate for the next generation. In this work, the DE steps are developed as follows:
  • Mutation. Three individuals from the population are randomly selected and the i -the mutant vector v k i is generated with this equation:
    v k i g + 1 = F x k r 2 g x k r 3 g + x k r 1 g
    where F is the mutation parameter, r 1 , r 2 , and r 3 identify the three selected individuals among N , k identifies the k -th step of the PF, and g identifies the generation of the DE.
  • Crossover. This step performs the crossover operator to generate a candidate for the trial vector and the current individual. The j -th dimension of the candidate is calculated by the following:
    u k i , j g + 1 = v k i , j g + 1 i f ( r a n d ( 0,1 ) C R o r ( j = j r a n d ) ) x k i , j g o t h e r w i s e
    where j r a n d is a random integer in the range of 1 , n , where n is the number of dimensions of the decision variables and C R is the crossover parameter.
  • Selection. In this step, the fitness function of u k i g + 1 and x k i are compared and the best current individual is selected.
    x k i g + 1 = u k i g + 1 i f ( f ( u k i g + 1 ) f ( x k i g ) ) x k i , j g o t h e r w i s e
  • End. The end condition is reached when a maximum number of iterations are executed or when the population’s average fitness meets requirements. Otherwise, the algorithm is repeated from step 1.
Considering the PSO algorithm, each particle of the swarm has a position X and a velocity V ; it is recalled that their evolution is classically based on the following formulas:
V i = w · V i + c 1 · r a n d · p b e s t i X i + c 2 · r a n d · g b e s t i X i         X i = X i + V i
where w is the inertia weight, p b e s t i is the best position of particle X i so far; g b e s t i is the best position of the swarm so far; rand is a random number between (0,1); and c 1 and c 2 are acceleration constants.
The DE algorithm and PSO have both advantages and disadvantages; for example, the differential algorithms have (i) a good global search capability due to the mutation step, which increases the diversity of the population, (ii) good local search capability, which is improved by the crossover step, and (iii) a memory function due to the selection process; however, the convergence speed is slow. On the other hand, PSO is characterized by a faster convergence rate, but it has the drawback that it might be stuck in a local optimum without converging to the global one because the diversity of the population can be lost very easily. Then, to exploit the characteristics of PSO, i.e., the high rate of convergence and the possibility to explore a local minimum quickly, Equation (113) is simplified as follows:
V i = w · V i + c 1 · r a n d · g b e s t i X i X i = X i + V i
After the aforementioned considerations, the hybrid algorithm DEPSO can be represented with the following steps:
  • Set t = 0 , and randomly initialize all individual positions X i and velocities V i .
  • t = t + 1 .
  • Calculate individual fitness function f ( X i ) .
  • Update the position of particles using the equation for mutation (110) and selection (112).
  • Select the individual in the best position of the swarm so far.
  • Update the position and velocity of particles according to (114).
  • Perform crossover and selection using (111) and (112), respectively.
  • If the counter is larger than a threshold, then exit; otherwise, return to step 2.
Authors claim that DEPSO avoids particle degradation, improves positioning accuracy, reduces the number of particles needed for positioning, and enhances the diversity of particles and robustness of the system.

3.10. Bat Algorithm

The bat algorithm (BA) idea is inspired by the echolocation behavior of bats to find food; in nature, they usually emit short pulses, and when they encounter food, they change the frequency to tune the echolocation time and increase the location accuracy of the prey [68]. The parameters of loudness and frequency determine the global search capability, and if a bat is near its scope, the impulse rate increases, whereas loudness decreases. In the standard bat algorithm, each individual i has a defined position x i ( t ) and velocity v i ( t ) in the search space, which are updated iteration by iteration with the following set of equations:
x i t + 1 = x i t + v i ( t + 1 )
v i t + 1 = v i t + ( x i t p t ) · f i
f i = f m i n + ( f m a x f m i n ) · β
where t is the t -th iteration, p t is the current global optimal solution, f i is the frequency of the i -th individual, f m i n = 0 , f m a x = 1 , and β is a random vector in the range [0, 1] with uniform distribution.
To implement the local search strategy, the following equation is considered:
x i t + 1 = x i t + ε A ¯ ( t )
where ε is a random number in the range [−1, 1] and A ¯ ( t ) is the average loudness of the population.
The global search is controlled by loudness A i t + 1 and pulse rate r i t + 1 , which are updated with the following equations:
A i t + 1 = α A i ( t )
r i t + 1 = r i t + ( 1 exp   ( γ t ) )
where 0 < α < 1 is the amplitude decrease rate and is a constant term, and γ > 0 is a constant. The initial values of loudness A i 0 and pulse rate r i 0 are set in the initialization.
The steps of the standard algorithm are as follows:
  • Initialize, for each bat, the position and velocity parameters and randomly generate the frequency with Equation (117).
  • Update the position and velocity of each bat with (115) and (116).
  • For each bat, generate a random number r a n d 1 in the range [0, 1];
    if r a n d 1 < r i ( t ) , then update the temp position with (118) and calculate the fitness value for the corresponding i -th bat.
  • For each bat, generate a random number r a n d 2 in the range [0, 1];
    if r a n d 2 < A i ( t ) and if F ( x i t ) < F ( p i t ) , where F is the fitness function, then update the loudness with (119) and the pulse rate with (120).
  • Sort each individual based on fitness values and save the best position.
  • If the final condition is met, then exit; otherwise, go to step 2.

Mobile Robot Localization Based on the Leader-Based Bat Algorithm

The authors in reference [69] propose a modified BA to solve the mobile robot global localization problem, which is often characterized by complex multimodal beliefs. In some problems, the standard BA might provide a premature convergence solution. Authors propose the leader-based bat algorithm (LBBA), introducing leaders who can influence the behavior of other individuals. This gives rise to more stable particle movements at multiple local optimum points, providing a convergence to the global optimum in more complex environments.
The LBBA is a nonparametric filter with N individuals, and each of them approximately corresponds to one region in the search space. The a priori states of individuals are represented by a set of random samples, and this enables the simultaneous representation of multiple optimal locations.
Unlike the standard BA, the LBBA uses both the best global g * and the best individual x * . The last one enables a higher diversity of solutions in the population, and, in addition, the performance and efficiency in the search for the global optimum are improved thanks to the adoption of leaders, who promote the creation of different colonies, decreasing the randomness of flight.
To determine the target position, the following object function is defined:
O x i = m i m m
where m i is the measurement performed by the simulated sensor of the i -th individual, and m is the measurement performed by the real robot sensor.
The main steps of a general LBBA are as follows:
  • Step 1. Initialization of x i , v i where i = 1 , , N . Define the number of leaders L = 1 , , L m a x .
    Define the number of leaders L = 1 , , L m a x .
    Initialization of frequency f i , impulse rate r i , and amplitude A i .
    Assess the fitness of each bat using the fitness function f i t ( x i ) .
    Determine leaders x L * with the fitness function.
    Define the best global bat and the relative best fitness: x * and f i t ( x * ) .
  • Step 2. Compute the following:
    -
    f i = f m i n + ( f m a x f m i n ) · β ;
    -
    v i t + 1 = v i t + ( x i t x L * ) · f i ;
    -
    x ^ i = x i t + v i t + 1 , where x ^ i is a temporary variable.
  • Step 3. If r a n d 1 > r i then x ^ i = x * + δ , where x * is the best global variable, and δ is a random displacement variable.
  • Step 4. Estimate the fitness of x ^ i using f i t x ^ i .
  • Step 5. Define new leaders x ^ L * .
  • Step 6. If f i t x ^ i f i t x i and r a n d 2 < A i , then compute the following:
    -
    x i = x ^ i ;
    -
    f i t x i = f i t ( x ^ i ) ;
    -
    r i = r ( 0 ) 1 e γ t ;
    -
    A i t + 1 = α A i ( t ) .
  • Step 7. If f i t x ^ i f i t x * , then compute x * = x ^ i , and f i t ( x * ) = f i t ( x ^ i ) .
  • Step 8. If the condition is met, return x * ; otherwise, repeat from step 2.
Authors claim that the LBBA has better performances when compared to other algorithms for robot localization such as PF and PSO, showing the ability to escape from local minima in highly complex environments. Moreover, it also succeeds in solving the robot’s kidnapping problem without premature convergences.

3.11. Artificial Bee Colony Optimization Algorithm

The Artificial Bee Colony (ABC) algorithm is a metaheuristic search algorithm inspired by the intelligent foraging behavior of honeybees in nature [70]. The algorithm is explained as follows [71]. The colony of artificial bees is composed of three groups: employed bees, onlookers, and scouts. The number of employed bees and onlooker bees is equal to the number of food sources. Every bee can only exploit one food source at a time. The scout bee was an employed bee that became a scout bee when it left the food source looking for a new one. A possible solution to the problem corresponds to a food source and the fitness of a solution corresponds to the nectar amount of a source. Each employed bee first explores the associated solution (or food source) according to their memories and explores possible solutions in corresponding neighborhoods. Then, they fly back to the hive and share the information about the food sources with the onlooker bees. The onlooker bees select the food sources, and a source is chosen using a probability criterion based on the value of fitness or amount of nectar: the food sources with more nectar have a higher probability of being selected. If a food source does not provide any better solution after several times that it is explored, the employed bee of that source becomes a scout bee and randomly selects a new source as it continues the work as an employed bee. The main steps of the algorithm are as follows [71]:
  • Initialization. A solution is considered a multi-dimensional vector x i and S N solutions are randomly generated at the beginning; the components of the vector are generated by the following equation:
    x i j = x j m i n + r a n d ( 0,1 ) × ( x j m a x x j m i n )
    where i is one of the S N solutions, j is one of the D dimensions of the vector, the x j m i n and x j m a x represent the lower and upper bound of the j -th dimension, respectively, and r a n d ( 0,1 ) provides a random number in the range [0, 1].
  • Employed bee phase. Each employed bee is associated with a solution x i and then new solutions, v i , are explored (or new food sources are generated) using the following equation:
    v i j = x i j + r a n d ( 1,1 ) × ( x i j x k j )
    where v i j is the j -th component of a new solution, r a n d ( 1,1 ) provides a random number in the range [−1,1], x k j is the j-th component of another food source x k that is selected randomly from the remaining population, and k i . The old solution, x i j , and the new one, v i j , are compared and the employed bee brings back the information of the better one, i.e., the solution with a smaller objective value or the best fitness. It is called the greedy selection strategy. The fitness values, f i t i , are calculated with the following equation:
    f i t i = 1 1 + f i f i 0 1 + a b s ( f i ) f i < 0
    where f i is the output of the objective function calculated using the x i solution.
  • Onlooker bee phase. The employed bees share the fitness information with the onlooker bees. Each onlooker bee chooses a solution for itself using a probability criterion: the higher the probability of a solution, the higher the possibility of being chosen. The selection probability of the i -th solution is called the roulette wheel selection and the probability p i is calculated as follows:
    p i = f i t i m = 1 S N f i t m .
    After the selection, each onlooker bee generates a new solution in the surroundings of the chosen solution by using (123). Each generated solution is compared with the corresponding previous food source via the greedy selection strategy, and if the new one is better, it is set as x i j = v i j . The difference between the updating process of the employed bees and the onlooker bees is that the employed bees update all solutions, while the onlooker bees only update the selected one.
  • Scout bee phase. A solution may be explored several times without providing a better solution in the surroundings. After a pre-set number of explorations in the surroundings of the same solution without better results, the solution is abandoned, and then the corresponding employed bee transforms into a scout bee to randomly generate a new solution by using (122). Then, it continues as an employed bee.

Multistage Localization in Wireless Sensor Networks Using the Artificial Bee Colony Algorithm

The authors in reference [72] propose a method based on the ABC algorithm to solve the sensor localization problem in a wireless sensor network, providing some simulation results. In the considered scenario, N nodes are deployed in a square plan, and each of them has a communication range of r units. A few nodes of them are considered as anchor nodes and they are in a known position; the other nodes are deployed in random locations. A node in an unknown position with at least three non-collinear anchors in its range can estimate the distances from anchors by parameters such as RSSI, Time of Arrival (ToA), and so on; if at least three anchor nodes are not available, the sensor cannot be localized. Each localizable node executes the ABC algorithm to estimate the coordinates of its location; once a sensor in an unknown position has been localized, it is used as a new anchor node to identify the other sensors. The aim is to determine the locations of all or as many nodes as possible, and in such a way, the distributed localization problem is modelled as a two-dimensional optimization problem.
The ABC algorithm computes the estimated location ( u ^ j x , u ^ j y ) , finding the minimum of the mean localization error E j , which is expressed as follows:
E j = 1 3 i = 1 3 u ^ j x a i x 2 + u ^ j y a i y 2 d ^ i
where j is the j -th node in an unknown position, i is the i -th anchor node with known coordinates ( a i x , a i y ) , d ^ i is the distance measured (that includes noise and errors), and ( u ^ j x , u ^ j y ) are the unknown coordinates.
The authors compared the performances of the ABC and PSO algorithms in finding the solution to the specific problem. The ABC algorithm can result in more accurate localization than PSO, but requires a higher computational time.

3.12. Simulated Annealing

The SA algorithm is a random-search technique that is based on the analogy between the process of metal crystalline formation from the liquid phase, and the process of searching for an optimal minimum in a problem. If the cooling process is too rapid, then the metal will solidify into a sub-optimal configuration. If the cooling process takes the proper time, the crystals within the material will solidify optimally into a state of minimum energy, which is the ground state associated with the optimal global solution to the problem. This is a heuristic method that does not need any particular assumptions and can be used for non-linear and non-convex problems; in particular, it results in a suitable technique for solving combinatorial optimization problems.
The main steps of the SA algorithm are as follows [73]:
  • Generate a random initial solution vector x = x o for the system and then evaluate it using the objective function f ( x ) .
  • Initialize the temperature T = T 0 .
  • Perturbate the previous solution to obtain a neighboring x i + Δ x . Evaluate the new solution.
  • A new solution ( x i + Δ x ) is accepted as a new solution x i + 1 = x i + Δ x if f ( x i + Δ x ) is better than f ( x i ) .
  • Otherwise, accept the new solution x i + Δ x with a probability e Δ f T i ,
    where Δ f = f ( x i + Δ x ) f ( x i ) .
  • Reduce the system temperature according to the cooling schedule.
  • Repeat from step 3 until a stopping criterion is met (number of cycles or T i = 0 ) .
In the SA problem, temperature is a parameter that controls the probability in step 5. During the iterations, the lower the temperature, the lower the probability of accepting a bad solution. This mechanism helps the procedure to avoid being stuck in a global minimum at the beginning of the iterations, because it allows the algorithm to jump out of the minima looking for better solutions. To obtain a good result, it is important to properly set the initial value of temperature T . If the value is too high, then it takes more cycles to converge, and vice versa; if the value is too small, the algorithm may provide a local optimum.
In step 6, the choice of an appropriate cooling scheme is important for the algorithm’s success. In the literature, there are several types of schemes, and they can be classified as either monotonic or adaptive. The former consists of a static decrease in the temperature at every iteration independently of the goodness of the solutions and independently of the current explored neighborhood structure, whereas the last one considers the quality of transitions, adopting mechanisms for the temperature decrease that can adopt also reannealing techniques [73].

Simulated Annealing-Based Localization in Wireless Sensor Network

The authors in reference [74] propose a SA localization (SAL) method for wireless sensor networks. It aims to estimate the position of sensors by determining the position of a set of anchors and the measured distance between them and the sensors. However, several factors can influence distance measurements, such as synchronization, multipath, fading, Non-Line of Sight (NLoS) conditions, and other sources of interference. To estimate the real distance of the sensor with a minimum error, the authors propose to solve the optimization problem with the SAL method. The objective function is defined as follows:
C F = i = 1 N j N i d ^ i j d i j 2
where N is number of non-anchor nodes, N i is the set of neighbors of the i -th node, d ^ i j and d i j are the estimated distance and the measured distance of node i with its neighbor j , respectively.
The adopted cooling scheme for temperature is T i + 1 = α * T i , where α < 1 is the cooling rate that is determined empirically. The solutions are perturbated as Δ d i + 1 = β * Δ d i , where Δ d is chosen to span the whole set of allowable moves and β < 1 is used to bias the generation of random moves at a lower T .

3.13. Ant Colony Optimization

Ant Colony Optimization (ACO) is a probabilistic method, and it was initially proposed to solve the Travelling Salesman Problem (TSP); in general, ACO is suitable for solving problems in which the searched solution is the best path through a graph. It is inspired by the behavior of an ant seeking a path between its colony and food. The ants live in a colony, roam around it looking for food, and communicate with each other by using pheromones which can be released on the ground when insects are moving. Once an ant finds some food, the insect brings it as much as possible back to the colony and, during the travel, it releases pheromones in agreement with the quantity and quality of the found food. Other ants can smell it, and when they come out of the colony looking for food, they choose the path with higher pheromones. The higher the pheromone along a route, the higher the probability that an ant chooses that path. In such a way, more ants follow the path, more pheromones are deposited, and the path has a higher probability of being chosen. After some iterations, the ants will choose the shortest path, which is expected to be the optimum of the problem.
Mathematically, the algorithm is explained as follows [75]. ACO can be understood considering a simple graph, G = V , E , with two nodes, where V represents the nodes of the graph (the nest v s and the food source v d ), E represents two links, e 1 and e 2 , between nodes with length l 1 and l 2 , respectively, such that l 1 < l 2 . The artificial pheromone trails are modelled for each path by τ 1 and τ 2 and they indicate the strength of the pheromone trail on the corresponding path. Every ant chooses with a probability given as follows:
p i = τ i τ 1 + τ 2
where i identifies the i -th link; about the lengths, it is τ 1 > τ 2 , which means that link 1 has a higher probability. As for returning, every ant uses the same path chosen in the outward journey, and it updates the artificial pheromone value associated with the used link with the following equation:
τ i = τ i Q l i
where Q > 0 is a parameter of the model. It can be inferred that the shorter the path, the higher the amount of added pheromones. At each step (or iteration), all the ants are initially placed in node v s and each of them moves from the nest to the food source. The evaporation of the pheromone is simulated by the following:
τ i = ( 1 ρ )
where ρ is in the range 0,1 and regulates the pheromone evaporation. Finally, all ants return to the nest and reinforce their chosen path, as reported above. Iteration by iteration, the shortest path will be the one with the highest pheromones, and then, with the highest probability.
For real combinatory optimization problems with more than two nodes, the solution is constructed as follows [76]. Each of the m ants is randomly placed on a chosen node and then at each node, a state transition rule is iteratively applied. The unvisited nodes, the pheromone trail strength, and the length between two nodes bias the ant’s choice. The partial tour is stored in a memory called the tabu list which needs to determine at each construction step the nodes to be visited, to guarantee that a feasible solution is built, and to retrace the trail of the related ant once it has been completed. When all ants end up constructing a path, the pheromones are updated using evaporation and new deposition; short paths are updated with a higher quantity of pheromones. The algorithm is summarized as follows [76]:
  • Set parameters and initialize pheromone trails.
  • Construct solutions.
  • Apply local search.
  • Update trails.
  • If termination conditions are not met, repeat from step 2.
There are different versions of ACO and historically, the first presented version was called the Ant System (AS); some variants of the ACO algorithms (here not investigated) are the Elitist AS (EAS), Rank-based AS (RAS), MAX–MIN Ant System (MMAS), Ant Colony System (ACS), and Hyper-Cube Framework (HCF).

ANN Topology Optimized by ACO

The author in reference [77] deals with the problem of finding a continuous path for a robot from an initial position to a goal position where collisions against static objects must be avoided. The proposed system exploits an ANN that provides the optimum path, and it is trained by adopting the swarm intelligence-based reinforcement learning (SWIRL) method based on PSO and ACO. Because the training process of Multi-Layer Perceptron (MLP) for pattern classification problems can be considered as graph problems where the neurons are vertices and the connections are directed edges (links), PSO is exploited to provide the best ANN connection weights, whilst the ACO is used to optimize the topology of the ANN. In such a way, the SWIRL provides both an appropriate architecture for the problem and the trained connection weights of the network.
At each step of the iteration process, the PSO is applied to adjust ANN connection weights within a given topology structure. Focusing only on the ACO algorithm, it is used to allocate training iterations among a set of candidate network topologies. The desirability d ( i ) for the i -th ANN is given as follows:
d ( i ) = 1 h + 1
where h is the number of hidden nodes; each i -th ANN represents a topology of a set of ANNs with different numbers of hidden nodes. The pheromone concentration τ i is initialized to 0.1, and it is expressed as follows:
τ i , t + 1 = ρ τ i , t + n a ( i ) g ( i ) g s u m
where ρ is the rate of evaporation, n a is the number of ants returning from the i -th neural network, g ( i ) is the global best for i , g s u m is the sum of all the current global bests, and t is the index iteration of the ACO. Each ant represents one training process for the PSO. During each ACO step, the ants go out into the topology space of ANNs, and they choose an ANN using a probability criteria, given as follows:
p ( i ) = d i b τ i , t a d i b τ i , t a
where a and b are constant factors which control the relative influence of pheromones and desirability, respectively [77].

3.14. Whale Optimization Algorithm

The Whale Optimization Algorithm (WOA) is a metaheuristic method inspired by the hunting behavior of humpback whales, and it has been proposed by authors in reference [78]. The algorithm mimics three behaviors: (i) encircling prey or siege predation, (ii) spiral bubble-net feeding maneuver, and (iii) searching for prey. In the first one, a whale is in a location looking for prey to encircle; at the beginning, the locations (or solutions) are random, and the current best solution, defined as the best search agent, is found by exploiting the fitness function; the other search agents or candidate solutions move iteration by iteration towards the current best one. Mathematically, the behavior of each whale can be modelled by the following:
D = C · X * t X ( t )
X t + 1 = X * t A · D
where t is the t -th iteration, A and C are coefficient vectors, X * is the position vector of the best solution obtained so far, X is a candidate position vector, and the dot ( · ) represents the element-by-element multiplication. Furthermore, it is possible to express A = 2 a · r a and C = 2 · r , where a is linearly decreased from 2 to 0 iteration by iteration, and r is a random vector in the range [0, 1] that allows it to reach any position of the design space. The used vectors in this explanation have the same dimension as the input vector of the problem; however, the algorithm can be applied to a problem with a dimension greater than three even if the algorithm is inspired by the analogy of a moving whale in a three-dimensional space.
In the second behavior (spiral bubble-net feeding manoeuvre), two approaches are adopted: the shrinking encircling mechanism and the spiral updating position. The former needs to push the X vectors towards X * and it consists in decreasing the value of a iteration by iteration; in such a way, considering that r is random, the result is that A is composed by random values in the range [ a , a ]. As for the second approach, to mimic an encircling movement, a spiral equation is created between the position of a whale and the best position; it is expressed as follows:
X t + 1 = D i · e b l · cos   2 π l + X * ( t )
where D i = X * t X ( t ) indicates the distance of the i -th candidate solution (or whale) to the best solution obtained so far, b is a constant for defining the shape of the exponential spiral, and l is a random number in the range [−1, 1]. Considering the analogy with whales, they are swimming around their candidate location prey within a shrinking circle and along a spiral-shaped path simultaneously. To update the position of whales and to mimic the simultaneous behavior of the shrinking encircling mechanism and the spiral model, in the algorithm they are chosen randomly with a probability of 50% in each iteration. The two approaches are expressed as follows:
X t + 1 = X * t A · D D · e b l · cos   2 π l + X * ( t ) i f   p < 0.5 i f   p 0.5
where p is a random number in the range [0, 1].
In the search for prey behavior, whales look for a prey location, randomly giving the algorithm the global searching capability. It is expressed as follows:
D = C · X r a n d X ( t )
X t + 1 = X r a n d A · D
where X r a n d is a random position vector chosen from the current whale population.
The WOA pseudo code is composed of four main parts, and they are given as follows [78]:
  • Initialize the whale population X i with i = 1 , 2 , , n .
    Calculate the fitness of each search agent.
    Define the best search agent X * .
  • While t < m a x   n u m b e r   o f   i t e r a t i o n s :
    • For each search agent:
      Update a , A , C , l , and p .
      If p < 0.5 : {
      -
      If A < 0.5 , update the position of the current search agent by (134).
      -
      Elseif A 0.5 , select a random search agent X r a n d and update the position of the current search agent by (139).
      }
      Elseif p 0.5 , update the position of the current search agent by (136).
    • End of for loop.
    • Check if any search agent goes beyond the search space and amend it.
    • Calculate the fitness of each search agent.
    • Update X * whether a better solution has been found.
    • t = t + 1 .
  • End.
  • Return X * .
In this basic version of the proposed WOA, other evolutionary operations are not included, but they might have been included by allowing hybridization with other evolutionary search schemes.

Three Magnetic Targets Using Individual Memory-Based WOA and LM Algorithm

The authors in reference [79] deal with multi-permanent magnet localization, exploiting an array of magnetometers. The aim is to estimate the location of the permanent magnets inside the human body by measuring the magnetic field. To obtain the best estimation of the measurements, the authors propose an improved WOA combined with the LM algorithm.
Each magnetometer in the array measures the magnetic flux that is given by the linear superposition of the flux densities generated by each permanent magnet. The objective function E of the total error is obtained by considering the sum of the three errors between the measured values by the magnetometers and the theoretical values; it is given as follows:
E = l = 1 N B l x B l x 2 + B l y B l y 2 + B l z B l z 2
where each B l represents one of the triaxial components of the magnetic flux density measured by the l -th magnetometer, B l is the theoretical value of one of the triaxial components of the magnetic flux, and N is the total number of magnetometers. A generic component B l is given by the superimposition of the magnetic fields generated by all permanent magnets, and it includes the distance variables between each permanent magnet and each magnetometer.
The name of the proposed algorithm is the Individual Memory-based Whale Optimization and Levenberg–Marquardt (IMWO-LM) algorithm; the IMWO algorithm is used to provide an approximate solution of each magnet pose, which is given as a guess input to the LM algorithm, which is suitable for finding the local minima.
In the proposed IMWO, S is defined as the number of whales and each of them searches for a solution of dimension d ; the vector position for the i -th whale is defined as X i = [ x 1 i , x 2 i , , x d i ] . The three behaviors presented in the WOA are here modified to give whale the ability of self-memory and to always remember the previous best position. The IMWO behaviors are defined as follows:
  • Search for prey. The current search agent position is updated considering both the position of the randomly selected whale and the historical optimal position. It is given as follows:
    D r a n d i = C · X r a n d i X t , i D p b e s t i = C X p b e s t t 1 , i X t , i X t , i = X r a n d i A D r a n d i + A D p b e s t i    
    where t identifies the t -th iteration, i identifies the search agent, X t , i is the current position, X p b e s t t 1 , i is the historical optimal position, D p b e s t i is the random distance between the current position and its historical optimal position, X r a n d i is a randomly selected search agent, D r a n d i is the random distance between a randomly selected search agent and current search agent, C and A are weight vectors defined as in the WOA, and X t , i is the newly updated location of the i -th search agent.
  • Encircling prey or siege predation. The best candidate search agent X * of the current population is considered the target prey at each iteration; search agents have memory and can remember the previous best position. It is given as follows:
    D l e a d e r i = C X * X t , i D p b e s t i = C X p b e s t t 1 , i X t , i X t , i = X * A D l e a d e r i + A D p b e s t i
    where D l e a d e r i is the random distance between the optimal candidate search agent and the current search agent, and X * is the optimal candidate search agent in the whole population so far. In such a way, the current search agent is updated to different positions around X * and X p b e s t t 1 , i .
  • Spiral bubble-net feeding maneuver or bubble predation. The helix-shaped movement of the whale is obtained using the spiral equation, considering the position of the current search agent, its historical optimal position, and the position of the optimal candidate search agent. The current position is given as follows:
    X t i = r 1 D * + r 2 D p b e s t i · e b l · cos   2 π l + r 1 X * + r 2 X p b e s t t 1 , i
    where D * = X * X t , i is the distance between the current search agent and the optimal candidate search agent, b is a constant for defining the shape of the exponential spiral, l is linearly decreased from 1 to −1 throughout iterations, and r 1 and r 2 are random numbers in the range [0, 1].
The upper and lower boundary of the pose parameters are set, and at each iteration, the algorithm must check if the search agents violate them.

3.15. Summary of the Investigated Methods

In this section, Table 1 reports a summary of the investigated methods, where the main characteristics, the significative advantages and disadvantages, and the investigated applications are listed to have a comprehensive view of the conducted research.

4. Conclusions

This paper provides an extensive study related to the optimization methods used in the field of localization. In the first part of the article, a light introduction to the optimization problem was reported, providing a general mathematical framework for the discipline. Besides the SO problem, some concepts about the MO problem have been introduced as well, and they aim to provide a clear and complete view of the basis of the subject.
In the second part of the article, extensive research about some interesting algorithms used in localization has been reported, providing for each of them the concept of the algorithm, the related mathematics model, and an example of application. Both types of algorithms for local and global optimization have been described, and it can be noted that metaheuristic algorithms, such as GA, DE, or PSO, are the most used in the field of localization for optimization problems due to their capabilities to manage non-convex and non-linear problems and provide satisfactory results. Moreover, the GA (as well as its modified versions, such as the QGA) is the most versatile optimization tool that allows for different types of problems such as continuous, discrete, or mixed-integer types. However, the standard GA has a slow speed of convergence. In contrast, PSO has a higher speed, and by observing Table 1, it is the most used among the investigated methods. Furthermore, PSO was shown to be the most prone method that allows for hybridization with other types, such as DEPSO or SWIRL; the modified methods improved the performance in solving specific problems.
In general, by comparing metaheuristic algorithms to gradient-based algorithms (that are especially used for local optimum), it is noted that stochastic algorithms do not need any Jacobian (or Hessian) matrix inversion, demonstrating that they are very flexible and suitable optimization tools for many optimization problems. On the other hand, they need to tune the parameters for each specific solved problem.
In the end, it is highlighted that none of the investigated methods is guaranteed to provide a real global optimum.

Author Contributions

Conceptualization, M.S., P.S. and A.O.; funding acquisition, P.S. and A.O.; project administration, P.S. and A.O.; resources, M.S., P.S., J.S. and A.O.; supervision, P.S., J.S. and A.O.; writing—original draft, M.S.; writing—review and editing, P.S., J.S. and A.O. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the National Research, Development, and Innovation Fund of Hungary through project no. 142790 under the FK_22 funding scheme.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Talbi, E.G. Metaheuristics: From Design to Implementation; John Wiley & Sons: Hoboken, NJ, USA, 2009. [Google Scholar]
  2. Gogna, A.; Tayal, A. Metaheuristics: Review and Application. J. Exp. Theor. Artif. Intell. 2013, 25, 503–526. [Google Scholar] [CrossRef]
  3. Gilli, M.; Schumann, E. Heuristic Optimisation in Financial Modelling. Ann. Oper. Res. 2012, 193, 129–158. [Google Scholar] [CrossRef]
  4. Agrawal, N.; Rangaiah, G.P.; Ray, A.K.; Gupta, S.K. Multi-Objective Optimization of the Operation of an Industrial Low-Density Polyethylene Tubular Reactor Using Genetic Algorithm and Its Jumping Gene Adaptations. Ind. Eng. Chem. Res. 2006, 45, 3182–3199. [Google Scholar] [CrossRef]
  5. Kulkarni, N.K.; Patekar, S.; Bhoskar, T.; Kulkarni, O.; Kakandikar, G.M.; Nandedkar, V.M. Particle Swarm Optimization Applications to Mechanical Engineering—A Review. In Materials Today: Proceedings; Elsevier: Amsterdam, The Netherlands, 2015; Volume 2. [Google Scholar]
  6. Jordehi, A.R. Particle Swarm Optimisation (PSO) for Allocation of FACTS Devices in Electric Transmission Systems: A Review. Renew. Sustain. Energy Rev. 2015, 52, 1260–1267. [Google Scholar] [CrossRef]
  7. Lin, M.H. An Optimal Workload-Based Data Allocation Approach for Multidisk Databases. Data Knowl. Eng. 2009, 68, 499–508. [Google Scholar] [CrossRef]
  8. Odry, A.; Kecskes, I.; Csik, D.; Hashim, H.A.; Sarcevic, P. Adaptive Gradient-Descent Extended Kalman Filter for Pose Estimation of Mobile Robots with Sparse Reference Signals. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Kyoto, Japan, 23–27 October 2022; Volume 2022. [Google Scholar]
  9. Grottke, J.; Blankenbach, J. Evolutionary Optimization Strategy for Indoor Position Estimation Using Smartphones. Electronics 2021, 10, 618. [Google Scholar] [CrossRef]
  10. Yousefi, M.; Nejat Pishkenari, H.; Alasty, A. A Fast and Robust Magnetic Localization Technique Based on Elimination of the Orientation Variables from the Optimization. IEEE Sens. J. 2021, 21, 21885–21892. [Google Scholar] [CrossRef]
  11. Yu, B.; Zhu, H.; Xue, D.; Xu, L.; Zhang, S.; Li, B. A Dead Reckoning Calibration Scheme Based on Optimization with an Adaptive Quantum-Inspired Evolutionary Algorithm for Vehicle Self-Localization. Entropy 2022, 24, 1128. [Google Scholar] [CrossRef]
  12. Lei, J.; Fang, H.; Zhu, Y.; Chen, Z.; Wang, X.; Xue, B.; Yang, M.; Wang, N. GPR Detection Localization of Underground Structures Based on Deep Learning and Reverse Time Migration. NDT E Int. 2024, 143, 103043. [Google Scholar] [CrossRef]
  13. Su, Y.; Wang, J.; Li, D.; Wang, X.; Hu, L.; Yao, Y.; Kang, Y. End-to-End Deep Learning Model for Underground Utilities Localization Using GPR. Autom Constr 2023, 149, 104776. [Google Scholar] [CrossRef]
  14. Boyd, S.P.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  15. Venter, G. Review of Optimization Techniques. In Encyclopedia of Aerospace Engineering; John Wiley & Sons, Ltd.: Chichester, UK, 2010. [Google Scholar]
  16. Esmaeilzadeh Azar, F.; Perrier, M.; Srinivasan, B. A Global Optimization Method Based on Multi-Unit Extremum-Seeking for Scalar Nonlinear Systems. Comput. Chem. Eng. 2011, 35, 456–463. [Google Scholar] [CrossRef]
  17. Tawfik, A.S.; Badr, A.A.; Abdel-Rahman, I.F. Comparative Optimizer Rank and Score: A Modern Approach for Performance Analysis of Optimization Techniques. Expert Syst. Appl. 2016, 45, 118–130. [Google Scholar] [CrossRef]
  18. Lin, M.H.; Tsai, J.F.; Yu, C.S. A Review of Deterministic Optimization Methods in Engineering and Management. Math. Probl. Eng. 2012, 2012, 756023. [Google Scholar] [CrossRef]
  19. Engelbrecht, A.P. Computational Intelligence: An Introduction, 2nd ed.; John Wiley & Sons Ltd.: Hoboken, NJ, USA, 2007. [Google Scholar]
  20. Arora, J.S. Introduction to Optimum Design; Elsevier: Amsterdam, The Netherlands, 2004. [Google Scholar]
  21. Cox, S.E.; Haftka, R.T.; Baker, C.A.; Grossman, B.; Mason, W.H.; Watson, L.T. A Comparison of Global Optimization Methods for the Design of a High-Speed Civil Transport. J. Glob. Optim. 2001, 21, 415–432. [Google Scholar] [CrossRef]
  22. Kvasov, D.E.; Mukhametzhanov, M.S. Metaheuristic vs. Deterministic Global Optimization Algorithms: The Univariate Case. Appl. Math. Comput. 2018, 318, 245–259. [Google Scholar] [CrossRef]
  23. Abdel-Basset, M.; Abdel-Fatah, L.; Sangaiah, A.K. Metaheuristic Algorithms: A Comprehensive Review. In Computational Intelligence for Multimedia Big Data on the Cloud with Engineering Applications; Academic Press: Cambridge, MA, USA, 2018. [Google Scholar]
  24. Agrawal, P.; Abutarboush, H.F.; Ganesh, T.; Mohamed, A.W. Metaheuristic Algorithms on Feature Selection: A Survey of One Decade of Research (2009–2019). IEEE Access 2021, 9, 3056407. [Google Scholar] [CrossRef]
  25. Francisco, M.; Revollar, S.; Vega, P.; Lamanna, R. A Comparative Study of Deterministic and Stochastic Optimization Methods for Integrated Design of Processes. In Proceedings of the IFAC Proceedings Volumes (IFAC-PapersOnline), 16th Triennial World Congress, Prague, Czech Republic, 3–8 July 2005; Volume 38. [Google Scholar]
  26. Jones, D.R.; Perttunen, C.D.; Stuckman, B.E. Lipschitzian Optimization without the Lipschitz Constant. J. Optim. Theory Appl. 1993, 79, 157–181. [Google Scholar] [CrossRef]
  27. Deb, K. Multi-Objective Optimisation Using Evolutionary Algorithms: An Introduction. In Multi-Objective Evolutionary Optimisation for Product Design and Manufacturing; Springer: London, UK, 2011. [Google Scholar]
  28. De Weck, O. Multiobjective Optimization: History and Promise. In Proceedings of the The Third China-Japan-Korea Joint Symposium on Optimization of Structural and Mechanical Systems, Kanazawa, Japan, 30 October–2 November 2004. Invited Keynote Paper, GL2-2. [Google Scholar]
  29. Gunantara, N. A Review of Multi-Objective Optimization: Methods and Its Applications. Cogent Eng. 2018, 5, 1502242. [Google Scholar] [CrossRef]
  30. Marler, R.T.; Arora, J.S. Survey of Multi-Objective Optimization Methods for Engineering. Struct. Multidiscip. Optim. 2004, 26, 369–395. [Google Scholar] [CrossRef]
  31. Gunantara, N.; Sastra, N.P.; Hendrantoro, G. Cooperative Diversity Pathsselection Protocol Withmulti-Objective Criterion in Wireless Ad-Hoc Networks. Int. J. Appl. Eng. Res. 2014, 9, 22395–22407. [Google Scholar]
  32. Athan, T.W.; Papalambros, P.Y. A Note on Weighted Criteria Methods for Compromise Solutions in Multi-Objective Optimization. Eng. Optim. 1996, 27, 155–176. [Google Scholar] [CrossRef]
  33. Gerasimov, E.N.; Repko, V.N. Multicriterial Optimization. Sov. Appl. Mech. 1978, 14, 1179–1184. [Google Scholar] [CrossRef]
  34. Meza Juan, C. Newton’s Method. Wiley Interdiscip. Rev. Comput. Stat. 2011, 3, 75–78. [Google Scholar] [CrossRef]
  35. Lee, J.D.; Simchowitz, M.; Jordan, M.I.; Recht, B. Gradient Descent Only Converges to Minimizers. J. Mach. Learn. Res. 2016, 49, 1246–1257. [Google Scholar] [CrossRef]
  36. Smith, A.E.; Coit, D.W.; Baeck, T.; Fogel, D.; Michalewicz, Z. Penalty Functions. Handb. Evol. Comput. 1997, 97, C5. [Google Scholar]
  37. Wang, X.; Yan, L.; Zhang, Q. Research on the Application of Gradient Descent Algorithm in Machine Learning. In Proceedings of the 2021 International Conference on Computer Network, Electronic and Automation, ICCNEA 2021, Xi’an, China, 24–26 September 2021. [Google Scholar]
  38. Chang, X.W.; Yang, X.; Zhou, T. MLAMBDA: A Modified LAMBDA Method for Integer Least-Squares Estimation. J. Geod. 2005, 79, 552–565. [Google Scholar] [CrossRef]
  39. Cheng, Q.; Chen, W.; Sun, R.; Wang, J.; Weng, D. RANSAC-Based Instantaneous Real-Time Kinematic Positioning with GNSS Triple-Frequency Signals in Urban Areas. J. Geod. 2024, 98, 24. [Google Scholar] [CrossRef]
  40. Transtrum, M.K.; Sethna, J.P. Improvements to the Levenberg-Marquardt Algorithm for Nonlinear Least-Squares Minimization. arXiv 2012, arXiv:1201.5885. [Google Scholar]
  41. Levenberg, K. A Method for the Solution of Certain Non-Linear Problems in Least Squares. Q. Appl. Math. 1944, 2, 164–168. [Google Scholar] [CrossRef]
  42. Marquardt, D.W. An Algorithm for Least-Squares Estimation of Nonlinear Parameters. J. Soc. Ind. Appl. Math. 1963, 11, 431–441. [Google Scholar] [CrossRef]
  43. Hoffman Laurence, D.; Gerald, L. Bradley Calculus for Business, Economics, and the Social and Life Sciences, 10th ed.; McGraw-Hill Inc.: New York, NY, USA, 2010. [Google Scholar]
  44. Kalman, D. Leveling with Lagrange: An Alternate View of Constrained Optimization. Math. Mag. 2009, 82, 186–196. [Google Scholar] [CrossRef]
  45. Parkinson, A.R.; Balling, R.J.; Hedengren, J.D. Optimization Methods for Engineering Design Applications and Theory; Brigham Young University: Provo, UT, USA, 1972; Volume 94. [Google Scholar]
  46. Ghojogh, B.; Ghodsi, A.; Karray, F.; Crowley, M. KKT Conditions, First-Order and Second-Order Optimization, and Distributed Optimization: Tutorial and Survey. arXiv 2021, arXiv:2110.01858. [Google Scholar]
  47. Xue, K.; Li, J.; Xiao, N.; Liu, J.; Ji, X.; Qian, H. Improving the Robot Localization Accuracy Using Range-Only Data: An Optimization Approach. In Proceedings of the 2021 6th IEEE International Conference on Advanced Robotics and Mechatronics, ICARM 2021, Chongqing, China, 3–5 July 2021. [Google Scholar]
  48. Katoch, S.; Chauhan, S.S.; Kumar, V. A Review on Genetic Algorithm: Past, Present, and Future. Multimed. Tools Appl. 2021, 80, 8091–8126. [Google Scholar] [CrossRef] [PubMed]
  49. Kaya, M. The Effects of a New Selection Operator on the Performance of a Genetic Algorithm. Appl. Math. Comput. 2011, 217, 7669–7678. [Google Scholar] [CrossRef]
  50. Liang, Y.; Leung, K.S. Genetic Algorithm with Adaptive Elitist-Population Strategies for Multimodal Function Optimization. Appl. Soft Comput. J. 2011, 11, 2017–2034. [Google Scholar] [CrossRef]
  51. Umbarkar, A.J.; Sheth, P.D. Crossover Operators in Genetic Algorithms: A Review. ICTACT J. Soft Comput. 2015, 6, 1083–1092. [Google Scholar] [CrossRef]
  52. Lambora, A.; Gupta, K.; Chopra, K. Genetic Algorithm—A Literature Review. In Proceedings of the International Conference on Machine Learning, Big Data, Cloud and Parallel Computing: Trends, Prespectives and Prospects, COMITCon 2019, Faridabad, India, 14–16 February 2019. [Google Scholar]
  53. Lim, S.M.; Sultan, A.B.M.; Sulaiman, M.N.; Mustapha, A.; Leong, K.Y. Crossover and Mutation Operators of Genetic Algorithms. Int. J. Mach. Learn. Comput. 2017, 7, 9–12. [Google Scholar] [CrossRef]
  54. Frenzel, J.F. Genetic Algorithms. IEEE Potentials 1993, 12, 21–24. [Google Scholar] [CrossRef]
  55. Han, K.H.; Kim, J.H. Genetic Quantum Algorithm and Its Application to Combinatorial Optimization Problem. In Proceedings of the IEEE Conference on Evolutionary Computation, ICEC, La Jolla, CA, USA, 16–19 July 2000; Volume 2. [Google Scholar]
  56. Deng, W.; Shang, S.; Cai, X.; Zhao, H.; Song, Y.; Xu, J. An Improved Differential Evolution Algorithm and Its Application in Optimization Problem. Soft Comput. 2021, 25, 5277–5298. [Google Scholar] [CrossRef]
  57. Bhowmik, P.; Das, S.; Konar, A.; Das, S.; Nagar, A.K. A New Differential Evolution with Improved Mutation Strategy. In Proceedings of the 2010 IEEE World Congress on Computational Intelligence, WCCI 2010—2010 IEEE Congress on Evolutionary Computation, CEC 2010, Barcelona, Spain, 18–23 July 2010. [Google Scholar]
  58. Lin, C.; Qing, A.; Feng, Q. A Comparative Study of Crossover in Differential Evolution. J. Heuristics 2011, 17, 675–703. [Google Scholar] [CrossRef]
  59. Wu, Y.; Liu, X.; Guan, W.; Chen, B.; Chen, X.; Xie, C. High-Speed 3D Indoor Localization System Based on Visible Light Communication Using Differential Evolution Algorithm. Opt. Commun. 2018, 424, 177–189. [Google Scholar] [CrossRef]
  60. Sun, M.; Wang, Y.; Joseph, W.; Plets, D. Indoor Localization Using Mind Evolutionary Algorithm-Based Geomagnetic Positioning and Smartphone IMU Sensors. IEEE Sens. J. 2022, 22, 3155817. [Google Scholar] [CrossRef]
  61. Jie, J.; Zeng, J.; Ren, Y. Improved Mind Evolutionary Computation for Optimizations. In Proceedings of the World Congress on Intelligent Control and Automation (WCICA), Hangzhou, China, 15–19 June 2004; Volume 3. [Google Scholar]
  62. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; IEEE: Piscataway, NJ, USA, 1995; pp. 1942–1948. [Google Scholar]
  63. Li, N.; Chen, J.; Yuan, Y.; Tian, X.; Han, Y.; Xia, M. A Wi-Fi Indoor Localization Strategy Using Particle Swarm Optimization Based Artificial Neural Networks. Int. J. Distrib. Sens. Netw. 2016, 2016, 4583147. [Google Scholar] [CrossRef]
  64. Zhang, Y.; Hu, H.; Fu, W.; Jiang, H. Particle Swarm Optimization-Based Minimum Residual Algorithm for Mobile Robot Localization in Indoor Environment. Int. J. Adv. Robot. Syst. 2017, 14, 1729881417729277. [Google Scholar] [CrossRef]
  65. Havangi, R. Mobile Robot Localization Based on PSO Estimator. Asian J. Control 2019, 21, 2167–2178. [Google Scholar] [CrossRef]
  66. Maeda, Y.; Matsushita, N. Simultaneous Perturbation Particle Swarm Optimization Using FPGA. In Proceedings of the IEEE International Conference on Neural Networks—Conference Proceedings, Orlando, FL, USA, 12–17 August 2007. [Google Scholar]
  67. Huo, J.; Ma, L.; Yu, Y.; Wang, J. Hybrid Algorithm Based Mobile Robot Localization Using de and PSO. In Proceedings of the Chinese Control Conference, CCC, Xi’an, China, 26–28 July 2013. [Google Scholar]
  68. Wang, Y.; Wang, P.; Zhang, J.; Cui, Z.; Cai, X.; Zhang, W.; Chen, J. A Novel Bat Algorithm with Multiple Strategies Coupling for Numerical Optimization. Mathematics 2019, 7, 135. [Google Scholar] [CrossRef]
  69. Neto, W.A.; Pinto, M.F.; Marcato, A.L.M.; da Silva, I.C.; Fernandes, D.d.A. Mobile Robot Localization Based on the Novel Leader-Based Bat Algorithm. J. Control Autom. Electr. Syst. 2019, 30, 337–346. [Google Scholar] [CrossRef]
  70. Karaboga, D. An Idea Based on Honey Bee Swarm for Numerical Optimization; Technical Report TR06; Erciyes University: Kayseri, Türkiye, 2005. [Google Scholar]
  71. Zhao, Y.; Yan, Q.; Yang, Z.; Yu, X.; Jia, B. A Novel Artificial Bee Colony Algorithm for Structural Damage Detection. Adv. Civ. Eng. 2020, 2020, 3743089. [Google Scholar] [CrossRef]
  72. Kulkarni, V.R.; Desai, V.; Kulkarni, R.V. Multistage Localization in Wireless Sensor Networks Using Artificial Bee Colony Algorithm. In Proceedings of the 2016 IEEE Symposium Series on Computational Intelligence, SSCI 2016, Athens, Greece, 6–9 December 2016. [Google Scholar]
  73. Amine, K. Multiobjective Simulated Annealing: Principles and Algorithm Variants. Adv. Oper. Res. 2019, 2019, 8134674. [Google Scholar] [CrossRef]
  74. Kannan, A.A.; Mao, G.; Vucetic, B. Simulated Annealing Based Localization in Wireless Sensor Network. In Proceedings of the Proceedings—Conference on Local Computer Networks, LCN, Sydney, NSW, Australia, 17 November 2005; Volume 2005. [Google Scholar]
  75. Blum, C. Ant Colony Optimization: Introduction and Recent Trends. Phys. Life Rev. 2005, 2, 353–373. [Google Scholar] [CrossRef]
  76. Stützle, T.; Dorigo, M. ACO Algorithms for the Travelling Salesman Problem. Evol. Algorithms Eng. Comput. Sci. 1999, 4, 163–183. [Google Scholar]
  77. Mali, S. Mobile Robot Localization Using Multi-Objective Optimization. Int. J. Latest Technol. Eng. Manag. Appl. Sci. 2015, 4, 21–25. [Google Scholar]
  78. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  79. Lv, B.; Qin, Y.; Dai, H.; Su, S. Improving Localization Success Rate of Three Magnetic Targets Using Individual Memory-Based WO-LM Algorithm. IEEE Sens. J. 2021, 21, 3101299. [Google Scholar] [CrossRef]
Figure 1. Local and global minima representation for a generic optimization problem.
Figure 1. Local and global minima representation for a generic optimization problem.
Machines 12 00569 g001
Figure 2. The classification of the optimization problems.
Figure 2. The classification of the optimization problems.
Machines 12 00569 g002
Figure 3. Representation of the two spaces for a generic MO problem.
Figure 3. Representation of the two spaces for a generic MO problem.
Machines 12 00569 g003
Figure 4. The space of the objective functions for a generic problem with two objective functions.
Figure 4. The space of the objective functions for a generic problem with two objective functions.
Machines 12 00569 g004
Figure 5. Representation of POF for problems with 2 objective functions.
Figure 5. Representation of POF for problems with 2 objective functions.
Machines 12 00569 g005
Figure 6. Representation of weak and strong duality.
Figure 6. Representation of weak and strong duality.
Machines 12 00569 g006
Figure 7. Representation of the standard 1-crossover point operator concept.
Figure 7. Representation of the standard 1-crossover point operator concept.
Machines 12 00569 g007
Figure 8. The architecture of MEA.
Figure 8. The architecture of MEA.
Machines 12 00569 g008
Table 1. Summary of the investigated methods.
Table 1. Summary of the investigated methods.
Investigated
Methods
Main CharacteristicsSignificative
Advantages (ADV) and/or
Disadvantages (DIS)
Applications
Newton’s Method
  • Iterative method.
  • Fixed step size.
  • Suitable for unconstrained problems.
  • Second-order method.
  • Quadratic rate of convergence.
  • Needing of an initial design point.
  • High computation cost. (DIS)
  • Needing a twice differentiable function. (DIS)
  • Good initial starting point. (DIS)
  • No applications investigated for this method.
GD
  • Iterative method.
  • Variable step size.
  • First-order method.
  • Dealing with convex functions.
  • Unconstrained problems.
  • Convergence time linked to the parameter α .
  • Penalty function can be used. (ADV)
  • The parameter α needs to be set properly to avoid skipping the optimum. (DIS)
  • Function must at least once differentiable. (DIS)
  • It could skip the global optimum and converge to a local one. (DIS)
  • No applications investigated for this method.
Gauss–
Newton
  • Iterative method.
  • Suitable for convex non-linear least-squares problems.
  • It requires at least once-differentiable function.
  • Assumption: the inverse of the gradient exists.
  • It is suitable only for a specific class of problems. (DIS)
  • No applications investigated for this method.
ILS Method/
LAMBDA
  • Iterative method.
  • Suitable for ILS problems.
  • It is suitable only for a specific class of problems. (DIS)
  • Increase the resolution speed of the problem. (ADV)
  • To solve frequency ambiguity problems in GNSS applications [38,39].
LM
  • Iterative method.
  • Suitable for non-linear optimization problems.
  • It interpolates between the GN and the GD.
  • A good initial guess is needed. (DIS)
  • If the problem has several local optima, a wrong global optimum could be provided. (DIS)
  • No applications investigated for this method.
Lagrange Multipliers Method
  • Analytic method.
  • Suitable for constrained problem (with equality and/or inequality).
  • Differentiable functions are needed.
  • It guarantees only that an optimum has been found but does not guarantee that is the global one. (DIS)
  • No applications investigated for this method.
Lagrange
Duality
  • Analytic method.
  • It transforms a minimization or maximization problem into its dual one.
  • It could provide an easier problem to solve. (ADV)
  • To estimate the measured range of radio source anchors [47].
GA
  • Iterative method.
  • Metaheuristic method.
  • Suitable for constrained and unconstrained problems.
  • Inspired by natural selection of animals.
  • It works directly with strings of bits
  • It deals with discontinuous, nondifferentiable, stochastic, or highly non-linear functions.
  • It is suitable for mixed-integer type problems.
  • It can solve many types of different problems. (ADV)
  • Large solutions space can be spanned. (DIS)
  • Slow solving speed. (DIS)
  • Each considered problem needs a specific tunning. (DIS)
  • Easy to tune. (ADV)
  • Computation of parameters to superimpose information from different sensors and radio signal [9].
QGA
  • It is a variation of the GA.
  • Unlike GA, QGA deals with Q-bits and therefore can exploits superimposition of states that allows to reduce the solution time convergence.
  • The solving speed is N times faster than the speed of classic GA. (ADV)
  • High complexity. (DIS)
  • To calibrate the parameters of a wheeled vehicle to compensate the systematic errors [11]
DE
  • Iterative method.
  • Metaheuristic method.
  • Suitable for constrained and unconstrained problems.
  • Not inspired by natural selection of animals.
  • It works directly with vectors
  • It deals with continuous problems.
  • Good global search capability. (ADV)
  • Good local search capability. (ADV)
  • A memory function. (ADV)
  • A slow convergence speed. (DIS)
  • Performance affected by the trial vector. (DIS)
  • Needing a tuning procedure. (DIS)
  • To calculate the z-coordinate of the receiver in VLC system for localization purposes [59].
MEA
  • Iterative method.
  • Metaheuristic method.
  • Suitable for constrained and unconstrained problems.
  • Inspired by learning modes and human mind activities.
  • It deals with discontinuous, non-differentiable, stochastic, non-linear, non-convex functions.
  • Fast rate of convergence. (ADV)
  • It can deal with data from sensors that are used as population generator when work at high frequency. (ADV)
  • To update the forward or backward movement of a walker in a PDR system [60].
PSO
  • Iterative method.
  • Metaheuristic method.
  • Suitable for unconstrained and constrained (by using penalization methods) problems.
  • Inspired by natural swarm behaviors of animals.
  • It deals with discontinuous, non-differentiable, stochastic, non-linear, non-convex functions.
  • Fast rate of convergence. (ADV)
  • Needing a tuning procedure to set the parameters. (DIS)
  • Possibility to be stuck in a local optimum without converging to the global. (DIS)
  • To train an ANN (used in a fingerprint localization system) [63].
  • To estimate the RSSI range by minimizing the squared residual between the real and the measured distances [64].
  • To solve the localization problem modelled as a Markov process in which the position is extracted by maximizing the a posteriori probability density. The approach is an alternative to EKF or PF [65].
  • To estimate the best ANN connection weights during the training process for an on moving robot that is able to avoid collisions [77]
DEPSO
  • Hybrid algorithm in which the concepts of DE and PSO are fused.
  • Better exploration of local optima. (ADV)
  • Better convergence speed than classic DE. (ADV)
  • To reduce the particle number and the computational time complexity in a PF [67].
BA
  • Iterative method.
  • Metaheuristic method.
  • Suitable for unconstrained and constrained (by using penalization methods) problems.
  • Inspired by the echolocation behavior
  • It deals with discontinuous, non-differentiable, stochastic, non-linear, non-convex functions.
  • Fast rate of convergence. (ADV)
  • Good ability to explore the search space and to quit from local optima. (ADV)
  • Standard BA might provide a premature convergence, that it is improved by adopting LBBA. (DIS)
  • Needing a tuning procedure. (DIS)
  • To localize the robot position by minimizing the error between the real and the measured position [69].
ABC
  • Iterative method.
  • Metaheuristic method.
  • Suitable for unconstrained and constrained problems.
  • Inspired by the intelligent foraging behavior of honeybees.
  • It deals with discontinuous, non-differentiable, stochastic, non-linear, non-convex functions.
  • Good ability to explore the search space. (ADV)
  • Slow rate of convergence. (DIS)
  • Needing a tuning procedure. (DIS)
  • To solve the sensor localization problem in a wireless sensor network to get new anchors; the location is estimated by minimizing the mean localization error [72].
SA
  • Iterative method.
  • Metaheuristic method.
  • Suitable for unconstrained and constrained problems.
  • Inspired by the process of metal crystalline formation from the liquid phase.
  • It deals with discontinuous, non-differentiable, stochastic, non-linear, non-convex functions.
  • Good ability to explore the search space. (ADV)
  • Slow rate of convergence. (DIS)
  • Needing a careful tuning procedure. (DIS)
  • To estimate the real position of sensors by exploiting the anchors positions and the measured distances between them and the sensors [74].
ACO
  • Iterative method.
  • Metaheuristic method.
  • Inspired by the behavior of ants.
  • Suitable to solve the problems in which the best path needs to be selected.
  • Solve a specific class of problems. (DIS)
  • To optimize the best ANN topology during the training process for an on moving robot that is able to avoid collisions [77]
WOA
  • Iterative method.
  • Metaheuristic method.
  • Suitable for unconstrained and constrained problems.
  • Inspired by the behavior of a whale to capture a prey
  • It deals with discontinuous, non-differentiable, stochastic, non-linear, non-convex functions.
  • Slow convergence speed. (DIS)
  • In some cases, it is hybridized with other algorithms to look for local optima. (DIS)
  • To estimate the location of the permanent magnets inside the human body by measuring the magnetic field; the total error is estimated by minimizing the sum of the three errors between the measured values and the theoretical values of the magnetic field [79].
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Stefanoni, M.; Sarcevic, P.; Sárosi, J.; Odry, A. Optimization Techniques in the Localization Problem: A Survey on Recent Advances. Machines 2024, 12, 569. https://doi.org/10.3390/machines12080569

AMA Style

Stefanoni M, Sarcevic P, Sárosi J, Odry A. Optimization Techniques in the Localization Problem: A Survey on Recent Advances. Machines. 2024; 12(8):569. https://doi.org/10.3390/machines12080569

Chicago/Turabian Style

Stefanoni, Massimo, Peter Sarcevic, József Sárosi, and Akos Odry. 2024. "Optimization Techniques in the Localization Problem: A Survey on Recent Advances" Machines 12, no. 8: 569. https://doi.org/10.3390/machines12080569

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop