main
main
main
unconstrained optimization
Abstract—In order to accurately solve the unconstrained optimization problem with penalized terms in the objective
optimization problem in which the objective function is nonlinear, function and solve the problem as an unconstrained problem.
a new optimization method different from both linear search
and trust region is presented in this paper. This new method is In many nonlinear optimization problems, the objective
based on the family of proximal algorithms, we customize our function has a large number of local minima (and maxima),
algorithm to solve different unconstrained optimization problems, finding an arbitrary local optimum is relatively straightforward
and we verify the theoretical of the proposed method via different by using classical local optimization methods, but finding the
numerical examples where we compare the new algorithm with global minimum (or maximum) of a function is far more diffi-
some existing state-of-the-art algorithms. Finally the simulation cult where the analytical methods are frequently not applicable,
results proved the performance of the algorithm and indicates
and the use of numerical solution strategies often leads to
the advantage of the proposed algorithm in the case where the
hypotheses are verified. very hard challenge. In our case, where the objective function
is not quadratic function, then many optimization methods
Keywords—Optimization, Unconstrained optimization problems, use strategies to ensure that some subsequence of iterations
Proximal algorithm. converges to an optimal solution. There are two main families
of global strategies for solving problem 1, the first and the
I. I NTRODUCTION most popular for ensuring convergence relies on line searches
and the second and increasingly popular is trust region.
Optimization refers to finding an argument that minimizes
or maximizes a given function (quadratic function, nonlinear At every iteration k, the following iter x(k+1) in both line
function, the sum of squares function, etc.), many optimization searches and trust region is construct as:
algorithms seek to find the solution by solving either linear or
nonlinear equations by defining the derivative of the objective x(k+1) = x(k) + ⇢(k) d(k) (2)
function equal to zero [1].
Unconstrained optimization problems consider the problem by controlling the step length ⇢(k) and the direction d(k)
of minimizing an objective function that depends on real to forcing the descent condition:
variables with no restrictions on their values.
f (x(k) + ⇢(k) d(k) ) < f (x(k) ) (3)
Mathematically, let x 2 RN be a real vector with N 1 both strategies differ in the order in which they choose the
components and let f : RN 7! R be a convex smooth function, direction and the distance of the move, Line searches strategy
Then our unconstrained optimization problem is: chooses the direction first, followed by the distance, conversely
of the Trust-region strategy that firstly chooses the maximum
min f (x) (1) distance, followed by the direction.
x
It is well known in the philosophy of optimization that
Unconstrained optimization problems arise directly in sev- strategies for rapid convergence and robustness of the
eral applications in Numerical Optimization [2] and in Cloud algorithms is often in direct conflict with the requirements
Computing [3] [4], but they also arise indirectly from re- of computer storage and speed, for example the gradient
formulations of constrained optimization problems that has a method is simple to compute, but it is quite slow and find
wide range of applications in Energy efficient communications only local minimum [8]. There is also the newton’s method
[6] [7], and also in Internet of Things (IoT) [5] to recover that minimizes a function using knowledge of its second
in a stable way a matrix from the matrix of detection data, derivative, it can be faster for good initial guesses when the
note that constrained problems can often be transformed into second derivative is known and easy to compute, however,
unconstrained problems with the help of Lagrange multipliers the analytic expression for the second derivative is often
[13], And it is practical to replace the constraints of an complicated or intractable, and the computation of the second
derivative also requires a lot of computation.
test point where k.k2 is the usual Euclidean norm. The function min-
e(k) = x(k) + s(k) ,
x (19) imized on the righthand side is strongly convex and not
everywhere infinite, so it has a unique minimizer for every v 2
and compared to the value predicted by the local approxima- RN . Figure 2 represents what a proximal operator does. The
tion m(k) (x). If a sufficient reduction of the objective function black lines are level curves of a convex function f, the blue line
is obtained, the test point is accepted as the next iterate and indicates the boundary of its domain. Evaluating the proximal
the CRR increased or maintained constant. Otherwise, the test at the blue points moves them to the corresponding red points.
point is rejected and the CRR is contracted, in the hope that the The three points in the domain of the function stay in the
local approximation will give better predictions over a smaller domain and move towards the minimum of the function, while
region [10]. the other two move to the boundary of the domain and towards
Formally the algorithm model based on trust region is the minimum of the function. The parameter µ controls the
provided by algorithm 2 stretch to which the proximal operator maps points towards
the minimum of f, with larger values of µ associated with
Algorithm 2 : Let a starting point x(0) , an initial CRR (0) mapped points near the minimum, and smaller values giving
and the constants ⌘1 , ⌘2 , 1 , 2 which satisfy the conditions a smaller movement towards the minimum.
between two iterates while the second confines the iterated
x(k+1) around x(k) by a constraint (the trust region). Where,
on the one hand, the intensity of the penalty is governed by
the parameter µ(k) which adapts from iteration to iteration, We
have, on the other hand, a Confidence Region Radius (k) also
varying during the iterations which determines the size of the
”containment” region. Small values of µ(k) provoke a strong
penalty whose effects are similar to a trust region of small
radius (k) .
Regularization of proximal methods acts as a penalization -i.e.
a form of constraint that is weak and easy to parameterize-
Whereas trust region use a strong, impracticable and impass-
able constraint.
V. C ONCLUSION
In this article, we presented an algorithm with a progression
strategy for the proximal point parameter. This algorithm has
the advantage of being fast and robust compared to other
iterative optimization algorithms.
The main advantage of using the proximal algorithm with
the progression strategy is it’s flexibility and simplicity.
In addition, we have made a complete comparison of
the simulations results of the first order iterative algorithms
presented in the literature with those of our algorithm, which
proved the robustness of our algorithm in terms of accuracy
and speed of convergence.
Fig. 4. Performance index according to the number of iterations
Furthermore, the simulation results between the proximal
point with a constant progression step and with our proposed
strategy for advance this parameter showed a good improve-
an error Of 10 12 . These results allow us to conclude that the
ment in the accuracy of the results.
algorithm of the conjugate gradient is more efficient than the
others. In conclusion, the proximal point algorithm seems to be a
competitive candidate for optimization algorithms. We expect
In the second simulation we consider a matrix of size that the proximal point can be extended to other more complex
60⇥60 and we compare the convergence of the three relevant functions and also to other applications.
algorithms of literature with our proposed algorithm, to situate
our algorithm in relation to the algorithms already studied [2] [3] [4] [5]
previously. Figure 4 shows the convergence of these algo- [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16]
rithms. By comparing the four curves obtained in this figure,
we observe that our proposed algorithm with the progress
strategy requires a minimum of iterations with respect to the R EFERENCES
other algorithms. All these results allow us to conclude that [1] G. K. Smyth, “Optimization and nonlinear equations,” Wiley StatsRef:
our algorithm is more efficient than the others. Statistics Reference Online, pp. 1–9, 2015.
[2] R. K. Joseph Sloan, David Kesler and A. Rahimi, “A numerical
For the final simulation, we consider a matrix of size optimization-based methodology for application robustification: Trans-
60⇥80 and we compare the convergence of the two proposed forming applications for error tolerance,” in IEEE/IFIP International
Conference on Dependable Systems and Networks, August 2010.
algorithms, the results of the simulations show us that the
[3] S. B. Stefania Sardellitti, Gesualdo Scutari, “Joint optimization of radio
proximal with our strategy of advancing the parameter µ(k) and computational resources for multicell mobile-edge computing,”
gives accurate results even better than with the proximal with IEEE Trans. on Signal and Information Processing over Networks, Dec
a constant µ(k) , and we can also observe that the proximal 2014.
[4] R. B. Leila Ismail, “Implementation and performance evaluation of a
distributed conjugate gradient method in a cloud computing environ-
ment,” Software: Practice and Experience, vol. 43, pp. 281–304, March
2013.
[5] Y. X. et al., “Cognitive internet of things: A new paradigm beyond
connection,” IEEE Internet of Things Journal, vol. 1, April 2014.
[6] A. G. S Cui and A. Bahai, “Energy-constrained modulation optimiza-
tion,” IEEE Transactions on Wireless Communications, vol. 4, Sept
2005.
[7] R. Z. CK Ho, “Optimal energy allocation for wireless communications
with energy harvesting constraints,” IEEE Transactions on Signal Pro-
cessing, 2012.
[8] J. Nocedal and S. J. Wright, Numerical Optimization. Springer Series
in Operations Research. Springer, 1999.
[9] J. Stoer and R. Bulirsch, Introduction to numerical analysis. Springer-
Verlag, New York, third dition, 2002.
[10] G. e. a. Conn, A. R., Trust-Region Methods. MPS/SIAM Series on
Optimization. SIAM, Philadelphia, USA, 2000.
[11] B. Martinet, Algorithms for solving optimization and minimax problems.
PhD thesis, University of Grenoble, 1972.
[12] R. T. Rockafellar, “Monotone operators and the proximal point algo-
rithm,” SIAM Journal on Control and Optimization, pp. 877–898, 1976.
[13] D. Bertsekas, Nonlinear Programming. Springer Series in Operations
Research. Springer, 1999.
[14] A. Ahookhosh, M., “An efficient nonmonotone trust-region method for
unconstrained optimization,” Numerical Algorithms, p. 523?540, 2012.
[15] N. Parikh and S. Boyd, “Proximal algorithms,” Found. Trends Optim,
vol. 1, 2013.
[16] S. Boyd and L. Vandenberghe., “Convex optimization,” Cambridge
University Press, 2004.