|Start Page Number:||251|
|End Page Number:||261|
|Publication Date:||Oct 2016|
|Journal:||Journal of Optimization Theory and Applications|
|Authors:||Shen Yuan, Wang Hongyong|
|Keywords:||programming: convex, heuristics|
The augmented Lagrangian method is a classic and efficient method for solving constrained optimization problems. However, its efficiency is still, to a large extent, dependent on how efficient the subproblem be solved. When an accurate solution to the subproblem is computationally expensive, it is more practical to relax the subproblem. Specifically, when the objective function has a certain favorable structure, the relaxed subproblem yields a closed‐form solution that can be solved efficiently. However, the resulting algorithm usually suffers from a slower convergence rate than the augmented Lagrangian method. In this paper, based on the relaxed subproblem, we propose a new algorithm with a faster convergence rate. Numerical results using the proposed approach are reported for three specific applications. The output is compared with the corresponding results from state‐of‐the‐art algorithms, and it is shown that the efficiency of the proposed method is superior to that of existing approaches.