Showing 11 - 20 of 8,005
In this paper we propose two new nonsymmetric primal-dual potential-reduction methods for conic problems. The methods are based on the primal-dual lifting [5]. This procedure allows to construct a strictly feasible primal-dual pair related by an exact scaling relation even if the cones are not...
Persistent link: https://www.econbiz.de/10005008570
In this paper we propose an accelerated version of the cubic regularization of Newton's method [6]. The original version, used for minimizing a convex function with Lipschitz-continuous Hessian, guarantees a global rate of convergence of order O(1/k exp.2), where k is the iteration counter. Our...
Persistent link: https://www.econbiz.de/10005065351
In this paper we analyze computational performance of dual trigonometric generating functions on some integer programming problems. We show that if the number of equality constraints is fixed, then this technique allows to solve the problems in time, which is polynomial in the dimension of the...
Persistent link: https://www.econbiz.de/10005065423
Persistent link: https://www.econbiz.de/10010926394
Persistent link: https://www.econbiz.de/10010926470
Persistent link: https://www.econbiz.de/10010926562
Persistent link: https://www.econbiz.de/10010926637
Persistent link: https://www.econbiz.de/10010926644
In this paper, we develop new subgradient methods for solving nonsmooth convex optimization problems. These methods are the first ones, for which the whole sequence of test points is endowed with the worst-case performance guarantees. The new methods are derived from a relaxed estimating...
Persistent link: https://www.econbiz.de/10010927696
We consider a new class of huge-scale problems, the problems with sparse subgradients. The most important functions of this type are piece-wise linear. For optimization problems with uniform sparsity of corresponding linear operators, we suggest a very efficient implementation of subgradient...
Persistent link: https://www.econbiz.de/10010610488