On Shor's r-Algorithm for Problems with Constraints

Introduction. Nonsmooth optimization problems arise in a wide range of applications, including engineering, finance, and deep learning, where activation functions often have discontinuous derivatives, such as ReLU. Conventional optimization algorithms developed primarily for smooth problems face dif...

Full description

Bibliographic Details
Main Authors: Vladimir Norkin, Anton Kozyriev
Format: Article
Language:English
Published: V.M. Glushkov Institute of Cybernetics 2023-09-01
Series:Кібернетика та комп'ютерні технології
Subjects:
Online Access:http://cctech.org.ua/13-vertikalnoe-menyu-en/494-abstract-23-3-2-arte
_version_ 1797637777949458432
author Vladimir Norkin
Anton Kozyriev
author_facet Vladimir Norkin
Anton Kozyriev
author_sort Vladimir Norkin
collection DOAJ
description Introduction. Nonsmooth optimization problems arise in a wide range of applications, including engineering, finance, and deep learning, where activation functions often have discontinuous derivatives, such as ReLU. Conventional optimization algorithms developed primarily for smooth problems face difficulties when applied to nonsmooth contexts due to discontinuities and other associated irregularities. Possible approaches to overcome these problems include smoothing of functions and applying non-smooth optimization techniques. In particular, Shor's r-algorithm (Shor, Zhurbenko (1971), Shor (1979)) with space stretching in the direction of the difference of two subsequent subgradients is a competitive non-smooth optimization method (Bagirov et al. (2014)). However, the original r-algorithm is designed to minimize unconstrained convex functions. The goal of the work. The standard technique for applying this algorithm to problems with constraints consists in the use of exact non-smooth penalty functions (Eremin (1967), Zangwill (1967)). At the same time, it is necessary to correctly choose (quite large) the penalty coefficient of the penalty functions. Norkin (2020, 2022), Galvan et al. (2021) propose the so-called projective exact penalty functions method, which theoretically does not require choice of the penalty coefficient. The purpose of the present work is to study an applicability of the new exact projective non-smooth penalty functions method for solving conditional problems of non-smooth optimization by Shor's r-algorithm. The results. In this paper, the original optimization problem with convex constraints is first transformed into an unconstrained problem by the projective penalty function method, and then the r-algorithm is used to solve the transformed problem. The results of testing this approach on problems with linear constraints using a program implemented in Matlab are presented. The results of the present study show that the standard method of non-smooth penalties combined with Shor's r-algorithm is fast, due to the use of the provided program to calculate the subgradients, but it requires the correct selection of the penalty parameter. The projective penalty method is slow because in this study it uses finite differences to calculate the gradients, but it is quite stable with respect to the choice of the penalty parameter. Further research will be aimed at investigating the differential properties of the projection mapping and reducing the time of computing subgradients for account of parallel calculations.
first_indexed 2024-03-11T12:54:14Z
format Article
id doaj.art-1b240698f9154d26a64462772c763a34
institution Directory Open Access Journal
issn 2707-4501
2707-451X
language English
last_indexed 2024-03-11T12:54:14Z
publishDate 2023-09-01
publisher V.M. Glushkov Institute of Cybernetics
record_format Article
series Кібернетика та комп'ютерні технології
spelling doaj.art-1b240698f9154d26a64462772c763a342023-11-03T21:39:07ZengV.M. Glushkov Institute of CyberneticsКібернетика та комп'ютерні технології2707-45012707-451X2023-09-013162210.34229/2707-451X.23.3.210-34229-2707-451X-23-3-2On Shor's r-Algorithm for Problems with ConstraintsVladimir Norkin0https://orcid.org/0000-0003-3255-0405Anton Kozyriev1V.M. Glushkov Institute of Cybernetics of the NAS of Ukraine, Kyiv,National Technical University of Ukraine "Ihor Sikorsky Kyiv Polytechnic Institute", KyivIntroduction. Nonsmooth optimization problems arise in a wide range of applications, including engineering, finance, and deep learning, where activation functions often have discontinuous derivatives, such as ReLU. Conventional optimization algorithms developed primarily for smooth problems face difficulties when applied to nonsmooth contexts due to discontinuities and other associated irregularities. Possible approaches to overcome these problems include smoothing of functions and applying non-smooth optimization techniques. In particular, Shor's r-algorithm (Shor, Zhurbenko (1971), Shor (1979)) with space stretching in the direction of the difference of two subsequent subgradients is a competitive non-smooth optimization method (Bagirov et al. (2014)). However, the original r-algorithm is designed to minimize unconstrained convex functions. The goal of the work. The standard technique for applying this algorithm to problems with constraints consists in the use of exact non-smooth penalty functions (Eremin (1967), Zangwill (1967)). At the same time, it is necessary to correctly choose (quite large) the penalty coefficient of the penalty functions. Norkin (2020, 2022), Galvan et al. (2021) propose the so-called projective exact penalty functions method, which theoretically does not require choice of the penalty coefficient. The purpose of the present work is to study an applicability of the new exact projective non-smooth penalty functions method for solving conditional problems of non-smooth optimization by Shor's r-algorithm. The results. In this paper, the original optimization problem with convex constraints is first transformed into an unconstrained problem by the projective penalty function method, and then the r-algorithm is used to solve the transformed problem. The results of testing this approach on problems with linear constraints using a program implemented in Matlab are presented. The results of the present study show that the standard method of non-smooth penalties combined with Shor's r-algorithm is fast, due to the use of the provided program to calculate the subgradients, but it requires the correct selection of the penalty parameter. The projective penalty method is slow because in this study it uses finite differences to calculate the gradients, but it is quite stable with respect to the choice of the penalty parameter. Further research will be aimed at investigating the differential properties of the projection mapping and reducing the time of computing subgradients for account of parallel calculations.http://cctech.org.ua/13-vertikalnoe-menyu-en/494-abstract-23-3-2-artesubgradient descentconstrained optimizationr-algorithmexact projective penalty
spellingShingle Vladimir Norkin
Anton Kozyriev
On Shor's r-Algorithm for Problems with Constraints
Кібернетика та комп'ютерні технології
subgradient descent
constrained optimization
r-algorithm
exact projective penalty
title On Shor's r-Algorithm for Problems with Constraints
title_full On Shor's r-Algorithm for Problems with Constraints
title_fullStr On Shor's r-Algorithm for Problems with Constraints
title_full_unstemmed On Shor's r-Algorithm for Problems with Constraints
title_short On Shor's r-Algorithm for Problems with Constraints
title_sort on shor s r algorithm for problems with constraints
topic subgradient descent
constrained optimization
r-algorithm
exact projective penalty
url http://cctech.org.ua/13-vertikalnoe-menyu-en/494-abstract-23-3-2-arte
work_keys_str_mv AT vladimirnorkin onshorsralgorithmforproblemswithconstraints
AT antonkozyriev onshorsralgorithmforproblemswithconstraints