Worst case complexity of direct search under convexity

In this paper we prove that the broad class of direct-search methods of directional type, based on imposing sufficient decrease to accept new iterates, exhibits the same worst case complexity bound and global rate of the gradient method for the unconstrained minimization of a convex and smooth funct...

ver descrição completa

Detalhes bibliográficos
Autor principal: Dodangeh, Mahdi (author)
Outros Autores: Vicente, Luís Nunes (author)
Formato: article
Idioma:eng
Publicado em: 2016
Texto completo:http://hdl.handle.net/10316/45246
País:Portugal
Oai:oai:estudogeral.sib.uc.pt:10316/45246
Descrição
Resumo:In this paper we prove that the broad class of direct-search methods of directional type, based on imposing sufficient decrease to accept new iterates, exhibits the same worst case complexity bound and global rate of the gradient method for the unconstrained minimization of a convex and smooth function. More precisely, it will be shown that the number of iterations needed to reduce the norm of the gradient of the objective function below a certain threshold is at most proportional to the inverse of the threshold. It will be also shown that the absolute error in the function values decay at a sublinear rate proportional to the inverse of the iteration counter. In addition, we prove that the sequence of absolute errors of function values and iterates converges r-linearly in the strongly convex case.