An evolutionary algorithm of linear complexity: application to training of deep neural networks

The performance of deep neural networks, such as Deep Belief Networks formed by Restricted Boltzmann Machines (RBMs), strongly depends on their training, which is the process of adjusting their parameters. This process can be posed as an optimization problem over n dimensions. However, typical networks contain tens of thousands of parameters, making this a High-Dimensional Problem (HDP). Although different optimization methods have been employed for this goal, the use of most of the Evolutionary Algorithms (EAs) becomes prohibitive due to their inability to deal with HDPs. For instance, the Covariance Matrix Adaptation Evolutionary Strategy (CMA-ES) which is regarded as one of the most effective EAs, exhibits the enormous disadvantage of requiring $O(n^2)$ memory and operations, making it unpractical for problems with more than a few hundred variables. In this paper, we introduce a novel EA that requires $O(n)$ operations and memory, but delivers competitive solutions for the training stage of RBMs with over one million variables, when compared against CMA-ES and the Contrastive Divergence algorithm, which is the standard method for training RBMs.

Datos y Recursos

Información Adicional

Campo Valor
Fuente https://scholar.google.com/citations?view_op=view_citation&hl=es&user=MG1jyREAAAAJ&pagesize=100&sortby=pubdate&citation_for_view=MG1jyREAAAAJ:70eg2SAEIzsC
Autor SI Valdez, A Rojas-Domínguez
Última actualización octubre 21, 2025, 09:02 (UTC)
Creado octubre 21, 2025, 09:02 (UTC)
Año 2019
Google Scholar URL https://scholar.google.com/citations?view_op=view_citation&hl=es&user=MG1jyREAAAAJ&pagesize=100&sortby=pubdate&citation_for_view=MG1jyREAAAAJ:70eg2SAEIzsC
Identificador hash 968ea34084e1
Lugar de publicación arXiv preprint arXiv:1907.05951, 2019
Tipo Publicación
Tipo de publicación Otro
URL directo https://arxiv.org/pdf/1907.05951