Efficient training of deep learning models...
URL: https://doi.org/10.1007/978-3-030-77004-4_14
Training of Deep Neural Networks (DNNs) is very computationally demanding and resources are typically spent on training-instances that do not provide the most benefit to a network’s learning; instead, the most relevant instances should be prioritized during training. Herein we present an improved version of the Adaptive Sampling (AS) method (Gopal, 2016) extended for the training of DNNs. As our main contribution we formulate a probability distribution for data instances that minimizes the variance of the gradient-norms w.r.t. the network’s loss function. Said distribution is combined with the optimal distribution for the data classes previously derived by Gopal and the improved AS is used to replace uniform sampling with the objective of accelerating the training of DNNs. Our proposal is comparatively evaluated against uniform sampling and against Online Batch Selection (Loshchilov & Hutter, 2015). Results from training a Convolutional Neural Network on the MNIST dataset with the Adadelta and Adam optimizers over different training batch-sizes show the effectiveness and superiority of our proposal.
Todavía no existen vistas creadas para este recurso.
Información adicional
Campo | Valor |
---|---|
Última actualización de los datos | 11 de octubre de 2025 |
Última actualización de los metadatos | 11 de octubre de 2025 |
Creado | 11 de octubre de 2025 |
Formato | HTML |
Licencia | No se ha provisto de una licencia |
Id | 25abfb23-4d55-4f3b-bc2e-e1c0efdc392d |
Package id | dd4b450d-e41a-40f9-b9ac-50ce286b734b |
State | active |