Trust-Region Minimization Algorithm for Training Responses (TRMinATR): The Rise of Machine Learning Techniques
Title | Trust-Region Minimization Algorithm for Training Responses (TRMinATR): The Rise of Machine Learning Techniques |
Publication Type | Conference Paper |
Year of Publication | 2018 |
Authors | Rafati, Jacob, DeGuchy, Omar, Marcia, Roummel F. |
Conference Name | 2018 26th European Signal Processing Conference (EUSIPCO) |
Date Published | Sept. 2018 |
Publisher | IEEE |
ISBN Number | 978-9-0827-9701-5 |
Keywords | approximation theory, computational complexity, Computer architecture, computer theory, Deep Learning, Europe, gradient descent methods, gradient methods, Hessian approximations, Hessian matrices, Hessian matrix inversion, human factors, learning (artificial intelligence), limited memory BFGS quasiNewton method, Limited-memory BFGS, Line-search methods, machine learning, matrix inversion, memory storage, Neural networks, Newton method, nonconvex functions, optimisation, Optimization, pubcrawl, Quasi-Newton methods, resilience, Resiliency, Scalability, security, Signal processing algorithms, storage management, Training, training responses, TRMinATR, Trust-region methods, trust-region minimization algorithm |
Abstract | Deep learning is a highly effective machine learning technique for large-scale problems. The optimization of nonconvex functions in deep learning literature is typically restricted to the class of first-order algorithms. These methods rely on gradient information because of the computational complexity associated with the second derivative Hessian matrix inversion and the memory storage required in large scale data problems. The reward for using second derivative information is that the methods can result in improved convergence properties for problems typically found in a non-convex setting such as saddle points and local minima. In this paper we introduce TRMinATR - an algorithm based on the limited memory BFGS quasi-Newton method using trust region - as an alternative to gradient descent methods. TRMinATR bridges the disparity between first order methods and second order methods by continuing to use gradient information to calculate Hessian approximations. We provide empirical results on the classification task of the MNIST dataset and show robust convergence with preferred generalization characteristics. |
URL | https://ieeexplore.ieee.org/document/8553243 |
DOI | 10.23919/EUSIPCO.2018.8553243 |
Citation Key | rafati_trust-region_2018 |
- Scalability
- Neural networks
- Newton method
- nonconvex functions
- optimisation
- optimization
- pubcrawl
- Quasi-Newton methods
- resilience
- Resiliency
- memory storage
- security
- Signal processing algorithms
- storage management
- Training
- training responses
- TRMinATR
- Trust-region methods
- trust-region minimization algorithm
- Hessian matrices
- computational complexity
- computer architecture
- computer theory
- deep learning
- Europe
- gradient descent methods
- gradient methods
- Hessian approximations
- approximation theory
- Hessian matrix inversion
- Human Factors
- learning (artificial intelligence)
- limited memory BFGS quasiNewton method
- Limited-memory BFGS
- Line-search methods
- machine learning
- matrix inversion