Paper
13 May 2019 Relating information complexity and training in deep neural networks
Alex Gain, Hava Siegelmann
Author Affiliations +
Abstract
Deep Neural Networks may be costly to train, and if testing error is too large, retraining may be required, unless lifelong learning methods are applied. Crucial to addressing learning at the edge, without access to powerful cloud computing, is the notion of problem difficulty for non-standard data domains. While it is known that training is harder for classes that are more entangled, the complexity of data points was not previously studied as an important contributor to training dynamics. We analyze data points by their information complexity and relate the complexity of the data to the test error. We elucidate training dynamics of DNNs, demonstrating that high complexity datapoints contribute to the error of the network, and that training DNNs consist of two important aspects - (1) Minimization of error due to high complexity datapoints, and (2) Margin decrease where entanglement of classes occurs. Whereas data complexity may be ignored when training in a cloud, it must be considered as part of the setting when training at the edge.
© (2019) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Alex Gain and Hava Siegelmann "Relating information complexity and training in deep neural networks", Proc. SPIE 10982, Micro- and Nanotechnology Sensors, Systems, and Applications XI, 109822H (13 May 2019); https://doi.org/10.1117/12.2520172
Lens.org Logo
CITATIONS
Cited by 1 scholarly publication.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Error analysis

Image information entropy

Clouds

Neural networks

Image filtering

Binary data

Computer science

Back to Top