Iclr2020: Compression based bound for non-compressed network: unified generalization error analysis of large compressible deep neural network - Download as a PDF or view online for free
1) The document presents a new compression-based bound for analyzing the generalization error of large deep neural networks, even when the networks are not explicitly compressed.
2) It shows that if a trained network's weights and covariance matrices exhibit low-rank properties, then the network has a small intrinsic dimensionality and can be efficiently compressed.
3) This allows deriving a tighter generalization bound than existing approaches, providing insight into why overparameterized networks generalize well despite having more parameters than training examples.
A gradient optimization and manifold preserving based binary neural network for point cloud - ScienceDirect
ICLR 2020
Imaging With Equivariant Deep Learning
PAC-Bayesian Bound for Gaussian Process Regression and Multiple Kernel Additive Model
Minimax optimal alternating minimization \ for kernel nonparametric tensor learning
PDF] Compression based bound for non-compressed network: unified generalization error analysis of large compressible deep neural network
Single-model uncertainty quantification in neural network potentials does not consistently outperform model ensembles
Jokyokai2
GitHub - shaohua0116/ICLR2020-OpenReviewData: Script that crawls meta data from ICLR OpenReview webpage. Tutorials on installing and using Selenium and ChromeDriver on Ubuntu.
AI, Free Full-Text