Publication

Export 129 results:
Filters: Author is T. Poggio  [Clear All Filters]
2020
Mhaskar, H. & Poggio, T. An analysis of training and generalization errors in shallow and deep networks. Neural Networks 121, 229 - 241 (2020).
Reddy, M. Vuyyuru, Banburski, A., Pant, N. & Poggio, T. Biologically Inspired Mechanisms for Adversarial Robustness. (2020).PDF icon CBMM_Memo_110.pdf (3.14 MB)
Poggio, T., Liao, Q. & Banburski, A. Complexity Control by Gradient Descent in Deep Networks. Nature Communications 11, (2020).PDF icon s41467-020-14663-9.pdf (431.68 KB)
Malkin, E., Deza, A. & Poggio, T. CUDA-Optimized real-time rendering of a Foveated Visual System. Shared Visual Representations in Human and Machine Intelligence (SVRHM) workshop at NeurIPS 2020 (2020). at <https://arxiv.org/abs/2012.08655>PDF icon Foveated_Drone_SVRHM_2020.pdf (13.44 MB)PDF icon v1 (12/15/2020) (14.7 MB)
Banburski, A. et al. Dreaming with ARC. Learning Meets Combinatorial Algorithms workshop at NeurIPS 2020 (2020).PDF icon CBMM Memo 113.pdf (1019.64 KB)
Rangamani, A., Rosasco, L. & Poggio, T. For interpolating kernel machines, the minimum norm ERM solution is the most stable. (2020).PDF icon CBMM_Memo_108.pdf (1015.14 KB)PDF icon Better bound (without inequalities!) (1.03 MB)
Mhaskar, H. & Poggio, T. Function approximation by deep networks. Communications on Pure & Applied Analysis 19, 4085 - 4095 (2020).PDF icon 1534-0392_2020_8_4085.pdf (514.57 KB)
Deza, A., Liao, Q., Banburski, A. & Poggio, T. Hierarchically Local Tasks and Deep Convolutional Networks. (2020).PDF icon CBMM_Memo_109.pdf (2.12 MB)
Poggio, T. & Liao, Q. Implicit dynamic regularization in deep networks. (2020).PDF icon TPR_ver2.pdf (2.29 MB)PDF icon Substantial edits (1.52 MB)PDF icon Edits that are extensive but minor in content (1.98 MB)PDF icon Extending theory, setting a post (2 MB)PDF icon Fine tuning (2.01 MB)PDF icon Corrections in Appendix about Neural Collapse (2.01 MB)PDF icon Small edits clarifying role of weight decay (2.39 MB)PDF icon Added: prove NC for multiclass+theorem on connected global minima (2.4 MB)
Poggio, T. & Cooper, Y. Loss landscape: SGD has a better view. (2020).PDF icon CBMM-Memo-107.pdf (1.03 MB)PDF icon Typos and small edits, ver11 (955.08 KB)PDF icon Small edits, corrected Hessian for spurious case (337.19 KB)
Poggio, T. & Banburski, A. An Overview of Some Issues in the Theory of Deep Networks. IEEJ Transactions on Electrical and Electronic Engineering 15, 1560 - 1571 (2020).
Han, Y., Roig, G., Geiger, G. & Poggio, T. Scale and translation-invariance for novel objects in human vision. Scientific Reports 10, (2020).PDF icon s41598-019-57261-6.pdf (1.46 MB)
Poggio, T. Stable Foundations for Learning: a framework for learning theory (in both the classical and modern regime). (2020).PDF icon Original file (584.54 KB)PDF icon Corrected typos and details of "equivalence" CV stability and expected error for interpolating machines. Added Appendix on SGD.  (905.29 KB)PDF icon Edited Appendix on SGD. (909.19 KB)PDF icon Deleted Appendix. Corrected typos etc (880.27 KB)PDF icon Added result about square loss and min norm (898.03 KB)
Poggio, T., Banburski, A. & Liao, Q. Theoretical issues in deep networks. Proceedings of the National Academy of Sciences 201907369 (2020). doi:10.1073/pnas.1907369117PDF icon PNASlast.pdf (915.3 KB)
2019
Mhaskar, H. & Poggio, T. An analysis of training and generalization errors in shallow and deep networks. (2019).PDF icon CBMM-Memo-098.pdf (687.36 KB)PDF icon CBMM Memo 098 v4 (08/2019) (2.63 MB)
Xiao, W., Chen, H., Liao, Q. & Poggio, T. Biologically-plausible learning algorithms can scale to large datasets. International Conference on Learning Representations, (ICLR 2019) (2019).PDF icon gk7779.pdf (721.53 KB)
Adler, A., Araya-Polo, M. & Poggio, T. Deep Recurrent Architectures for Seismic Tomography. 81st EAGE Conference and Exhibition 2019 (2019).
Poggio, T., Kur, G. & Banburski, A. Double descent in the condition number. (2019).PDF icon Fixing typos, clarifying error in y, best approach is crossvalidation (837.18 KB)PDF icon Incorporated footnote in text plus other edits (854.05 KB)PDF icon Deleted previous discussion on kernel regression and deep nets: it will appear, extended, in a separate paper (795.28 KB)PDF icon correcting a bad typo (261.24 KB)PDF icon Deleted plot of condition number of kernel matrix: we cannot get a double descent curve  (769.32 KB)
Banburski, A. et al. Dynamics & Generalization in Deep Networks -Minimizing the Norm. NAS Sackler Colloquium on Science of Deep Learning (2019).
Zhang, J., Han, Y., Poggio, T. & Roig, G. Eccentricity Dependent Neural Network with Recurrent Attention for Scale, Translation and Clutter Invariance . Vision Science Society (2019).

Pages