Publication

Found 906 results
[ Author(Asc)] Title Type Year
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z 
P
Pouncy, T. & Gershman, S. J. Inductive biases in theory-based reinforcement learning. Cognitive Psychology 138, 101509 (2022).
Pouncy, T., Tsividis, P. & Gershman, S. J. What Is the Model in Model‐Based Planning?. Cognitive Science 45, (2021).
Ponce, C. R. et al. Evolving Images for Visual Neurons Using a Deep Generative Network Reveals Coding Principles and Neuronal Preferences. Cell 177, 1009 (2019).PDF icon Author's last draft (20.26 MB)
Poggio, T. What if.. (2015).PDF icon What if.pdf (2.09 MB)
Poggio, T., Rosasco, L., Shashua, A., Cohen, N. & Anselmi, F. Notes on Hierarchical Splines, DCLNs and i-theory. (2015).PDF icon CBMM Memo 037 (1.83 MB)
Poggio, T. et al. Theory of Deep Learning III: explaining the non-overfitting puzzle. (2017).PDF icon CBMM-Memo-073.pdf (2.65 MB)PDF icon CBMM Memo 073 v2 (revised 1/15/2018) (2.81 MB)PDF icon CBMM Memo 073 v3 (revised 1/30/2018) (2.72 MB)PDF icon CBMM Memo 073 v4 (revised 12/30/2018) (575.72 KB)
Poggio, T. Deep Learning: mathematics and neuroscience. (2016).PDF icon Deep Learning- mathematics and neuroscience.pdf (1.25 MB)
Poggio, T. A Perspective: Sparse Compositionality and Efficiently Computable Intelligence. (2026).PDF icon Perspective_SPCOMP-9.pdf (170.23 KB)
Poggio, T., Kur, G. & Banburski, A. Double descent in the condition number. (2019).PDF icon Fixing typos, clarifying error in y, best approach is crossvalidation (837.18 KB)PDF icon Incorporated footnote in text plus other edits (854.05 KB)PDF icon Deleted previous discussion on kernel regression and deep nets: it will appear, extended, in a separate paper (795.28 KB)PDF icon correcting a bad typo (261.24 KB)PDF icon Deleted plot of condition number of kernel matrix: we cannot get a double descent curve  (769.32 KB)
Poggio, T. & Anselmi, F. Visual Cortex and Deep Networks: Learning Invariant Representations. 136 (The MIT Press, 2016). at <https://mitpress.mit.edu/books/visual-cortex-and-deep-networks>
Poggio, T. & Fraser, M. Compositional sparsity of learnable functions. Bulletin of the American Mathematical Society 61, 438-456 (2024).
Poggio, T. Stable Foundations for Learning: a framework for learning theory (in both the classical and modern regime). (2020).PDF icon Original file (584.54 KB)PDF icon Corrected typos and details of "equivalence" CV stability and expected error for interpolating machines. Added Appendix on SGD.  (905.29 KB)PDF icon Edited Appendix on SGD. (909.19 KB)PDF icon Deleted Appendix. Corrected typos etc (880.27 KB)PDF icon Added result about square loss and min norm (898.03 KB)
Poggio, T. A. & Xu, M. On efficiently computable functions, deep networks and sparse compositionality. (2025).PDF icon Deep_sparse_networks_approximate_efficiently_computable_functions.pdf (223.15 KB)
Poggio, T. & Liao, Q. Explicit regularization and implicit bias in deep network classifiers trained with the square loss. arXiv (2020). at <https://arxiv.org/abs/2101.00072>
Poggio, T. Deep Leaning: Mathematics and Neuroscience. A Sponsored Supplement to Science Brain-Inspired intelligent robotics: The intersection of robotics and neuroscience, 9-12 (2016).
Poggio, T. How Deep Sparse Networks Avoid the Curse of Dimensionality: Efficiently Computable Functions are Compositionally Sparse. (2022).PDF icon v1.0 (984.15 KB)PDF icon v5.7 adding in context learning etc (1.16 MB)
Poggio, T., Liao, Q. & Xu, M. Implicit dynamic regularization in deep networks. (2020).PDF icon v1.2 (2.29 MB)PDF icon v.59 Update on rank (2.43 MB)
Poggio, T. From Marr’s Vision to the Problem of Human Intelligence. (2021).PDF icon CBMM-Memo-118.pdf (362.19 KB)
Poggio, T., Liao, Q. & Banburski, A. Complexity Control by Gradient Descent in Deep Networks. Nature Communications 11, (2020).PDF icon s41467-020-14663-9.pdf (431.68 KB)
Poggio, T. Associative Memory as the Core of Intelligence in Technology and Evolution. (2026).PDF icon Review_On_Associative_Memories-14.pdf (245.78 KB)
Poggio, T. & Liao, Q. Theory II: Landscape of the Empirical Risk in Deep Learning. (2017).PDF icon CBMM Memo 066_1703.09833v2.pdf (5.56 MB)
Poggio, T. & Liao, Q. Theory II: Deep learning and optimization. Bulletin of the Polish Academy of Sciences: Technical Sciences 66, (2018).PDF icon 03_775-788_00920_Bpast.No_.66-6_31.12.18_K2.pdf (5.43 MB)
Poggio, T. From Associative Memories to Powerful Machines. (2021).PDF icon v1.0 (1.01 MB)PDF icon v1.3Section added August 6 on self attention (3.9 MB)
Poggio, T. Is Research in Intelligence an Existential Risk?. (2014).PDF icon Is Research in Intelligence an Existential Risk.pdf (571.42 KB)
Poggio, T., Anselmi, F. & Rosasco, L. I-theory on depth vs width: hierarchical function composition. (2015).PDF icon cbmm_memo_041.pdf (1.18 MB)

Pages