Publication
Fast, invariant representation for human action in the visual system. (2016). at <http://arxiv.org/abs/1601.01358>
CBMM Memo 042 (3.03 MB)
Feature learning in deep classifiers through Intermediate Neural Collapse. (2023).
Feature_Learning_memo.pdf (2.16 MB)
For HyperBFs AGOP is a greedy approximation to gradient descent. (2024).
CBMM-Memo-148.pdf (1.06 MB)
For interpolating kernel machines, the minimum norm ERM solution is the most stable. (2020).
CBMM_Memo_108.pdf (1015.14 KB)
Better bound (without inequalities!) (1.03 MB)
Formation of Representations in Neural Networks. (2024).
CBMM-Memo-150.pdf (4.03 MB)
Foveation-based Mechanisms Alleviate Adversarial Examples. (2016).
cbmm_memo_044.pdf (11.48 MB)
From Associative Memories to Powerful Machines. (2021).
v1.0 (1.01 MB)
v1.3Section added August 6 on self attention (3.9 MB)
From Marr’s Vision to the Problem of Human Intelligence. (2021).
CBMM-Memo-118.pdf (362.19 KB)
On Generalization Bounds for Neural Networks with Low Rank Layers. (2024).
CBMM-Memo-151.pdf (697.31 KB)
Group Invariant Deep Representations for Image Instance Retrieval. (2016).
CBMM-Memo-043.pdf (2.66 MB)
Hierarchically Local Tasks and Deep Convolutional Networks. (2020).
CBMM_Memo_109.pdf (2.12 MB)
Holographic Embeddings of Knowledge Graphs. (2015).
holographic-embeddings.pdf (677.87 KB)
A Homogeneous Transformer Architecture. (2023).
CBMM Memo 143 v2 (1.1 MB)
How Deep Sparse Networks Avoid the Curse of Dimensionality: Efficiently Computable Functions are Compositionally Sparse. (2022).
v1.0 (984.15 KB)
v5.7 adding in context learning etc (1.16 MB)
How Important is Weight Symmetry in Backpropagation?. (2015).
1510.05067v3.pdf (615.32 KB)
Implicit dynamic regularization in deep networks. (2020).
v1.2 (2.29 MB)
v.59 Update on rank (2.43 MB)
On Invariance and Selectivity in Representation Learning. (2015).
CBMM Memo No. 029 (812.07 KB)
The Invariance Hypothesis Implies Domain-Specific Regions in Visual Cortex. (2014). doi:10.1101/004473
CBMM Memo 004_new.pdf (2.25 MB)
I-theory on depth vs width: hierarchical function composition. (2015).
cbmm_memo_041.pdf (1.18 MB)
The Janus effects of SGD vs GD: high noise and low rank. (2023).
Updated with appendix showing empirically that the main results extend to deep nonlinear networks (2.95 MB)
Small updates...typos... (616.82 KB)
Learning An Invariant Speech Representation. (2014).
CBMM-Memo-022-1406.3884v1.pdf (1.81 MB)
Learning Functions: When Is Deep Better Than Shallow. (2016). at <https://arxiv.org/pdf/1603.00988v4.pdf>
Loss landscape: SGD has a better view. (2020).
CBMM-Memo-107.pdf (1.03 MB)
Typos and small edits, ver11 (955.08 KB)
Small edits, corrected Hessian for spurious case (337.19 KB)
Multiplicative Regularization Generalizes Better Than Additive Regularization. (2025).
CBMM Memo 158.pdf (4.8 MB)
Musings on Deep Learning: Properties of SGD. (2017).
CBMM Memo 067 v2 (revised 7/19/2017) (5.88 MB)
CBMM Memo 067 v3 (revised 9/15/2017) (5.89 MB)
CBMM Memo 067 v4 (revised 12/26/2017) (5.57 MB)
]