Title | A Homogeneous Transformer Architecture |
Publication Type | CBMM Memos |
Year of Publication | 2023 |
Authors | Gan, Y, Poggio, T |
Number | 143 |
Date Published | 09/2023 |
Abstract | While the Transformer architecture has made a substantial impact in the field of machine learning, it is unclear what purpose each component serves in the overall architecture. Heterogeneous nonlinear circuits such as multi-layer RELU networks are interleaved with layers of soft-max units. We introduce here a homogeneous architecture based on Hyper Radial Basis Function (HyperBF) units. Evalua- tions on CIFAR10, CIFAR100, and Tiny ImageNet demonstrate a performance comparable to standard vision transformers. |
DSpace@MIT |
Download:
CBMM Memo 143 v2
CBMM Memo No:
143 Associated Module:
CBMM Relationship:
- CBMM Funded