Multiplicative Regularization Generalizes Better Than Additive Regularization

TitleMultiplicative Regularization Generalizes Better Than Additive Regularization
Publication TypeCBMM Memos
Year of Publication2025
AuthorsDubach, R, Abdallah, MS, Poggio, T
Number158
Date Published07/2025
Abstract

We investigate the effectiveness of multiplicative versus additive (L2) regularization in deep neural networks, focusing on convolutional neural networks for classification. While additive methods constrain the sum of squared weights, multiplicative regularization directly penalizes the product of layerwise Frobenius norms, a quantity theoretically linked to tighter Rademacher-based generalization bounds. Through experiments on binary classification tasks in a controlled setup, we observe that multiplicative regularization consistently yields wider margin distributions, stronger rank suppression in deeper layers, and improved robustness to label noise. Under 20% label corruption, multiplicative regularization preserves margins that are 5.2% higher and achieves 3.59% higher accuracy compared to additive regularization in our main network architecture. Furthermore, multiplicative regularization achieves a 3.53% boost in test performance for multiclass classification compared to additive regularization. Our analysis of training dynamics shows that directly constraining the global product of norms leads to flatter loss landscapes that correlate with greater resilience to overfitting. These findings highlight the practical benefits of multiplicative penalties for improving generalization and stability in deep models.

DSpace@MIT

https://hdl.handle.net/1721.1/159862

Download:  PDF icon CBMM Memo 158.pdf
CBMM Memo No:  158

Associated Module: 

CBMM Relationship: 

  • CBMM Funded