Unknown

Dataset Information

0

Complexity control by gradient descent in deep networks.


ABSTRACT: Overparametrized deep networks predict well, despite the lack of an explicit complexity control during training, such as an explicit regularization term. For exponential-type loss functions, we solve this puzzle by showing an effective regularization effect of gradient descent in terms of the normalized weights that are relevant for classification.

SUBMITTER: Poggio T 

PROVIDER: S-EPMC7039878 | biostudies-literature | 2020 Feb

REPOSITORIES: biostudies-literature

altmetric image

Publications

Complexity control by gradient descent in deep networks.

Poggio Tomaso T   Liao Qianli Q   Banburski Andrzej A  

Nature communications 20200224 1


Overparametrized deep networks predict well, despite the lack of an explicit complexity control during training, such as an explicit regularization term. For exponential-type loss functions, we solve this puzzle by showing an effective regularization effect of gradient descent in terms of the normalized weights that are relevant for classification. ...[more]

Similar Datasets

| S-EPMC6605414 | biostudies-literature
| S-EPMC8563972 | biostudies-literature
| S-EPMC6611597 | biostudies-literature
| S-EPMC11244533 | biostudies-literature
| S-EPMC2933224 | biostudies-literature
| S-EPMC6923147 | biostudies-literature
| S-EPMC10542513 | biostudies-literature
| S-EPMC7402485 | biostudies-literature
| S-EPMC7714173 | biostudies-literature
| S-EPMC9198964 | biostudies-literature