We state concentration inequalities for the output of the hidden layers of a stochastic deep neural network (SDNN), as well as for the output of the whole SDNN. These results allow us to introduce an expected classifier (EC), and to give probabilistic upper bound for the classification error of the EC. We also state the optimal number of layers for the SDNN via an optimal stopping procedure. We apply our analysis to a stochastic version of a feedforward neural network with ReLU activation function.
%0 Journal Article
%1 Caprio2022-ad
%A Caprio, Michele
%A Mukherjee, Sayan
%D 2022
%I arXiv
%K
%T Concentration inequalities and optimal number of layers for stochastic deep neural networks
%X We state concentration inequalities for the output of the hidden layers of a stochastic deep neural network (SDNN), as well as for the output of the whole SDNN. These results allow us to introduce an expected classifier (EC), and to give probabilistic upper bound for the classification error of the EC. We also state the optimal number of layers for the SDNN via an optimal stopping procedure. We apply our analysis to a stochastic version of a feedforward neural network with ReLU activation function.
@article{Caprio2022-ad,
abstract = {We state concentration inequalities for the output of the hidden layers of a stochastic deep neural network (SDNN), as well as for the output of the whole SDNN. These results allow us to introduce an expected classifier (EC), and to give probabilistic upper bound for the classification error of the EC. We also state the optimal number of layers for the SDNN via an optimal stopping procedure. We apply our analysis to a stochastic version of a feedforward neural network with ReLU activation function.},
added-at = {2024-09-10T11:56:37.000+0200},
author = {Caprio, Michele and Mukherjee, Sayan},
biburl = {https://puma.scadsai.uni-leipzig.de/bibtex/20cb3f4ed0e6138228a7114fa26b3f4f0/scadsfct},
interhash = {94d39b8aaeef63b6c0bf1001ceef3a57},
intrahash = {0cb3f4ed0e6138228a7114fa26b3f4f0},
keywords = {},
publisher = {arXiv},
timestamp = {2024-09-10T15:15:57.000+0200},
title = {Concentration inequalities and optimal number of layers for stochastic deep neural networks},
year = 2022
}