可微函数
人工神经网络
激活函数
一般化
噪音(视频)
梯度下降
概率逻辑
计算机科学
随机共振
随机梯度下降算法
规范(哲学)
人工智能
数学
数学分析
法学
图像(数学)
政治学
作者
Fabing Duan,François Chapeau‐Blondeau,Derek Abbott
标识
DOI:10.1016/j.chaos.2023.114363
摘要
This paper proposes a flexible probabilistic activation function that enhances the training and operation of artificial neural networks by intentionally injecting noise to gain additional control over the response of each neuron. During the learning phase, the level of injected noise is iteratively optimized by gradient-descent, realizing a form of adaptive stochastic resonance. From simple hard-threshold non-differentiable neuronal responses, controlled injection of noise gives access to a wide range of useful activation functions, with sufficient differentiability to enable gradient-descent learning for both the neuron and the injected-noise levels. Experimental results on function approximation demonstrate injected noise generally converging to non-vanishing optimal levels associated with improved generalization abilities in the neural networks. A theoretical explanation of the generalization improvement based on the path norm bound is presented. With injected noise in the deep neural network, experimental results on classifying images also obtain non-vanishing optimal noise levels to achieve better testing accuracies. The proposed probabilistic activation functions show the potential of adaptive stochastic resonance for useful applications in machine learning.
科研通智能强力驱动
Strongly Powered by AbleSci AI