激活函数
乙状窦函数
MNIST数据库
计算机科学
水准点(测量)
人工神经网络
人工智能
双曲函数
深度学习
非线性系统
功能(生物学)
模式识别(心理学)
数学
物理
数学分析
生物
进化生物学
量子力学
大地测量学
地理
作者
Serhat Kılıçarslan,Mete Çelik
标识
DOI:10.1016/j.eswa.2021.114805
摘要
In deep learning models, the inputs to the network are processed using activation functions to generate the output corresponding to these inputs. Deep learning models are of particular importance in analyzing big data with numerous parameters and forecasting and are useful for image processing, natural language processing, object recognition, and financial forecasting. Sigmoid and tangent activation functions, which are traditional activation functions, are widely used in deep learning models. However, the sigmoid and tangent activation functions face the vanishing gradient problem. In order to overcome this problem, the ReLU activation function and its derivatives were proposed in the literature. However, there is a negative region problem in these activation functions. In this study, novel RSigELU activation functions, such as single-parameter RSigELU (RSigELUS) and double-parameter (RSigELUD), which are a combination of ReLU, sigmoid, and ELU activation functions, were proposed. The proposed RSigELUS and RSigELUD activation functions can overcome the vanishing gradient and negative region problems and can be effective in the positive, negative, and linear activation regions. Performance evaluation of the proposed RSigELU activation functions was performed on the MNIST, Fashion MNIST, CIFAR-10, and IMDb Movie benchmark datasets. Experimental evaluations showed that the proposed activation functions perform better than other activation functions.
科研通智能强力驱动
Strongly Powered by AbleSci AI