计算机科学
卷积神经网络
树(集合论)
人工智能
残余物
推论
深度学习
频道(广播)
激活函数
领域(数学)
模式识别(心理学)
领域(数学分析)
人工神经网络
机器学习
算法
数学
数学分析
计算机网络
纯数学
作者
Hua Shen,Zhiwei Wang,Jixin Zhang,Mingwu Zhang
标识
DOI:10.1016/j.ins.2024.120131
摘要
Deep learning (DL) has demonstrated exceptional success across various domains, including computer vision, natural language processing, and speech recognition. However, the training and inference processes of DL models typically require substantial computational resources and storage space, presenting a significant challenge within the Internet of Things (IoT) domain. This study contributes theoretically to the field of lightweight DL by proposing L-Net, a lightweight convolutional neural network designed specifically for low-compute devices. The L-Net addresses challenges associated with channel interaction disparities and vanishing gradients. To further improve the network performance, we introduce the residual enhanced channel attention (or R-ECA) module, which combines a bypass mechanism derived from simplified residual learning with the attention mechanism's cross-channel interaction. Additionally, we replace the rectified linear unit function (or ReLU) with an exponential linear unit (or ELU) function to enhance the network's nonlinear expression capability and training speed. We conducted object recognition experiments and compared the accuracy and prediction stability of L-Net with well-known models, such as AlexNet, VGG11, SqueezeNet, ResNet, and MobileNet, to assess its efficacy. Using the CIFAR-10 dataset and our custom dataset of apple tree leaf diseases, our experimental results demonstrate that, with relatively smaller model parameters, L-Net performs exceptionally well in terms of mean Average Precision (mAP), achieving 0.906. Furthermore, when applied to our custom dataset, L-Net exhibits relatively consistent performance across various dataset splits under different ratios, outperforming the majority of models.
科研通智能强力驱动
Strongly Powered by AbleSci AI