配方
计算机科学
品味
人工智能
图像(数学)
自然语言处理
融合
模式识别(心理学)
计算机视觉
语言学
心理学
历史
神经科学
哲学
考古
作者
Yawei Chen,Min Cao,Wen-Jing Gao
出处
期刊:2020 5th International Conference on Smart Grid and Electrical Automation (ICSGEA)
日期:2020-06-01
卷期号:: 203-208
被引量:4
标识
DOI:10.1109/icsgea51094.2020.00050
摘要
It is difficult for taste classification of Chinese recipe to achieve satisfactory results based on single-modal data. However, there are few studies on multimodal analysis in this field. In this paper, we put forward to tackle taste classification for Chinese recipe based on image and text fusion algorithms. Firstly, visual features and textual features are extracted from different models, including a convolutional neural network (CNN) constructed for visual feature extraction and a pretrained word2vec model combined with a multi-layer perception network for textual feature extraction. Secondly, two fusion strategies, called feature-level and decision-level fusion, are designed to perform multimodal fusion for the final taste prediction. Several experiments are carried out with K-fold cross-validation to verify the effectiveness of our proposed model. The results show that the multimodal fusion model for taste classification is superior to those based on single-modal features. Besides, compared with feature-level fusion, decision-level fusion performs better in the task of taste classification for Chinese recipe.
科研通智能强力驱动
Strongly Powered by AbleSci AI