计算机科学
元数据
潜在Dirichlet分配
集合(抽象数据类型)
主题模型
过程(计算)
人工智能
动力学(音乐)
音乐剧
感知
多媒体
机器学习
万维网
艺术
物理
视觉艺术
神经科学
操作系统
生物
程序设计语言
声学
作者
Khaled Boughanmi,Asim Ansari
标识
DOI:10.1177/00222437211016495
摘要
The success of creative products depends on the felt experience of consumers. Capturing such consumer reactions requires the fusing of different types of experiential covariates and perceptual data in an integrated modeling framework. In this article, the authors develop a novel multimodal machine learning framework that combines multimedia data (e.g., metadata, acoustic features, user-generated textual data) in creative product settings and apply it to predict the success of musical albums and playlists. The authors estimate the proposed model on a unique data set collected using different online sources. The model integrates different types of nonparametrics to flexibly accommodate diverse types of effects. It uses penalized splines to capture the nonlinear impact of acoustic features and a supervised hierarchical Dirichlet process to represent crowd sourced textual tags, and it captures dynamics via a state-space specification. The authors show the predictive superiority of the model with respect to several benchmarks. The results illuminate the dynamics of musical success over the past five decades. The authors then use the components of the model for marketing decisions such as forecasting the success of new albums, conducting album tuning and diagnostics, constructing playlists for different generations of music listeners, and providing contextual recommendations.
科研通智能强力驱动
Strongly Powered by AbleSci AI