Intelligent Music Content Generation Model Based on Multimodal Situational Sentiment Perception
Abstract
To further examine the interrelationship between music, emotion, and scene, as well as to furnish novel technical assistance for music creation, the study devised a multimodal sentiment analysis model for auditory and visual features with deep learning. Based on this model, a new music content generation model was proposed, which improved upon the traditional Transformer architecture. The experimental results indicated that the minimum values of mean absolute error, root mean square error and mean absolute percentage error of the research-designed multimodal sentiment analysis architecture were 0.149, 0.166, and 0.140 respectively. The maximum value of R-squared was 0.961. The experimentally constructed multimodal sentiment analysis dataset effectively improved the performance of the model. The model performed well on Precision-Recall curve, receiver operating characteristic curve. The sentiment recognition accuracy was up to 0.98, and the recognition efficiency was high. Meanwhile, the music generated by the improved Transformer structure was closest to the dataset in terms of pitch and melody variation, with a minimum difference margin of 0.86%. The generated music performed better in terms of smoothness, coherence, and percentage of completeness. Using this model for music generation, the highest values of hit rate and normalized discounted cumulative gain could be 93.984% and 91.566%. The mean inverse rank could be up to 0.89. This study deepens the mechanism of music emotion generation, captures the emotion and context of music more accurately, and promotes the development of the fields of emotion computing and sentiment recognition.DOI:
https://doi.org/10.31449/inf.v49i5.6846Downloads
Published
How to Cite
Issue
Section
License
Authors retain copyright in their work. By submitting to and publishing with Informatica, authors grant the publisher (Slovene Society Informatika) the non-exclusive right to publish, reproduce, and distribute the article and to identify itself as the original publisher.
All articles are published under the Creative Commons Attribution license CC BY 3.0. Under this license, others may share and adapt the work for any purpose, provided appropriate credit is given and changes (if any) are indicated.
Authors may deposit and share the submitted version, accepted manuscript, and published version, provided the original publication in Informatica is properly cited.







