Submit Search
Upload
Locations
•
Download as PPTX, PDF
•
0 likes
•
179 views
H
hazelbaz
Follow
Report
Share
Report
Share
1 of 3
Download now
Recommended
Presentation2
Presentation2
KatiePearce
These slides present a novel content-based system that utilizes the perceived emotion of multimedia content as a bridge to connect music and video. Specifically, we propose a novel machine learning framework, called Acousticvisual Emotion Gaussians (AVEG), to jointly learn the tripartite relationship among music, video, and emotion from an emotion-annotated corpus of music videos. For a music piece (or a video sequence), the AVEG model is applied to predict its emotion distribution in a stochastic emotion space from the corresponding low-level acoustic (resp. visual) features. Finally, music and video are matched by measuring the similarity between the two corresponding emotion distributions, based on a distance measure such as KL divergence.
ACM Multimedia 2012 Grand Challenge: Music Video Generation
ACM Multimedia 2012 Grand Challenge: Music Video Generation
Ju-Chiang Wang
Newspapers need to establish a consistent brand and deliver a quality-reading experience to an audience scattered across an ever-expanding universe of devices and platforms. This ain't easy. After two years of research, we're ready to share what we've learned.
Publish Once, Brand Everywhere
Publish Once, Brand Everywhere
laurendreier
Paragraphs
Paragraphs
Laura Cardona
Teoría Discursova- tipos de párrafos
Teoría Discursova- tipos de párrafos
Laura Cardona
Trabajo realizado para la asignatura teg y sociedad de la Universidad ECCI
Politica publica de tic
Politica publica de tic
Javier Perdomo
Este es un trabajo elaborado para la asignatura de tecnología y sociedad de la universidad ECCI
Politica publica de tic
Politica publica de tic
Javier Perdomo
Say you like me
Say you like me
hazelbaz
Recommended
Presentation2
Presentation2
KatiePearce
These slides present a novel content-based system that utilizes the perceived emotion of multimedia content as a bridge to connect music and video. Specifically, we propose a novel machine learning framework, called Acousticvisual Emotion Gaussians (AVEG), to jointly learn the tripartite relationship among music, video, and emotion from an emotion-annotated corpus of music videos. For a music piece (or a video sequence), the AVEG model is applied to predict its emotion distribution in a stochastic emotion space from the corresponding low-level acoustic (resp. visual) features. Finally, music and video are matched by measuring the similarity between the two corresponding emotion distributions, based on a distance measure such as KL divergence.
ACM Multimedia 2012 Grand Challenge: Music Video Generation
ACM Multimedia 2012 Grand Challenge: Music Video Generation
Ju-Chiang Wang
Newspapers need to establish a consistent brand and deliver a quality-reading experience to an audience scattered across an ever-expanding universe of devices and platforms. This ain't easy. After two years of research, we're ready to share what we've learned.
Publish Once, Brand Everywhere
Publish Once, Brand Everywhere
laurendreier
Paragraphs
Paragraphs
Laura Cardona
Teoría Discursova- tipos de párrafos
Teoría Discursova- tipos de párrafos
Laura Cardona
Trabajo realizado para la asignatura teg y sociedad de la Universidad ECCI
Politica publica de tic
Politica publica de tic
Javier Perdomo
Este es un trabajo elaborado para la asignatura de tecnología y sociedad de la universidad ECCI
Politica publica de tic
Politica publica de tic
Javier Perdomo
Say you like me
Say you like me
hazelbaz
One of the most exciting but challenging endeavors in music research is to develop a computational model that comprehends the affective content of music signals and organizes a music collection according to emotion. In this paper, we propose a novel \emph{acoustic emotion Gaussians} (AEG) model that defines a proper generative process of emotion perception in music. As a generative model, AEG permits easy and straightforward interpretations of the model learning processes. To bridge the acoustic feature space and music emotion space, a set of \emph{latent feature classes}, which are learned from data, is introduced to perform the end-to-end semantic mappings between the two spaces. Based on the space of latent feature classes, the AEG model is applicable to both automatic music emotion annotation and emotion-based music retrieval. To gain insights into the AEG model, we also provide illustrations of the model learning process. A comprehensive performance study is conducted to demonstrate the superior accuracy of AEG over its predecessors, using two emotion annotated music corpora MER60 and MTurk. Our results show that the AEG model outperforms the state-of-the-art methods in automatic music emotion annotation. Moreover, for the first time a quantitative evaluation of emotion-based music retrieval is reported.
The Acoustic Emotion Gaussians Model for Emotion-based Music Annotation and R...
The Acoustic Emotion Gaussians Model for Emotion-based Music Annotation and R...
Ju-Chiang Wang
Computational modeling of music emotion has been addressed primarily by two approaches: the categorical approach that categorizes emotions into mood classes and the dimensional approach that regards emotions as numerical values over a few dimensions such as valence and activation. Being two extreme scenarios (discrete/continuous), the two approaches actually share a unified goal of understanding the emotion semantics of music. This paper presents the first computational model that unifies the two semantic modalities under a probabilistic framework, which makes it possible to explore the relationship between them in a computational way. With the proposed framework, mood labels can be mapped into the emotion space in an unsupervised and content-based manner, without any training ground truth annotations for the semantic mapping. Such a function can be applied to automatically generate a semantically structured tag cloud in the emotion space. To demonstrate the effectiveness of the proposed framework, we qualitatively evaluate the mood tag clouds generated from two emotion-annotated corpora, and quantitatively evaluate the accuracy of the categorical-dimensional mapping by comparing the results with those created by psychologists, including the one proposed by Whissell & Plutchik and the one defined in the Affective Norms for English Words (ANEW).
Exploring the Relationship Between Multi-Modal Emotion Semantics of Music
Exploring the Relationship Between Multi-Modal Emotion Semantics of Music
Ju-Chiang Wang
Recently, plenty of full-length concert videos have become available on video-sharing websites such as YouTube. As each video generally contains multiple songs, natural questions that arise include “what is the set list?” and “when does each song begin and end?” Indeed, many full concert videos on YouTube contain song lists and timecodes contributed by uploaders and viewers. However, newly uploaded content and videos of lesser-known artists typically lack this metadata. Manually labeling such metadata would be labor-intensive, and thus an automated solution is desirable. In this paper, we define a novel research problem, automatic set list segmentation of full concert videos, which calls for techniques in music information retrieval (MIR) such as audio fingerprinting, cover song identification, musical event detection, music alignment, and structural segmentation. Moreover, we propose a greedy approach that sequentially identifies a song from a database of studio versions and simultaneously estimates its probable boundaries in the concert. We conduct preliminary evaluations on a collection of 20 full concerts and 1,152 studio tracks. Our result demonstrates the effectiveness of the proposed greedy algorithm.
Automatic Set List Identification and Song Segmentation of Full-Length Concer...
Automatic Set List Identification and Song Segmentation of Full-Length Concer...
Ju-Chiang Wang
In the music information retrieval (MIR) research, developing a computational model that comprehends the affective content of music signal and utilizes such a model to organize music collections have been an essential topic. Emotion perception in music is in nature subjective. Consequently, building a general emotion recognition system that performs equally well for every user could be insufficient. In contrast, it would be more desirable for one's personal computer/device being able to understand his/her perception of music emotion. In our previous work, we have developed the acoustic emotion Gaussians (AEG) model, which can learn the broad emotion perception of music from general users. Such a general music emotion model, called the background AEG model in this paper, can recognize the perceived emotion of unseen music from a general point of view. In this paper, we go one step further to realize the personalized music emotion modeling by adapting the background AEG model with a limited number of emotion annotations provided by a target user in an online and dynamic fashion. A novel maximum a posteriori (MAP)-based algorithm is proposed to achieve this in a probabilistic framework. We carry out quantitative evaluations on a well-known emotion annotated corpus, MER60, to validate the effectiveness of the proposed method for personalized music emotion recognition.
Personalized Music Emotion Recognition via Model Adaptation
Personalized Music Emotion Recognition via Model Adaptation
Ju-Chiang Wang
Lee Harvey Oswald
Lee Harvey Oswald
Louise Hannecart
Evidencias deivy galvis vasquez.
Evidencias deivy galvis vasquez.
Deyvi Galvis Vasquez
Choreography ;
Choreography ;
hazelbaz
More Related Content
Viewers also liked
One of the most exciting but challenging endeavors in music research is to develop a computational model that comprehends the affective content of music signals and organizes a music collection according to emotion. In this paper, we propose a novel \emph{acoustic emotion Gaussians} (AEG) model that defines a proper generative process of emotion perception in music. As a generative model, AEG permits easy and straightforward interpretations of the model learning processes. To bridge the acoustic feature space and music emotion space, a set of \emph{latent feature classes}, which are learned from data, is introduced to perform the end-to-end semantic mappings between the two spaces. Based on the space of latent feature classes, the AEG model is applicable to both automatic music emotion annotation and emotion-based music retrieval. To gain insights into the AEG model, we also provide illustrations of the model learning process. A comprehensive performance study is conducted to demonstrate the superior accuracy of AEG over its predecessors, using two emotion annotated music corpora MER60 and MTurk. Our results show that the AEG model outperforms the state-of-the-art methods in automatic music emotion annotation. Moreover, for the first time a quantitative evaluation of emotion-based music retrieval is reported.
The Acoustic Emotion Gaussians Model for Emotion-based Music Annotation and R...
The Acoustic Emotion Gaussians Model for Emotion-based Music Annotation and R...
Ju-Chiang Wang
Computational modeling of music emotion has been addressed primarily by two approaches: the categorical approach that categorizes emotions into mood classes and the dimensional approach that regards emotions as numerical values over a few dimensions such as valence and activation. Being two extreme scenarios (discrete/continuous), the two approaches actually share a unified goal of understanding the emotion semantics of music. This paper presents the first computational model that unifies the two semantic modalities under a probabilistic framework, which makes it possible to explore the relationship between them in a computational way. With the proposed framework, mood labels can be mapped into the emotion space in an unsupervised and content-based manner, without any training ground truth annotations for the semantic mapping. Such a function can be applied to automatically generate a semantically structured tag cloud in the emotion space. To demonstrate the effectiveness of the proposed framework, we qualitatively evaluate the mood tag clouds generated from two emotion-annotated corpora, and quantitatively evaluate the accuracy of the categorical-dimensional mapping by comparing the results with those created by psychologists, including the one proposed by Whissell & Plutchik and the one defined in the Affective Norms for English Words (ANEW).
Exploring the Relationship Between Multi-Modal Emotion Semantics of Music
Exploring the Relationship Between Multi-Modal Emotion Semantics of Music
Ju-Chiang Wang
Recently, plenty of full-length concert videos have become available on video-sharing websites such as YouTube. As each video generally contains multiple songs, natural questions that arise include “what is the set list?” and “when does each song begin and end?” Indeed, many full concert videos on YouTube contain song lists and timecodes contributed by uploaders and viewers. However, newly uploaded content and videos of lesser-known artists typically lack this metadata. Manually labeling such metadata would be labor-intensive, and thus an automated solution is desirable. In this paper, we define a novel research problem, automatic set list segmentation of full concert videos, which calls for techniques in music information retrieval (MIR) such as audio fingerprinting, cover song identification, musical event detection, music alignment, and structural segmentation. Moreover, we propose a greedy approach that sequentially identifies a song from a database of studio versions and simultaneously estimates its probable boundaries in the concert. We conduct preliminary evaluations on a collection of 20 full concerts and 1,152 studio tracks. Our result demonstrates the effectiveness of the proposed greedy algorithm.
Automatic Set List Identification and Song Segmentation of Full-Length Concer...
Automatic Set List Identification and Song Segmentation of Full-Length Concer...
Ju-Chiang Wang
In the music information retrieval (MIR) research, developing a computational model that comprehends the affective content of music signal and utilizes such a model to organize music collections have been an essential topic. Emotion perception in music is in nature subjective. Consequently, building a general emotion recognition system that performs equally well for every user could be insufficient. In contrast, it would be more desirable for one's personal computer/device being able to understand his/her perception of music emotion. In our previous work, we have developed the acoustic emotion Gaussians (AEG) model, which can learn the broad emotion perception of music from general users. Such a general music emotion model, called the background AEG model in this paper, can recognize the perceived emotion of unseen music from a general point of view. In this paper, we go one step further to realize the personalized music emotion modeling by adapting the background AEG model with a limited number of emotion annotations provided by a target user in an online and dynamic fashion. A novel maximum a posteriori (MAP)-based algorithm is proposed to achieve this in a probabilistic framework. We carry out quantitative evaluations on a well-known emotion annotated corpus, MER60, to validate the effectiveness of the proposed method for personalized music emotion recognition.
Personalized Music Emotion Recognition via Model Adaptation
Personalized Music Emotion Recognition via Model Adaptation
Ju-Chiang Wang
Lee Harvey Oswald
Lee Harvey Oswald
Louise Hannecart
Evidencias deivy galvis vasquez.
Evidencias deivy galvis vasquez.
Deyvi Galvis Vasquez
Choreography ;
Choreography ;
hazelbaz
Viewers also liked
(7)
The Acoustic Emotion Gaussians Model for Emotion-based Music Annotation and R...
The Acoustic Emotion Gaussians Model for Emotion-based Music Annotation and R...
Exploring the Relationship Between Multi-Modal Emotion Semantics of Music
Exploring the Relationship Between Multi-Modal Emotion Semantics of Music
Automatic Set List Identification and Song Segmentation of Full-Length Concer...
Automatic Set List Identification and Song Segmentation of Full-Length Concer...
Personalized Music Emotion Recognition via Model Adaptation
Personalized Music Emotion Recognition via Model Adaptation
Lee Harvey Oswald
Lee Harvey Oswald
Evidencias deivy galvis vasquez.
Evidencias deivy galvis vasquez.
Choreography ;
Choreography ;
Locations
1.
Locations
2.
In a forest
3.
In town
Download now