The document discusses similarity measures used in music information retrieval systems. It defines music information retrieval as searching for music objects using musical queries. Some applications of MIR discussed are music search and recommendation. The document outlines different methods for calculating musical similarity, including text-based, audio feature-based, semantic concept-based, and multimodal fusion approaches. It concludes by noting future directions for similarity measures in MIR.
The document discusses music information retrieval (MIR) and content-based approaches. It describes how MIR deals with intrinsic characteristics of music like pitch, intensity, and timbre. Key concepts covered include music formats, dimensions of music language, and different types of music users. The document also summarizes Shazam's process for identifying songs from short audio clips, which involves fingerprinting audio files and matching fingerprints to identify songs.
Music Information Retrieval (MIR) is an interdisciplinary field that retrieves information from music. MIR aims to make the world's music accessible by using techniques from computer science, information retrieval, audio processing, musicology and more. MIR applications include music document retrieval, recommender systems, emotion detection and more. Music document retrieval identifies music through metadata like Table of Contents or content-based fingerprinting like Shazam. Emotion detection in music aims to classify emotions in music but faces challenges due to subjective human emotion and requires multi-disciplinary techniques.
Query By Humming - Music Retrieval TechniqueShital Kat
This seminar report summarizes query by humming technology. The basic architecture involves extracting melodic information from a hummed input, transcribing it, and comparing it to melodic contours in a database. Challenges include imperfect user queries and accurately capturing pitches from hums. Popular query by humming applications include Shazam, SoundHound, and Midomi. The report also discusses file formats like WAV and MIDI, and the Parsons code algorithm for representing melodies.
This radio trail targets adults interested in nature by using David Attenborough's formal and relaxed narration over animal sound effects, ambient sounds, and non-diegetic background music in a 37-second clip promoting a BBC Radio 4 documentary. The voiceover is louder than the other sound elements like music and archival footage.
slides presented at a three-hour local AI music course in Taiwan in Oct 2021; part 1: a brief introduction to music information retrieval (+analysis, +generation)
Music Information Retrieval is about retrieving information from music entities.
The slides will introduce the basic concepts of the music language, passing through different kind of music representations and it will end up describing some low level features that are used when dealing with music entities.
Real Time Drum Augmentation with Physical ModelingBen Eyes
This document discusses augmenting acoustic drums with physical modeling to create new sounds and performances. It summarizes previous research that used convolution or spectral processing to digitally process drum sounds. The author then describes his own project that uses a physical model of strings as a VST plugin to process drum sounds from a snare drum and rototoms in real time. An interview with the percussionist discusses the collaborative composition process and how playing with the system required experimenting with extended techniques. The author concludes that future work will involve developing their own drum models and exploring new interfaces like facial recognition to control sound parameters.
The document discusses music information retrieval (MIR) and content-based approaches. It describes how MIR deals with intrinsic characteristics of music like pitch, intensity, and timbre. Key concepts covered include music formats, dimensions of music language, and different types of music users. The document also summarizes Shazam's process for identifying songs from short audio clips, which involves fingerprinting audio files and matching fingerprints to identify songs.
Music Information Retrieval (MIR) is an interdisciplinary field that retrieves information from music. MIR aims to make the world's music accessible by using techniques from computer science, information retrieval, audio processing, musicology and more. MIR applications include music document retrieval, recommender systems, emotion detection and more. Music document retrieval identifies music through metadata like Table of Contents or content-based fingerprinting like Shazam. Emotion detection in music aims to classify emotions in music but faces challenges due to subjective human emotion and requires multi-disciplinary techniques.
Query By Humming - Music Retrieval TechniqueShital Kat
This seminar report summarizes query by humming technology. The basic architecture involves extracting melodic information from a hummed input, transcribing it, and comparing it to melodic contours in a database. Challenges include imperfect user queries and accurately capturing pitches from hums. Popular query by humming applications include Shazam, SoundHound, and Midomi. The report also discusses file formats like WAV and MIDI, and the Parsons code algorithm for representing melodies.
This radio trail targets adults interested in nature by using David Attenborough's formal and relaxed narration over animal sound effects, ambient sounds, and non-diegetic background music in a 37-second clip promoting a BBC Radio 4 documentary. The voiceover is louder than the other sound elements like music and archival footage.
slides presented at a three-hour local AI music course in Taiwan in Oct 2021; part 1: a brief introduction to music information retrieval (+analysis, +generation)
Music Information Retrieval is about retrieving information from music entities.
The slides will introduce the basic concepts of the music language, passing through different kind of music representations and it will end up describing some low level features that are used when dealing with music entities.
Real Time Drum Augmentation with Physical ModelingBen Eyes
This document discusses augmenting acoustic drums with physical modeling to create new sounds and performances. It summarizes previous research that used convolution or spectral processing to digitally process drum sounds. The author then describes his own project that uses a physical model of strings as a VST plugin to process drum sounds from a snare drum and rototoms in real time. An interview with the percussionist discusses the collaborative composition process and how playing with the system required experimenting with extended techniques. The author concludes that future work will involve developing their own drum models and exploring new interfaces like facial recognition to control sound parameters.
Mehfil : Song Recommendation System Using Sentiment DetectedIRJET Journal
This document describes a song recommendation system called Mehfil that uses sentiment analysis to recommend songs based on a user's detected mood. It has three main modules: sentiment analysis using facial recognition and emotion detection on images via a deep learning model, music recommendation by classifying songs based on audio features and assigning mood labels, and integration using the Spotify API to generate personalized playlists based on the detected sentiment. The system aims to make creating mood-based playlists easier by analyzing a user's facial expression in real-time with their webcam to infer their mood and select an appropriate playlist of songs. It discusses the technologies used like Haar Cascade for face detection, MobileNetV2 for sentiment classification, and the Spotify API for music metadata and
The document discusses a survey on music recognition searching applications, finding that while users appreciate the convenience of such apps, limitations include over half refusing to help enlarge databases and the top reason for not using an app being required fees, suggesting challenges in building large databases and monetizing without compromising user experience.
Knn a machine learning approach to recognize a musical instrumentIJARIIT
An outline is provided of a proposed system to recognize musical instruments using machine learning techniques. The system first extracts features from audio files using the MIR toolbox in Matlab. It then uses a hybrid feature selection method and vector quantization to identify instruments. Specifically, the key audio descriptors are selected and feature vectors are generated and matched to standard vectors to classify the instrument. The k-nearest neighbors algorithm is used for classification. Preliminary results show the system can accurately recognize instruments based on extracted acoustic features.
This document discusses music data mining. It provides an overview of music data mining tasks and approaches. Some key tasks discussed are music similarity search, clustering, and music sequence mining. Music similarity search involves finding music files similar to a given file based on features and similarity measures. Clustering divides music data into groups of similar objects based on pairwise similarity. Music sequence mining finds frequent patterns that appear in the ordered sequences of elements in music data instances.
Spotify provides personalized music recommendations to over 100 million active users based on their listening history and the listening history of similar users. It utilizes various recommendation approaches, including collaborative filtering using latent factor models to create lower-dimensional representations of users and songs. Spotify also uses natural language processing models on playlist data and deep learning on audio features to power recommendations. Personalizing music at Spotify's massive scale across 30 million tracks presents challenges around cold starts, repeated consumption, and measuring recommendation quality.
survey on Hybrid recommendation mechanism to get effective ranking results fo...Suraj Ligade
These days clients are having exclusive
requirements towards advancements, they need to hunt tunes
in such circumstances where they are not ready to recall tunes
title or melody related points of interest. Recovery of music or
melodies substance is one of the hardest errands and testing
work in the field of Music Information Retrieval (MIR). There
are different looking techniques created and executed, yet
these seeking strategies are no more ready to inquiry tunes
which required by the clients and confronting different issues
like programmed playlist creation, music suggestion or music
pursuit are connected issues. In past framework client seek
the tune with the assistance of tune title, craftsman name and
whatever other related points of interest so this strategy is
exceptionally tedious. To beat this issue singing so as to look
tune or murmuring a segment of it is the most regular
approach to seek the tune. This hunt strategy is the most
helpful when client don't have entry to sound gadget or client
can't review the traits of the tune such as tune title, name of
craftsman, name of collection. In proposed framework client
have not stress over recalling the tune data and this technique
is not tedious. In this strategy we utilize the data from a
client's hunt history and in addition the normal properties of
client's comparative foundations. Cross breed proposal
component utilizes the substance construct recovery
framework situated in light of utilization of the sound data
such as tone, pitch, mood. This component used to get exact
result to the client. The more imperative idea is clients ready
to work their gadgets without manual information orders by
hand. It is simple and basic system to perform music look.
J-P. Fauconnier, J. Roumier. Musonto - A Semantic Search Engine Dedicated to ...MusicNet
Jean-Philippe Fauconnier (Université Catholique de Louvain, Belgium) and Joseph Roumier (CETIC, Belgium).
Music Linked Data Workshop, 12 May 2011, JISC, London.
Music Recommendation System using Euclidean, Cosine Similarity, Correlation D...IRJET Journal
This document describes a music recommendation system that uses various distance and similarity algorithms like Euclidean distance, cosine similarity, and correlation distance. It integrates these recommendation models with a user-friendly web interface built using the Flask framework. The system is evaluated based on how accurately and relevantly it can recommend songs. In summary, the project created an effective music recommendation system that utilizes different similarity metrics and Flask for web integration.
Internet radio provides new opportunities for niche audiences and customized playlists but also new challenges for radio networks and record companies. It reduces the power of radio networks as gatekeepers and enables low-cost, tailored streaming for very small music tastes. However, record companies want to maintain control and royalty streams and limit streaming functionality to mimic traditional radio. New forms of interaction between listeners, artists and stations are also emerging using tools like Twitter.
This document discusses the history and evolution of music discovery from phonographs to modern technologies like smartphones and MP3 players. It outlines several recommendation technologies used for music discovery, including collaborative filtering based on user behavior, annotations from users, and content analysis. The document also discusses issues with recommendation systems like relevance, variety, scalability, and privacy. It provides an overview of resources that can be used to build music recommendation systems, including MusicBrainz, Last.fm, and Echo Nest for content analysis and recommendations.
This document provides an overview of a dissertation on Emofy, a classical music recommender system. The summary includes:
- Emofy is a music recommender system that recommends classical Indian music based on the user's mood by classifying moods and associating different ragas and genres with different moods.
- The dissertation discusses collecting and labeling a dataset of classical music, extracting features to classify mood, and using machine learning algorithms like random forests to achieve over 90% accuracy in mood classification.
- The recommended system uses mood classification to map users to appropriate ragas and playlists of classical music tracks on Spotify aimed at therapeutic applications.
This document discusses multimodal music mood classification using audio, lyrics, and social tags. It outlines research questions around using social tags to develop mood taxonomies, determining the most useful lyric features for classification, and whether lyrics or audio perform better. The author finds that combining lyrics and audio improves classification effectiveness and efficiency. Contributions include establishing empirically derived mood categories and the largest dataset combining audio, lyrics and tags. Future work may explore multimedia and multimodal approaches.
Echoes, Whispers, and Footsteps from the Conflux of Sonic Interaction Design ...Eoin Brazil
This document discusses selecting and classifying sounds for interaction design and public spaces. It summarizes previous studies that explored mapping human activity to actions using sound, considering concurrent auditory presentations, and using metaphors from listener responses. The document recommends verifying sound mappings through ratings and comparisons, and visually depicting the human perceptual space. It also describes projects that sonified exhibitions and public spaces like train stations to enhance experiences.
This document summarizes a talk on integrating listening into music collection interfaces. It discusses how current commercial interfaces rely mainly on text searches and recommendations. It argues that listening should be integrated as it allows for faster and more effective navigation of music collections. The document then reviews several academic projects that have incorporated listening into interfaces, including passive listening interfaces like Mused, 3D browsing interfaces like SoundTorch, and landscapes like Neptune and Sonixplorer that combine visualization and audio.
This document provides an introduction to audio content analysis (ACA). It discusses how ACA aims to automatically extract content information from audio signals to enable content-driven services. Example applications of ACA include speech recognition, music transcription, and noise pollution monitoring. The document also outlines related fields like music information retrieval and computational auditory scene analysis, and notes that ACA has historically progressed from mechanical devices to data-driven machine learning systems and now deep neural networks.
Colloque IMT -04/04/2019- L'IA au cœur des mutations industrielles - "Machine...I MT
Colloque IMT -04/04/2019- L'IA au cœur des mutations industrielles - "Machine Listening: L'intelligence artificielle pour les sons et la musique". Présentation par Gaël Richard
Applications of AI and NLP to advance Music Recommendations on Voice AssistantsAI Publications
This paper builds on the popular use case of music requests on voice assistants like Siri, Google Assistant, Alexa, and others and explores the different AI and NLP techniques. The paper particularly focuses on how each of these techniques can be applied in the context of musical recommendations and experiences on voice assistants. It enumerates specific problems in the space of music recommendations and illustrates how specific techniques like multi-armed bandits can be applied.
Case-based Sequential Ordering of Songs for Playlist RecommendationUri Levanon
A model of contiguous sequential patterns of songs, in order to acquire the knowledge about which songs sound well together in a playlist.
A research paper that I have presented in September 2006 at the European Conference on Case Based Reasoning (ECCBR '06).
By Claudio Baccigalupo and Enric Plaza
This document discusses analyzing music data extracted from Spotify using various data science techniques. It describes collecting audio feature data for Billboard top songs from 1958-2017 and Grammy winning albums from 1959-2018. Exploratory analysis was performed on song attributes like duration, tempo and loudness. Principal component analysis and k-means clustering were used to group songs based on similarities in audio features like energy, danceability and acousticness. The analysis revealed patterns in attributes of popular music over time that could aid music recommendation systems.
This document provides an overview of natural language processing (NLP). It defines NLP, discusses common NLP tasks such as part-of-speech tagging and machine translation, and explains why NLP is challenging due to various ambiguities in natural language. The document also briefly discusses related fields like linguistics, machine learning, and information retrieval, and concludes by noting that it only covers an introduction to NLP and does not discuss solutions or the current state of the field.
This document provides an overview of natural language processing (NLP). It begins with examples of NLP applications like translation and question answering. It then discusses the backgrounds in artificial intelligence, linguistics, and the web. The document outlines several common NLP tasks like part-of-speech tagging, named-entity recognition, word sense disambiguation, and parsing. It also discusses challenges like ambiguity in natural language. The document concludes with a discussion of why NLP is difficult due to ambiguity at both the linguistic and acoustic levels.
More Related Content
Similar to CSTalks - Music Information Retrieval - 23 Feb
Mehfil : Song Recommendation System Using Sentiment DetectedIRJET Journal
This document describes a song recommendation system called Mehfil that uses sentiment analysis to recommend songs based on a user's detected mood. It has three main modules: sentiment analysis using facial recognition and emotion detection on images via a deep learning model, music recommendation by classifying songs based on audio features and assigning mood labels, and integration using the Spotify API to generate personalized playlists based on the detected sentiment. The system aims to make creating mood-based playlists easier by analyzing a user's facial expression in real-time with their webcam to infer their mood and select an appropriate playlist of songs. It discusses the technologies used like Haar Cascade for face detection, MobileNetV2 for sentiment classification, and the Spotify API for music metadata and
The document discusses a survey on music recognition searching applications, finding that while users appreciate the convenience of such apps, limitations include over half refusing to help enlarge databases and the top reason for not using an app being required fees, suggesting challenges in building large databases and monetizing without compromising user experience.
Knn a machine learning approach to recognize a musical instrumentIJARIIT
An outline is provided of a proposed system to recognize musical instruments using machine learning techniques. The system first extracts features from audio files using the MIR toolbox in Matlab. It then uses a hybrid feature selection method and vector quantization to identify instruments. Specifically, the key audio descriptors are selected and feature vectors are generated and matched to standard vectors to classify the instrument. The k-nearest neighbors algorithm is used for classification. Preliminary results show the system can accurately recognize instruments based on extracted acoustic features.
This document discusses music data mining. It provides an overview of music data mining tasks and approaches. Some key tasks discussed are music similarity search, clustering, and music sequence mining. Music similarity search involves finding music files similar to a given file based on features and similarity measures. Clustering divides music data into groups of similar objects based on pairwise similarity. Music sequence mining finds frequent patterns that appear in the ordered sequences of elements in music data instances.
Spotify provides personalized music recommendations to over 100 million active users based on their listening history and the listening history of similar users. It utilizes various recommendation approaches, including collaborative filtering using latent factor models to create lower-dimensional representations of users and songs. Spotify also uses natural language processing models on playlist data and deep learning on audio features to power recommendations. Personalizing music at Spotify's massive scale across 30 million tracks presents challenges around cold starts, repeated consumption, and measuring recommendation quality.
survey on Hybrid recommendation mechanism to get effective ranking results fo...Suraj Ligade
These days clients are having exclusive
requirements towards advancements, they need to hunt tunes
in such circumstances where they are not ready to recall tunes
title or melody related points of interest. Recovery of music or
melodies substance is one of the hardest errands and testing
work in the field of Music Information Retrieval (MIR). There
are different looking techniques created and executed, yet
these seeking strategies are no more ready to inquiry tunes
which required by the clients and confronting different issues
like programmed playlist creation, music suggestion or music
pursuit are connected issues. In past framework client seek
the tune with the assistance of tune title, craftsman name and
whatever other related points of interest so this strategy is
exceptionally tedious. To beat this issue singing so as to look
tune or murmuring a segment of it is the most regular
approach to seek the tune. This hunt strategy is the most
helpful when client don't have entry to sound gadget or client
can't review the traits of the tune such as tune title, name of
craftsman, name of collection. In proposed framework client
have not stress over recalling the tune data and this technique
is not tedious. In this strategy we utilize the data from a
client's hunt history and in addition the normal properties of
client's comparative foundations. Cross breed proposal
component utilizes the substance construct recovery
framework situated in light of utilization of the sound data
such as tone, pitch, mood. This component used to get exact
result to the client. The more imperative idea is clients ready
to work their gadgets without manual information orders by
hand. It is simple and basic system to perform music look.
J-P. Fauconnier, J. Roumier. Musonto - A Semantic Search Engine Dedicated to ...MusicNet
Jean-Philippe Fauconnier (Université Catholique de Louvain, Belgium) and Joseph Roumier (CETIC, Belgium).
Music Linked Data Workshop, 12 May 2011, JISC, London.
Music Recommendation System using Euclidean, Cosine Similarity, Correlation D...IRJET Journal
This document describes a music recommendation system that uses various distance and similarity algorithms like Euclidean distance, cosine similarity, and correlation distance. It integrates these recommendation models with a user-friendly web interface built using the Flask framework. The system is evaluated based on how accurately and relevantly it can recommend songs. In summary, the project created an effective music recommendation system that utilizes different similarity metrics and Flask for web integration.
Internet radio provides new opportunities for niche audiences and customized playlists but also new challenges for radio networks and record companies. It reduces the power of radio networks as gatekeepers and enables low-cost, tailored streaming for very small music tastes. However, record companies want to maintain control and royalty streams and limit streaming functionality to mimic traditional radio. New forms of interaction between listeners, artists and stations are also emerging using tools like Twitter.
This document discusses the history and evolution of music discovery from phonographs to modern technologies like smartphones and MP3 players. It outlines several recommendation technologies used for music discovery, including collaborative filtering based on user behavior, annotations from users, and content analysis. The document also discusses issues with recommendation systems like relevance, variety, scalability, and privacy. It provides an overview of resources that can be used to build music recommendation systems, including MusicBrainz, Last.fm, and Echo Nest for content analysis and recommendations.
This document provides an overview of a dissertation on Emofy, a classical music recommender system. The summary includes:
- Emofy is a music recommender system that recommends classical Indian music based on the user's mood by classifying moods and associating different ragas and genres with different moods.
- The dissertation discusses collecting and labeling a dataset of classical music, extracting features to classify mood, and using machine learning algorithms like random forests to achieve over 90% accuracy in mood classification.
- The recommended system uses mood classification to map users to appropriate ragas and playlists of classical music tracks on Spotify aimed at therapeutic applications.
This document discusses multimodal music mood classification using audio, lyrics, and social tags. It outlines research questions around using social tags to develop mood taxonomies, determining the most useful lyric features for classification, and whether lyrics or audio perform better. The author finds that combining lyrics and audio improves classification effectiveness and efficiency. Contributions include establishing empirically derived mood categories and the largest dataset combining audio, lyrics and tags. Future work may explore multimedia and multimodal approaches.
Echoes, Whispers, and Footsteps from the Conflux of Sonic Interaction Design ...Eoin Brazil
This document discusses selecting and classifying sounds for interaction design and public spaces. It summarizes previous studies that explored mapping human activity to actions using sound, considering concurrent auditory presentations, and using metaphors from listener responses. The document recommends verifying sound mappings through ratings and comparisons, and visually depicting the human perceptual space. It also describes projects that sonified exhibitions and public spaces like train stations to enhance experiences.
This document summarizes a talk on integrating listening into music collection interfaces. It discusses how current commercial interfaces rely mainly on text searches and recommendations. It argues that listening should be integrated as it allows for faster and more effective navigation of music collections. The document then reviews several academic projects that have incorporated listening into interfaces, including passive listening interfaces like Mused, 3D browsing interfaces like SoundTorch, and landscapes like Neptune and Sonixplorer that combine visualization and audio.
This document provides an introduction to audio content analysis (ACA). It discusses how ACA aims to automatically extract content information from audio signals to enable content-driven services. Example applications of ACA include speech recognition, music transcription, and noise pollution monitoring. The document also outlines related fields like music information retrieval and computational auditory scene analysis, and notes that ACA has historically progressed from mechanical devices to data-driven machine learning systems and now deep neural networks.
Colloque IMT -04/04/2019- L'IA au cœur des mutations industrielles - "Machine...I MT
Colloque IMT -04/04/2019- L'IA au cœur des mutations industrielles - "Machine Listening: L'intelligence artificielle pour les sons et la musique". Présentation par Gaël Richard
Applications of AI and NLP to advance Music Recommendations on Voice AssistantsAI Publications
This paper builds on the popular use case of music requests on voice assistants like Siri, Google Assistant, Alexa, and others and explores the different AI and NLP techniques. The paper particularly focuses on how each of these techniques can be applied in the context of musical recommendations and experiences on voice assistants. It enumerates specific problems in the space of music recommendations and illustrates how specific techniques like multi-armed bandits can be applied.
Case-based Sequential Ordering of Songs for Playlist RecommendationUri Levanon
A model of contiguous sequential patterns of songs, in order to acquire the knowledge about which songs sound well together in a playlist.
A research paper that I have presented in September 2006 at the European Conference on Case Based Reasoning (ECCBR '06).
By Claudio Baccigalupo and Enric Plaza
This document discusses analyzing music data extracted from Spotify using various data science techniques. It describes collecting audio feature data for Billboard top songs from 1958-2017 and Grammy winning albums from 1959-2018. Exploratory analysis was performed on song attributes like duration, tempo and loudness. Principal component analysis and k-means clustering were used to group songs based on similarities in audio features like energy, danceability and acousticness. The analysis revealed patterns in attributes of popular music over time that could aid music recommendation systems.
Similar to CSTalks - Music Information Retrieval - 23 Feb (20)
This document provides an overview of natural language processing (NLP). It defines NLP, discusses common NLP tasks such as part-of-speech tagging and machine translation, and explains why NLP is challenging due to various ambiguities in natural language. The document also briefly discusses related fields like linguistics, machine learning, and information retrieval, and concludes by noting that it only covers an introduction to NLP and does not discuss solutions or the current state of the field.
This document provides an overview of natural language processing (NLP). It begins with examples of NLP applications like translation and question answering. It then discusses the backgrounds in artificial intelligence, linguistics, and the web. The document outlines several common NLP tasks like part-of-speech tagging, named-entity recognition, word sense disambiguation, and parsing. It also discusses challenges like ambiguity in natural language. The document concludes with a discussion of why NLP is difficult due to ambiguity at both the linguistic and acoustic levels.
CSTalks-Sensor-Rich Mobile Video Indexing and Search-17Augcstalks
The document describes a system called GeoVid for managing and searching geo-referenced video content. GeoVid uses sensors in mobile devices to automatically collect location, direction, and other contextual data while recording video. This sensor data is associated with video frames to model the viewable scene. GeoVid then indexes and stores this sensor-enriched video data to support spatiotemporal search and retrieval of relevant video segments by location, trajectory, or other metadata. The system aims to improve searchability of user-generated video through automated geo-tagging without requiring manual user input.
9/24/2011 28
Approach – Search (2)
• Index FOV models using R-tree
This document discusses visualizing software behavior through system traces. It describes how system traces can contain huge amounts of information about software execution but are difficult for humans to understand. The author introduces a tool called WinResMon that records Windows system events and traces. Various techniques for visualizing these traces are presented, including dot plots, histograms, and barcodes, which help identify patterns, anomalies, and differences between traces to better understand software behavior. Examples analyzing file copying, a software build, and idle machines are provided.
The document discusses polymorphic heterogeneous multi-core systems as a solution to limitations in instruction-level parallelism (ILP) and thread-level parallelism (TLP) approaches for improving single-core performance. It proposes an architecture with cores that can dynamically reconfigure their internal structure and collaborate to best match software requirements. The cores are connected to a reconfigurable fabric that implements custom instructions to further speed up programs. Experimental results show this approach achieves speedups and better load balancing compared to homogeneous multi-core systems. Future work is needed to study overhead and implement dynamic scheduling.
The document proposes a framework for recommendations based on analyzing relationships between users, items, tags, and ratings (quaternary relationships). It models these relationships using a 4-order tensor and applies Higher-Order Singular Value Decomposition (HOSVD) to reveal latent semantic associations. This allows generating recommendations for users, items, tags, and predicting ratings. Experimental results on a movie dataset show the proposed quaternary approach outperforms methods using only ternary relationships.
CSTalks - Object detection and tracking - 25th Maycstalks
Object detection is a fundamental step in most of the video analysis applications. There are many research challenges involved in automatic object detection, depending on different scenarios. The most prevalent application of object detection is in the field of multimedia surveillance. In this talk we will discuss the common problems in the object detection in a surveillance video. Further, we will discuss the Gaussian Mixture Model (GMM) based object detection method. While object detection is the basic step of video analysis, higher level semantic interpretation of the scene requires trajectory information. Most of the suspicious event detection methods use tracking as the basic building block. In the second part of the talk, we will discuss particle filter based method of object tracking. To summarize, the aim of the talk is two-fold: (1) Discuss common problems in object detection and tracking (2) Hands on experience of how to use classical methods of GMM and particle filtering in problem solving.
CSTalks - The Multicore Midlife Crisis - 30 Marcstalks
The document discusses the multicore midlife crisis as processors move to multiple cores to cope with Moore's Law. As core counts increase, the memory bandwidth does not scale accordingly, creating a memory wall problem. Solutions proposed include increasing cache sizes, improving memory speeds, and better caching techniques. Future multicore designs may focus more on heterogeneous cores tailored for different workloads rather than increasing core counts uniformly. Research challenges include coping with heterogeneity, improving data locality given slow memory speeds, and software techniques to help address issues like cache coherence.
This document summarizes machine learning concepts including supervised and unsupervised learning techniques. It discusses fundamental questions in machine learning like how to build systems that improve with experience. Key problems covered include classification, regression, and clustering. Challenges like overfitting, model complexity, and optimization techniques like gradient descent are also summarized. Open problems in machine learning like transferring learned knowledge between tasks and preserving privacy in data mining are mentioned.
CSTalks - Real movie recommendation - 9 Marcstalks
This document proposes a new approach to movie recommendation that considers temporal dynamics and local user ratings. The current best approach is collaborative filtering with temporal dynamics, but this new approach clusters users based on their individual monitoring and behavior over time. It also clusters movies based on their global and dynamic class ratings. The model would monitor users, user-user patterns, user-movie patterns, and movie-movie patterns over time to update recommendations and predictions. This is aimed to provide more accurate recommendations by considering how user preferences can change over time.
The document discusses career paths after obtaining a PhD, including academia, industry, and other options. It provides an outline of the typical academic career path, from postdoc to obtaining assistant professor positions. Interview tips are given for applying to assistant professor roles. Pursuing industry careers or entrepreneurship are also addressed. The conclusion emphasizes the importance of publishing, networking, and maintaining high quality research throughout the PhD and beyond.
This document discusses the past, present, and future of peer-to-peer (P2P) networks. It describes how P2P networks emerged to help alleviate load on servers and make use of the interconnectivity between nodes. Popular early P2P file sharing applications like Napster, Gnutella, and BitTorrent are discussed. While P2P file sharing declined due to legal issues, P2P protocols now see growing use for media streaming, communication, and within data centers. The document predicts P2P will continue evolving with technologies like IPv6 and play an increasing role in areas like social networks and peer production.
The document summarizes a presentation on Named Data Networking (NDN) given by Mostafa Rezazad. It discusses the motivation for NDN, which is to make data and services rather than locations the primary objects on the network. This allows for benefits like redundancy elimination, easier mobility, and more inherent security. An overview is provided of NDN's packet types, node structure, name structure, and routing approach.
The document summarizes a presentation on modeling and verifying timed concurrent systems. It introduces a simple coffee vending machine model as an example and specifies properties of it using temporal logic. Timed automata are discussed as a way to model systems with timing parameters. The talk will cover extending the coffee machine model to make it parametric and synthesizing parameters to satisfy given properties.
This document discusses research in GPU computing. It provides an introduction to GPU computing, including how GPUs were originally for graphics processing but are now used more broadly through frameworks like CUDA and OpenCL. It discusses advantages of GPUs like their large number of cores compared to CPUs. Open problems in the field are also outlined, such as developing new data structures and algorithms suitable for massive parallelism. The document suggests GPU computing will continue growing in importance as computing moves towards more highly multithreaded architectures.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
5. Definition
Music Information Retrieval
Is the process of searching for, and finding, music objects, or
part of objects, via a query framed musically and/or in musical
terms.
Music objects: Recordings (wav, mp3, etc.), scores, parts, etc.
Musically framed query: Singing, humming, keyboad, notation-based,
MIDI files, sound files, etc.
Music terms: Genre, style, tempo, bibliography, etc.
Applications
Music search, recommendation, identification ,etc.
5
7. Applications -- Search
Text-based Music Search
Compare a textual query with the metadata
Is adopted by most existing systems.
Examples: Last.fm, Musicovery, …
7
8. Applications -- Search
Content-based Music Search
Compare audio query with audio content
Query-by-humming/singing/recording: midomi
8
9. Applications -- Search
Content-based Music Search
Compare rhythm tapped with audio content
Query-by-tapping: SongTapper
9
10. Applications -- Search
User’s information need Online Offline
(intention) : Query
Explicit Query: text, audio, etc. Intention Gap
Query
Formation
Similarity measure: Descriptors Documents
Query Music documents in the Match
Descriptor
Extraction
Database Indexing
Semantic Gap Index Descriptors
Ranking: relevant documents by
domain specific criterions (no. of Ranking
hits ). Ranked List
Presentation
Results
10
12. Applications – Recommendation
Collaborative-Filtering-based Recommendation
Last.fm: what you (and others ) listen to and like,
Amazon: customers who shopped for … also shopped for …
12
13. Applications – Recommendation
Collaborative-Filtering-based Recommendation
Last.fm: what you (and others ) listen to and like,
Amazon: customers who shopped for … also shopped for …
Example:
Users: A, B, and C
Music: 1, 2, …, 8
Small similarity Large similarity
C A B
Similarity
4
Similarity 1 Measure
1
Measure
2 2
6 3 3
8 4 5
Recommend to user A
13
14. Applications -- Recommendation
Audio Content-based Recommendation
Recommend songs which have similar audio content to the
songs that you like.
Pandora:
Music database Music Experts
User
Listen
Instrument:
Instrument: Similarity Instrument:
Vocal:
Vocal: Vocal:
Structure:
Measure
Structure: …Structure:
… …
400 Attributes/song
Recommendations
14
15. Applications -- Recommendation
User’s information need Online / Offline Offline
(intention): User Profile
Implicit user profiles: ratings, Intention Gap
Profile
Capture
listening history, etc.
Descriptors Documents
Similarity measure:
Descriptor
User profiles Music Match Extraction
documents/other user profiles Semantic Gap Index
Indexing
Descriptors
Ranking: relevant documents Ranking
by some domain specific
Ranked List
criterions (no. of hits).
Presentation
Results
15
16. Similarity Measure
one of the most fundamental concepts in MIR
Online / Offline Offline
Closely related to
User Profile
/query
What information music Profile/query
Intention Gap
Capture
contains.
How this information is
Descriptors Documents
represented. Match
Descriptor
Extraction
How to match between themSemantic Gap
Indexing
Index Descriptors
Ranking
Ranked List
Presentation
Results
16
17. Music Information Plane
Similarity can be measure
from different aspects.
Song1: New favorite -
Alison Krauss
Song2: She is Beautiful -
Andrew W.K.
Song1 Song2
Female Male
Dissimilar Gentle Aggressive
Slow fast
Guitar
Similar Tempo: ~162 BPM
(Beat Per Minute)
* O. C. Herrada. Music recommendation and discovery in
Music the long tail. PhD thsis. 2008. 17
19. Similarity Measure Methods
Text-based Method: Okapi BM-25 Ranking
Given: queryQ, containing keywords q1, …, qn, music documents: bag of words.
BM25 ranking function can be formulated as:
f(qi, D) is qi’s term frequency (tf) in document D.
|D| is the length of document D in words.
avgdl is the average document length in the collection.
k1 and b are free parameters, usually set as k1=2.0 and b=0.75.
IDF(qi) is the inverse document frequency (idf), calculated as:
. The query term appears in this document frequently. f (qi, D)
. And it doesn’t appear in other document. IDF
19
20. Similarity Measure Methods
Text-based Method:
Pros
Simple & efficient
Cons
Affected by noisy/wrong texts
Songs with no text cannot be retrieved
Require high-level domain knowledge to create good metadata
“Text retrieval on audio metadata” not pure music retrieval
20
24. Existing Works
Audio Feature-based Method
Audio Feature extraction Distribution modeling Model Comparison
Use low-level feature directly
Pitch, loudness, MFCC (Blum et al.[3], 1999)
Histogram of MFCC (Foote[4], 1997)
Spectrum, rhythm, chord changesingleVector (Tzanetakis [5], 2002)
Low-level features higher-level features.
Cluster MFCC=>model comparison (Aucouturier[6], 2002)
MFCC => Gaussian Mixture Models => model comparison
MFCC =>“anchor space”, compare probability models (Berenzweig et al.
[7], 2003)
24
25. Similarity Measure Methods
Audio Feature-based Method
Audio Feature extraction Distribution modeling Model Comparison
Euclidean /Cosine distance (uniform-length feature vectors)
Distance between two probability distributions
Kullback-Leibler divergence (KL Distance) / relative entropy
No closed form for Gaussian Distributions
Centroid distance: Euclidean distance between the overall means;
Sampling based method: compute the likelihood of one model given points
sampled from another; very computationally expensive;
Earth-Mover’s distance
Berenzweig, A., Logan, B., Ellis, D. P., and Whitman, B. P., A Large-Scale Evaluation of Acoustic and Subjective Music-Similarity
Measures. Computer Music Journal. 28, 2004, 25
26. Similarity Measure Methods
Audio Feature-based Method
Pros
Can deal with new songs with no or few texts.
Save human labors from annotating each song manually
Cons
Time complexity is relatively high.
Features ≠ audio piece: Two songs with very similar features may sounds
very different.
The average performance is reaching the glass ceiling of around 65% in
accuracy.
26
28. Similarity Measure Methods
Semantic Concept-based Method
Nature of user queries
Far beyond of bibliographic text and audio search
Semantically-rich
Syntactically- undetermined
e.g.: “Find me a classical and happy song”, or “Find me a song to relax”
“Find me some songs for parties/ weddings/ in churches” …
Collaborative(social) tagging is very popular on Web 2.0.
Users annotate their feelings or opinions to the music. Tags,
comments, etc.
28
29. Similarity Measure Methods
Semantic Concept-based Method
Tags VS user queries (Last.fm)
Tag Type Frequency Multi-tag search queries
Genre 68% 51%
Locale 12% 7%
Mood 5% 4%
Opinion 4% 2%
Instrumentation 4% 5%
Style 3% 26%
. Paul Lamere. Social tagging and music information
retrieval. Journal of New Music Research. 2008.
. Klaas Bosteels, Elias Pampalk, and Etienne E. Kerre. Music
retrieval based on social tags: a case study. ISMIR, 2008. 29
30. Similarity Measure Methods
Vocabulary: classical, jazz, … piano, violin, …, female, male, …
…
Model Model … Model
… Probability vector
Song1 Similarity
…
Song2
30
32. Similarity Measure Methods
Multimodal Method
Information keeps growing.
One of the most important ongoing trends:
Metadata
Audio Semantic
Content Concept
Users are important.
32
33. Similarity Measure Methods
Multimodal Method
Document Vectors
Customization
Fuzzy Music Semantic
Vector
B. Zhang, J. Shen, Q. Xiang, and Y. Wang. CompositMap: a novel framework for music similarity measure. ACM Multimedia,
2009. 33
35. Conclusion and Future Directions
What makes MIR (and the similarity measure) so
tricky?
Music information is
Multimodal: audio, metadata, social , …
Multicultural: e.g., modern art, Indian ragas, …
Multirepresentational: audio, MIDI, score, …
Multifaceted: melody, tempo, beat, …
…
Similarity can be measured from different aspects.
35
36. Conclusions and Future Directions
What do users really want?
Intention Gap
User interactions with the system.
Learn a good user preference modeling
What kind of music features can really capture this need?
Content –Tags Semantic Gap
Leverage more social data? Comments, ratings, groups, playlist, other
user created information, …
How to fuse multiple information effectively?
Identify the relevant/discriminative information aspects
Fusion Methods
36
38. References
[2] F[1] O. C. Herrada. Music recommendation and discovery in the long
tail. PhD thsis. 2008.
. Pachet. Knowledge management and musical metadata. Encyclopedia of
Knowledge Management. Idea Group, 2005.
[3] T. L. Blum, D. F. Keislar, J. A. Wheaton, and E. H. Wold. Method and article
of manufacture for content-based analysis, storage, retrieval, and
segmentation of audio information. U.S. Patent 5, 918, 223.
[4] J. T. Foote. Content-based retrieval of music and audio. SPIE, 1997.
[5] G. Tzanetakis. Manipulation, analysis, and retrieval system for audio
signals. PhD thsis, 2002.
[6] J. J. Aucouturier and F. Pachet. Music similarity measure: What’s the use?
International Symposium on Music information retrieval. 2002.
[7] A. Berenzweig, D. P. W. Ellis and S. Lawrence. Anchor space for
classification and similarity measurement for music. ICME 2003.
38
39. References
[8] B. Zhang, J. Shen, Q. Xiang and Y. Wang. CompositeMap: a
novel framework for music similarity measure. SIGIR, 2009.
[9] B. Whiteman and S. Lawrence. Inferring descriptions and
similarity for music from community metadata. International
computer music conference. 2002.
[10] M. Schedl, T. Pohle, P. Knees and G. Widmer. Assigning
and visualizing music genre by web-based co-occurrence
analysis. ISMIR 2006.
[11] B. Whitman and Paris Smaragdis. Combining musical and
cultural features for intelligent style detection. ISMIR 2002.
[12] L. Chen, P. Wright, and W. Nejdl. Improving music genre
classification using collaborative tagging data. WSDM, 2009.
[13] Benedikt Raes. Automatic generation of music metadata.
ISMIR, 2009.
39