Computational Melodic Analysis of Indian Art MusicSankalp Gulati
This document summarizes research on computational melodic analysis of Indian art music. Key areas discussed include tonic identification, predominant melody estimation, motif discovery, melodic similarity, melodic pattern networks, raga recognition using melodic patterns, and resources like datasets and demonstrations of the techniques.
Phrase-based Rāga Recognition Using Vector Space ModelingSankalp Gulati
This document describes an approach for automatic raga recognition in Indian art music using phrase-based vector space modeling. It involves discovering melodic patterns from a Carnatic music collection through intra-recording and inter-recording analysis. The patterns are then clustered into communities based on their network of similarities. Features are extracted from the pattern-recording relationships using term frequency-inverse document frequency weighting. These features are used to build a feature matrix for raga recognition.
Mining Melodic Patterns in Large Audio Collections of Indian Art MusicSankalp Gulati
More info: http://mtg.upf.edu/node/3108
Abstract: Discovery of repeating structures in music is fundamental to its analysis, understanding and interpretation. We present a data-driven approach for the discovery of short-time melodic patterns in large collections of Indian art music. The approach first discovers melodic patterns within an audio recording and subsequently searches for their repetitions in the entire music collection. We compute similarity between melodic patterns using dynamic time warping (DTW). Furthermore, we investigate four different variants of the DTW cost function for rank refinement of the obtained results. The music collection used in this study comprises 1,764 audio recordings with a total duration of 365 hours. Over 13 trillion DTW distance computations are done for the entire dataset. Due to the computational complexity of the task, different lower bounding and early abandoning techniques are applied during DTW distance computation. An evaluation based on expert feedback on a subset of the dataset shows that the discovered melodic patterns are musically relevant. Several musically interesting relationships are discovered, yielding further scope for establishing novel similarity measures based on melodic patterns. The discovered melodic patterns can further be used in challenging computational tasks such as automatic raga recognition, composition identification and music recommendation.
Constantine Kotropoulos, Associate Professor, Aristotle University of Thessaloniki, Department of Informatics, Sparse and Low Rank Representations in Music Signal Analysis
Neural Substrates of Music Learning and EmotionsPsyche Loui
Neural Substrates of Music Learning and Emotions | Slides from my talk at The Origins of Music and Human Society, a Conference by Institute of Advanced Study in Toulouse and Royaumont Foundation at Royaumont Abbey, France | December 16, 2017
Discovery and Characterization of Melodic Motives in Large Audio Music Collec...Sankalp Gulati
Sankalp Gulati proposed a methodology for discovering and characterizing melodic motives in large audio music collections using domain knowledge of Indian art music. The methodology involves extracting pitch, loudness, and timbre features from audio signals, representing melodies, calculating melodic similarity, extracting repeated patterns as motives, and analyzing the extracted motives. Gulati aims to apply this methodology to a collection of over 550 hours of Indian art music audio and evaluate the results through listening tests and user feedback.
[Tutorial] Computational Approaches to Melodic Analysis of Indian Art MusicSankalp Gulati
Computational Approaches to Melodic Analysis of Indian Art Music discusses various computational approaches to analyzing the melody of Indian art music, including tonic identification and predominant pitch estimation. For tonic identification, the document describes using multipitch analysis of the audio signal along with a drone background to identify the tonic note, with reported 90% accuracy. It also discusses signal processing techniques used such as STFT and spectral peak picking. For predominant pitch estimation, it reviews various pitch estimation algorithms including autocorrelation-based, frequency-domain, and multipitch approaches, highlighting the YIN algorithm which produces fewer errors than other methods.
Landmark Detection in Hindustani Music MelodiesSankalp Gulati
More info: http://mtg.upf.edu/node/2998
Abstract: Musical melodies contain hierarchically organized events, where some events are more salient than others, acting as melodic landmarks. In Hindustani music melodies, an important landmark is the occurrence of a nyas. Occurrence of nyas is crucial to build and sustain the format of a rag and mark the boundaries of melodic motifs. Detection of nyas segments is relevant to tasks such as melody segmentation, motif discovery and rag recognition. However, detection of nyas segments is challenging as these segments do not follow explicit set of rules in terms of segment length, contour characteristics, and melodic context. In this paper we propose a method for the automatic detection of nyas segments in Hindustani music melodies. It consists of two main steps: a segmentation step that incorporates domain knowledge in order to facilitate the placement of nyas boundaries, and a segment classification step that is based on a series of musically motivated pitch contour features. The proposed method obtains significant accuracies for a heterogeneous data set of 20 audio music recordings containing 1257 nyas svar occurrences and total duration of 1.5 hours. Further, we show that the proposed segmentation strategy significantly improves over a classical piece-wise linear segmentation approach.
Computational Melodic Analysis of Indian Art MusicSankalp Gulati
This document summarizes research on computational melodic analysis of Indian art music. Key areas discussed include tonic identification, predominant melody estimation, motif discovery, melodic similarity, melodic pattern networks, raga recognition using melodic patterns, and resources like datasets and demonstrations of the techniques.
Phrase-based Rāga Recognition Using Vector Space ModelingSankalp Gulati
This document describes an approach for automatic raga recognition in Indian art music using phrase-based vector space modeling. It involves discovering melodic patterns from a Carnatic music collection through intra-recording and inter-recording analysis. The patterns are then clustered into communities based on their network of similarities. Features are extracted from the pattern-recording relationships using term frequency-inverse document frequency weighting. These features are used to build a feature matrix for raga recognition.
Mining Melodic Patterns in Large Audio Collections of Indian Art MusicSankalp Gulati
More info: http://mtg.upf.edu/node/3108
Abstract: Discovery of repeating structures in music is fundamental to its analysis, understanding and interpretation. We present a data-driven approach for the discovery of short-time melodic patterns in large collections of Indian art music. The approach first discovers melodic patterns within an audio recording and subsequently searches for their repetitions in the entire music collection. We compute similarity between melodic patterns using dynamic time warping (DTW). Furthermore, we investigate four different variants of the DTW cost function for rank refinement of the obtained results. The music collection used in this study comprises 1,764 audio recordings with a total duration of 365 hours. Over 13 trillion DTW distance computations are done for the entire dataset. Due to the computational complexity of the task, different lower bounding and early abandoning techniques are applied during DTW distance computation. An evaluation based on expert feedback on a subset of the dataset shows that the discovered melodic patterns are musically relevant. Several musically interesting relationships are discovered, yielding further scope for establishing novel similarity measures based on melodic patterns. The discovered melodic patterns can further be used in challenging computational tasks such as automatic raga recognition, composition identification and music recommendation.
Constantine Kotropoulos, Associate Professor, Aristotle University of Thessaloniki, Department of Informatics, Sparse and Low Rank Representations in Music Signal Analysis
Neural Substrates of Music Learning and EmotionsPsyche Loui
Neural Substrates of Music Learning and Emotions | Slides from my talk at The Origins of Music and Human Society, a Conference by Institute of Advanced Study in Toulouse and Royaumont Foundation at Royaumont Abbey, France | December 16, 2017
Discovery and Characterization of Melodic Motives in Large Audio Music Collec...Sankalp Gulati
Sankalp Gulati proposed a methodology for discovering and characterizing melodic motives in large audio music collections using domain knowledge of Indian art music. The methodology involves extracting pitch, loudness, and timbre features from audio signals, representing melodies, calculating melodic similarity, extracting repeated patterns as motives, and analyzing the extracted motives. Gulati aims to apply this methodology to a collection of over 550 hours of Indian art music audio and evaluate the results through listening tests and user feedback.
[Tutorial] Computational Approaches to Melodic Analysis of Indian Art MusicSankalp Gulati
Computational Approaches to Melodic Analysis of Indian Art Music discusses various computational approaches to analyzing the melody of Indian art music, including tonic identification and predominant pitch estimation. For tonic identification, the document describes using multipitch analysis of the audio signal along with a drone background to identify the tonic note, with reported 90% accuracy. It also discusses signal processing techniques used such as STFT and spectral peak picking. For predominant pitch estimation, it reviews various pitch estimation algorithms including autocorrelation-based, frequency-domain, and multipitch approaches, highlighting the YIN algorithm which produces fewer errors than other methods.
Landmark Detection in Hindustani Music MelodiesSankalp Gulati
More info: http://mtg.upf.edu/node/2998
Abstract: Musical melodies contain hierarchically organized events, where some events are more salient than others, acting as melodic landmarks. In Hindustani music melodies, an important landmark is the occurrence of a nyas. Occurrence of nyas is crucial to build and sustain the format of a rag and mark the boundaries of melodic motifs. Detection of nyas segments is relevant to tasks such as melody segmentation, motif discovery and rag recognition. However, detection of nyas segments is challenging as these segments do not follow explicit set of rules in terms of segment length, contour characteristics, and melodic context. In this paper we propose a method for the automatic detection of nyas segments in Hindustani music melodies. It consists of two main steps: a segmentation step that incorporates domain knowledge in order to facilitate the placement of nyas boundaries, and a segment classification step that is based on a series of musically motivated pitch contour features. The proposed method obtains significant accuracies for a heterogeneous data set of 20 audio music recordings containing 1257 nyas svar occurrences and total duration of 1.5 hours. Further, we show that the proposed segmentation strategy significantly improves over a classical piece-wise linear segmentation approach.
This study explored the physiological and psychological responses to different music frequencies. Participants listened to audio clips at frequencies of A=432Hz and A=440Hz while heart rate, blood oxygen levels, and preference were measured. There was no significant difference in heart rate or oxygen levels between conditions. However, participants significantly preferred and enjoyed the A=440Hz frequency more, which is the standard modern tuning. The results supported the hypothesis that standard tuning would be perceived as more enjoyable.
Trends in Singing Voice Research: An Innovative ApproachPedro Melo Pestana
This study analyzed over 750 scientific papers about singing voice published between 1949-2016 to identify trends in singing voice research. The number of published papers has grown over time but has slowed slightly after 2010. Early research focused on voice quality and demands on singers, using acoustic analysis. Recently, there has been more focus on vocal function rather than structure, and increased use of electroglottography and musical perception studies. The Journal of Voice published the most papers overall. Trend analysis found increasing interest in vocal functionality rather than organic structures after 2010.
The past, present and future of singing synthesisEji Warp
The document discusses the history and current state of speech and singing voice modeling. It addresses limitations in existing models, including the assumptions of quasi-static signals and oversimplified representations of aperiodic components. The author proposes a new "low level speech model" that models the harmonic, noise and transient components separately and incorporates a more accurate glottal/source model. This could enable applications like improved pitch shifting by revealing the instants of vocal fold closure. Overall, the document analyzes past approaches, outlines ongoing challenges, and presents a future direction of research toward a more robust low-level speech model.
Graphical visualization of musical emotionsPranay Prasoon
The document discusses graphical visualization of musical emotions using artificial neural networks. 13 audio features are extracted from Hindustani classical music clips labeled as happy or sad. An ANN model with backpropagation algorithm is trained on 70% of data, validated on 15% and tested on 15%. The model correctly classified 15 of 17 happy clips and 21 of 22 sad clips. Testing was repeated 10 times with over 90% accuracy each time, showing the model effectively recognizes musical emotions. Future work involves expanding the model to recognize additional emotions and incorporating physiological features.
Performance Comparison of Musical Instrument Family Classification Using Soft...Waqas Tariq
Nowadays, it appears essential to design automatic and efficacious classification algorithm for the musical instruments. Automatic classification of musical instruments is made by extracting relevant features from the audio samples, afterward classification algorithm is used (using these extracted features) to identify into which of a set of classes, the sound sample is possible to fit. The aim of this paper is to demonstrate the viability of soft set for audio signal classification. A dataset of 104 (single monophonic notes) pieces of Traditional Pakistani musical instruments were designed. Feature extraction is done using two feature sets namely perception based and mel-frequency cepstral coefficients (MFCCs). In a while, two different classification techniques are applied for classification task, which are soft set (comparison table) and fuzzy soft set (similarity measurement). Experimental results show that both classifiers can perform well on numerical data. However, soft set achieved accuracy up to 94.26% with best generated dataset. Consequently, these promising results provide new possibilities for soft set in classifying musical instrument sounds. Based on the analysis of the results, this study offers a new view on automatic instrument classification
Audio descriptive analysis of singer and musical instrument identification in...eSAT Journals
Abstract Music information retrieval (MIR) has reached to a reasonably stable state after advancement in the Low Level audio Descriptors (LLDs) and feature extraction techniques. The analysis of sound has now become simple by the continuous efforts and research of MIR community in the field of signal processing from last two decades. In north Indian classical music, a singer is accompanied by some instruments such as harmonium, violin or flute. These instruments are tuned in the same musical scale (pitch range) in which the singer is signing. Separate researches have been made in recent past to identify a musical instrument and a singer. In this paper, we have analyzed the low level audio descriptors, for singing voice and musical instrument sound together, that appears to human ear as similar with respect to ‘timbre’, to see if we could treat them same and use identification/ classification routines to classify them into their classes. We have used Hybrid Selection algorithm from wrapper technique(the one that uses classifier also in feature selection process) to identify and extract the features and K-Means and K nearest neighbor classifiers to classify and cross verify the accuracy of classification. The accuracy of classification achieved was 91.1% which clearly proves that musical instruments and singing voice that sounds similar in timbral aspect can be grouped together and classification is possible with mixed database of instruments and singing voices. Keywords: Music Information Retrieval (MIR), Timbre, Singing Voice, Low level Descriptors (LLD, North Indian Classical music. MIRTOOL BOX
This document discusses incorporating music into the social sciences and sciences. It suggests that music can be used to enhance learning and memory in several ways. Music can improve recall of words and texts when paired with musical elements. It may also facilitate learning by improving auditory processing and creating more positive emotional states. Several studies found that background music did not interfere with verbal learning as long as it was not too loud or complex. Overall, the document argues that music has applications to support learning across various disciplines beyond just music education.
Modeling of Song Pattern Similarity using Coefficient of VarianceGobinda Karmakar ☁
This paper proposes a system to automatically identify the raga and raga cycle of a song by analyzing note frequencies. It calculates the coefficient of variance of note frequencies to measure similarity between songs. If the coefficient of variance is between 0-1, the songs are from the same raga cycle and are similar. The system was tested on songs and accurately identified whether songs were from the same or different raga cycles based on their coefficient of variance.
There are two aspects of dissonance perception: learned/top-down and innate/bottom-up. Sensory dissonance can be modeled using either auditory models based on the auditory periphery or curve-mapping models based on empirical data. Computer programs that simulate sensory dissonance processing can estimate the degree of dissonance for a given sound. The models were tested on piano music, drone music, and synthesized chords by comparing their predictions of dissonance to participant ratings. The curve-mapping models predicted ratings reasonably well for isolated chords and drone music but not piano music, possibly due to non-sensory influences on ratings for more complex music.
BASIC PRINCIPLES OF NETWORKED MUSIC PERFORMANCEAlvaro Barbosa
Presentation at the 132nd Audio Engineering Society Convention in Budapest. The lecture was held during the Distributed Music Workshop with Alexander Carôt, Nathan Brock, Andrea Szigetvári, David Willyard and Karl Steinberg.
April 27th, 2012
The document outlines Olmo Cornelis' research on opportunities for symbiosis between Western and non-Western musical idioms from 2008-2014. It discusses his background, previous work digitizing an ethnomusicological archive, challenges in accurately describing non-Western music, and a proposed methodology using music information retrieval techniques to objectively analyze ethnic music and inform new compositions blending musical elements from different cultures.
Graded unit presentation - Maria Saraiva Maria Saraiva
This document outlines Maria Saraiva's work and development as a session musician for her HND2 Music Graded Unit 2. It describes the core units covered including songwriting, music research, guitar studies, music theory, live performance, and independent studio projects. Specific projects are detailed such as performing with the band Freaky Pouches, assisting a student with their degree show, recording cover songs, learning online promotion strategies, and studying music business topics. Examples of the work produced for these projects are available on Maria's portfolio.
Behavioral and DTI Studies on Normal and Impaired Learning of Musical StructurePsyche Loui
Behavioral and DTI Studies on Normal and Impaired Learning of Musical Structure
A talk for CogSci 2013 in Berlin, August 1, 2013
Youtube video is here: http://www.youtube.com/watch?v=h1PnbDIhOXA
This document discusses a method for extracting vocals from songs and converting them to instrumental covers using deep learning techniques. It involves using the Spleeter library to separate vocals from music tracks. The extracted vocals can then be converted to instrumental covers for different instruments using a DDSP (Differentiable Digital Signal Processing) library combined with pretrained convolutional neural networks. This allows generating instrumental covers from songs to help music students learn instruments without relying on professionals to create covers. The proposed approach could make a variety of instrumental covers more widely available and assist those learning music.
IRJET- Survey on Musical Scale IdentificationIRJET Journal
This document discusses a survey of various techniques for musical scale identification from digital audio signals. It begins with an introduction to important musical concepts like pitch, notes, scales, and ragas. It then reviews existing methods that use digital signal processing and machine learning to analyze features extracted from an audio signal to identify its musical scale. The document surveys several existing works and discusses opportunities to develop new algorithms that can automatically identify scales using mathematical methods applied to concepts from prior research. The goal is to enhance existing methods for musical scale identification from digital audio.
Application of Fisher Linear Discriminant Analysis to Speech/Music Classifica...Lushanthan Sivaneasharajah
This document describes research applying Fisher Linear Discriminant Analysis (LDA) and K-Nearest Neighbors (K-NN) algorithms to classify speech and music audio clips. It finds that Fisher LDA using single features like mel-frequency cepstral coefficients achieves classification error rates below 5%, outperforming K-NN. While combining multiple features does not improve LDA results, combining the outputs of LDA and K-NN classifiers using majority voting further lowers the error rate to 4.5%, demonstrating the benefit of classifier ensembles for this task.
The ancient Greeks, particularly Pythagoras, were the first to analyze music mathematically rather than just appreciating it as art. Pythagoras discovered that the pitch of a note was related to the length of the string producing it, and that this depended on how the string vibrated. This led to the science of acoustics. Pythagoras also developed one of the first mathematical musical scales by considering intervals of octaves and fifths. Later, it was shown that every musical pitch has a distinct frequency and wavelength that can be represented mathematically. Rhythm, tempo, note duration, and other elements of music also have mathematical representations involving fractions, ratios, and other relationships.
Gender and Mental Health - Counselling and Family Therapy Applications and In...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) CurriculumMJDuyan
(𝐓𝐋𝐄 𝟏𝟎𝟎) (𝐋𝐞𝐬𝐬𝐨𝐧 𝟏)-𝐏𝐫𝐞𝐥𝐢𝐦𝐬
𝐃𝐢𝐬𝐜𝐮𝐬𝐬 𝐭𝐡𝐞 𝐄𝐏𝐏 𝐂𝐮𝐫𝐫𝐢𝐜𝐮𝐥𝐮𝐦 𝐢𝐧 𝐭𝐡𝐞 𝐏𝐡𝐢𝐥𝐢𝐩𝐩𝐢𝐧𝐞𝐬:
- Understand the goals and objectives of the Edukasyong Pantahanan at Pangkabuhayan (EPP) curriculum, recognizing its importance in fostering practical life skills and values among students. Students will also be able to identify the key components and subjects covered, such as agriculture, home economics, industrial arts, and information and communication technology.
𝐄𝐱𝐩𝐥𝐚𝐢𝐧 𝐭𝐡𝐞 𝐍𝐚𝐭𝐮𝐫𝐞 𝐚𝐧𝐝 𝐒𝐜𝐨𝐩𝐞 𝐨𝐟 𝐚𝐧 𝐄𝐧𝐭𝐫𝐞𝐩𝐫𝐞𝐧𝐞𝐮𝐫:
-Define entrepreneurship, distinguishing it from general business activities by emphasizing its focus on innovation, risk-taking, and value creation. Students will describe the characteristics and traits of successful entrepreneurs, including their roles and responsibilities, and discuss the broader economic and social impacts of entrepreneurial activities on both local and global scales.
This study explored the physiological and psychological responses to different music frequencies. Participants listened to audio clips at frequencies of A=432Hz and A=440Hz while heart rate, blood oxygen levels, and preference were measured. There was no significant difference in heart rate or oxygen levels between conditions. However, participants significantly preferred and enjoyed the A=440Hz frequency more, which is the standard modern tuning. The results supported the hypothesis that standard tuning would be perceived as more enjoyable.
Trends in Singing Voice Research: An Innovative ApproachPedro Melo Pestana
This study analyzed over 750 scientific papers about singing voice published between 1949-2016 to identify trends in singing voice research. The number of published papers has grown over time but has slowed slightly after 2010. Early research focused on voice quality and demands on singers, using acoustic analysis. Recently, there has been more focus on vocal function rather than structure, and increased use of electroglottography and musical perception studies. The Journal of Voice published the most papers overall. Trend analysis found increasing interest in vocal functionality rather than organic structures after 2010.
The past, present and future of singing synthesisEji Warp
The document discusses the history and current state of speech and singing voice modeling. It addresses limitations in existing models, including the assumptions of quasi-static signals and oversimplified representations of aperiodic components. The author proposes a new "low level speech model" that models the harmonic, noise and transient components separately and incorporates a more accurate glottal/source model. This could enable applications like improved pitch shifting by revealing the instants of vocal fold closure. Overall, the document analyzes past approaches, outlines ongoing challenges, and presents a future direction of research toward a more robust low-level speech model.
Graphical visualization of musical emotionsPranay Prasoon
The document discusses graphical visualization of musical emotions using artificial neural networks. 13 audio features are extracted from Hindustani classical music clips labeled as happy or sad. An ANN model with backpropagation algorithm is trained on 70% of data, validated on 15% and tested on 15%. The model correctly classified 15 of 17 happy clips and 21 of 22 sad clips. Testing was repeated 10 times with over 90% accuracy each time, showing the model effectively recognizes musical emotions. Future work involves expanding the model to recognize additional emotions and incorporating physiological features.
Performance Comparison of Musical Instrument Family Classification Using Soft...Waqas Tariq
Nowadays, it appears essential to design automatic and efficacious classification algorithm for the musical instruments. Automatic classification of musical instruments is made by extracting relevant features from the audio samples, afterward classification algorithm is used (using these extracted features) to identify into which of a set of classes, the sound sample is possible to fit. The aim of this paper is to demonstrate the viability of soft set for audio signal classification. A dataset of 104 (single monophonic notes) pieces of Traditional Pakistani musical instruments were designed. Feature extraction is done using two feature sets namely perception based and mel-frequency cepstral coefficients (MFCCs). In a while, two different classification techniques are applied for classification task, which are soft set (comparison table) and fuzzy soft set (similarity measurement). Experimental results show that both classifiers can perform well on numerical data. However, soft set achieved accuracy up to 94.26% with best generated dataset. Consequently, these promising results provide new possibilities for soft set in classifying musical instrument sounds. Based on the analysis of the results, this study offers a new view on automatic instrument classification
Audio descriptive analysis of singer and musical instrument identification in...eSAT Journals
Abstract Music information retrieval (MIR) has reached to a reasonably stable state after advancement in the Low Level audio Descriptors (LLDs) and feature extraction techniques. The analysis of sound has now become simple by the continuous efforts and research of MIR community in the field of signal processing from last two decades. In north Indian classical music, a singer is accompanied by some instruments such as harmonium, violin or flute. These instruments are tuned in the same musical scale (pitch range) in which the singer is signing. Separate researches have been made in recent past to identify a musical instrument and a singer. In this paper, we have analyzed the low level audio descriptors, for singing voice and musical instrument sound together, that appears to human ear as similar with respect to ‘timbre’, to see if we could treat them same and use identification/ classification routines to classify them into their classes. We have used Hybrid Selection algorithm from wrapper technique(the one that uses classifier also in feature selection process) to identify and extract the features and K-Means and K nearest neighbor classifiers to classify and cross verify the accuracy of classification. The accuracy of classification achieved was 91.1% which clearly proves that musical instruments and singing voice that sounds similar in timbral aspect can be grouped together and classification is possible with mixed database of instruments and singing voices. Keywords: Music Information Retrieval (MIR), Timbre, Singing Voice, Low level Descriptors (LLD, North Indian Classical music. MIRTOOL BOX
This document discusses incorporating music into the social sciences and sciences. It suggests that music can be used to enhance learning and memory in several ways. Music can improve recall of words and texts when paired with musical elements. It may also facilitate learning by improving auditory processing and creating more positive emotional states. Several studies found that background music did not interfere with verbal learning as long as it was not too loud or complex. Overall, the document argues that music has applications to support learning across various disciplines beyond just music education.
Modeling of Song Pattern Similarity using Coefficient of VarianceGobinda Karmakar ☁
This paper proposes a system to automatically identify the raga and raga cycle of a song by analyzing note frequencies. It calculates the coefficient of variance of note frequencies to measure similarity between songs. If the coefficient of variance is between 0-1, the songs are from the same raga cycle and are similar. The system was tested on songs and accurately identified whether songs were from the same or different raga cycles based on their coefficient of variance.
There are two aspects of dissonance perception: learned/top-down and innate/bottom-up. Sensory dissonance can be modeled using either auditory models based on the auditory periphery or curve-mapping models based on empirical data. Computer programs that simulate sensory dissonance processing can estimate the degree of dissonance for a given sound. The models were tested on piano music, drone music, and synthesized chords by comparing their predictions of dissonance to participant ratings. The curve-mapping models predicted ratings reasonably well for isolated chords and drone music but not piano music, possibly due to non-sensory influences on ratings for more complex music.
BASIC PRINCIPLES OF NETWORKED MUSIC PERFORMANCEAlvaro Barbosa
Presentation at the 132nd Audio Engineering Society Convention in Budapest. The lecture was held during the Distributed Music Workshop with Alexander Carôt, Nathan Brock, Andrea Szigetvári, David Willyard and Karl Steinberg.
April 27th, 2012
The document outlines Olmo Cornelis' research on opportunities for symbiosis between Western and non-Western musical idioms from 2008-2014. It discusses his background, previous work digitizing an ethnomusicological archive, challenges in accurately describing non-Western music, and a proposed methodology using music information retrieval techniques to objectively analyze ethnic music and inform new compositions blending musical elements from different cultures.
Graded unit presentation - Maria Saraiva Maria Saraiva
This document outlines Maria Saraiva's work and development as a session musician for her HND2 Music Graded Unit 2. It describes the core units covered including songwriting, music research, guitar studies, music theory, live performance, and independent studio projects. Specific projects are detailed such as performing with the band Freaky Pouches, assisting a student with their degree show, recording cover songs, learning online promotion strategies, and studying music business topics. Examples of the work produced for these projects are available on Maria's portfolio.
Behavioral and DTI Studies on Normal and Impaired Learning of Musical StructurePsyche Loui
Behavioral and DTI Studies on Normal and Impaired Learning of Musical Structure
A talk for CogSci 2013 in Berlin, August 1, 2013
Youtube video is here: http://www.youtube.com/watch?v=h1PnbDIhOXA
This document discusses a method for extracting vocals from songs and converting them to instrumental covers using deep learning techniques. It involves using the Spleeter library to separate vocals from music tracks. The extracted vocals can then be converted to instrumental covers for different instruments using a DDSP (Differentiable Digital Signal Processing) library combined with pretrained convolutional neural networks. This allows generating instrumental covers from songs to help music students learn instruments without relying on professionals to create covers. The proposed approach could make a variety of instrumental covers more widely available and assist those learning music.
IRJET- Survey on Musical Scale IdentificationIRJET Journal
This document discusses a survey of various techniques for musical scale identification from digital audio signals. It begins with an introduction to important musical concepts like pitch, notes, scales, and ragas. It then reviews existing methods that use digital signal processing and machine learning to analyze features extracted from an audio signal to identify its musical scale. The document surveys several existing works and discusses opportunities to develop new algorithms that can automatically identify scales using mathematical methods applied to concepts from prior research. The goal is to enhance existing methods for musical scale identification from digital audio.
Application of Fisher Linear Discriminant Analysis to Speech/Music Classifica...Lushanthan Sivaneasharajah
This document describes research applying Fisher Linear Discriminant Analysis (LDA) and K-Nearest Neighbors (K-NN) algorithms to classify speech and music audio clips. It finds that Fisher LDA using single features like mel-frequency cepstral coefficients achieves classification error rates below 5%, outperforming K-NN. While combining multiple features does not improve LDA results, combining the outputs of LDA and K-NN classifiers using majority voting further lowers the error rate to 4.5%, demonstrating the benefit of classifier ensembles for this task.
The ancient Greeks, particularly Pythagoras, were the first to analyze music mathematically rather than just appreciating it as art. Pythagoras discovered that the pitch of a note was related to the length of the string producing it, and that this depended on how the string vibrated. This led to the science of acoustics. Pythagoras also developed one of the first mathematical musical scales by considering intervals of octaves and fifths. Later, it was shown that every musical pitch has a distinct frequency and wavelength that can be represented mathematically. Rhythm, tempo, note duration, and other elements of music also have mathematical representations involving fractions, ratios, and other relationships.
Similar to Computational Approaches to Melodic Analysis of Indian Art Music (18)
Gender and Mental Health - Counselling and Family Therapy Applications and In...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) CurriculumMJDuyan
(𝐓𝐋𝐄 𝟏𝟎𝟎) (𝐋𝐞𝐬𝐬𝐨𝐧 𝟏)-𝐏𝐫𝐞𝐥𝐢𝐦𝐬
𝐃𝐢𝐬𝐜𝐮𝐬𝐬 𝐭𝐡𝐞 𝐄𝐏𝐏 𝐂𝐮𝐫𝐫𝐢𝐜𝐮𝐥𝐮𝐦 𝐢𝐧 𝐭𝐡𝐞 𝐏𝐡𝐢𝐥𝐢𝐩𝐩𝐢𝐧𝐞𝐬:
- Understand the goals and objectives of the Edukasyong Pantahanan at Pangkabuhayan (EPP) curriculum, recognizing its importance in fostering practical life skills and values among students. Students will also be able to identify the key components and subjects covered, such as agriculture, home economics, industrial arts, and information and communication technology.
𝐄𝐱𝐩𝐥𝐚𝐢𝐧 𝐭𝐡𝐞 𝐍𝐚𝐭𝐮𝐫𝐞 𝐚𝐧𝐝 𝐒𝐜𝐨𝐩𝐞 𝐨𝐟 𝐚𝐧 𝐄𝐧𝐭𝐫𝐞𝐩𝐫𝐞𝐧𝐞𝐮𝐫:
-Define entrepreneurship, distinguishing it from general business activities by emphasizing its focus on innovation, risk-taking, and value creation. Students will describe the characteristics and traits of successful entrepreneurs, including their roles and responsibilities, and discuss the broader economic and social impacts of entrepreneurial activities on both local and global scales.
ISO/IEC 27001, ISO/IEC 42001, and GDPR: Best Practices for Implementation and...PECB
Denis is a dynamic and results-driven Chief Information Officer (CIO) with a distinguished career spanning information systems analysis and technical project management. With a proven track record of spearheading the design and delivery of cutting-edge Information Management solutions, he has consistently elevated business operations, streamlined reporting functions, and maximized process efficiency.
Certified as an ISO/IEC 27001: Information Security Management Systems (ISMS) Lead Implementer, Data Protection Officer, and Cyber Risks Analyst, Denis brings a heightened focus on data security, privacy, and cyber resilience to every endeavor.
His expertise extends across a diverse spectrum of reporting, database, and web development applications, underpinned by an exceptional grasp of data storage and virtualization technologies. His proficiency in application testing, database administration, and data cleansing ensures seamless execution of complex projects.
What sets Denis apart is his comprehensive understanding of Business and Systems Analysis technologies, honed through involvement in all phases of the Software Development Lifecycle (SDLC). From meticulous requirements gathering to precise analysis, innovative design, rigorous development, thorough testing, and successful implementation, he has consistently delivered exceptional results.
Throughout his career, he has taken on multifaceted roles, from leading technical project management teams to owning solutions that drive operational excellence. His conscientious and proactive approach is unwavering, whether he is working independently or collaboratively within a team. His ability to connect with colleagues on a personal level underscores his commitment to fostering a harmonious and productive workplace environment.
Date: May 29, 2024
Tags: Information Security, ISO/IEC 27001, ISO/IEC 42001, Artificial Intelligence, GDPR
-------------------------------------------------------------------------------
Find out more about ISO training and certification services
Training: ISO/IEC 27001 Information Security Management System - EN | PECB
ISO/IEC 42001 Artificial Intelligence Management System - EN | PECB
General Data Protection Regulation (GDPR) - Training Courses - EN | PECB
Webinars: https://pecb.com/webinars
Article: https://pecb.com/article
-------------------------------------------------------------------------------
For more information about PECB:
Website: https://pecb.com/
LinkedIn: https://www.linkedin.com/company/pecb/
Facebook: https://www.facebook.com/PECBInternational/
Slideshare: http://www.slideshare.net/PECBCERTIFICATION
How to Setup Warehouse & Location in Odoo 17 InventoryCeline George
In this slide, we'll explore how to set up warehouses and locations in Odoo 17 Inventory. This will help us manage our stock effectively, track inventory levels, and streamline warehouse operations.
Walmart Business+ and Spark Good for Nonprofits.pdfTechSoup
"Learn about all the ways Walmart supports nonprofit organizations.
You will hear from Liz Willett, the Head of Nonprofits, and hear about what Walmart is doing to help nonprofits, including Walmart Business and Spark Good. Walmart Business+ is a new offer for nonprofits that offers discounts and also streamlines nonprofits order and expense tracking, saving time and money.
The webinar may also give some examples on how nonprofits can best leverage Walmart Business+.
The event will cover the following::
Walmart Business + (https://business.walmart.com/plus) is a new shopping experience for nonprofits, schools, and local business customers that connects an exclusive online shopping experience to stores. Benefits include free delivery and shipping, a 'Spend Analytics” feature, special discounts, deals and tax-exempt shopping.
Special TechSoup offer for a free 180 days membership, and up to $150 in discounts on eligible orders.
Spark Good (walmart.com/sparkgood) is a charitable platform that enables nonprofits to receive donations directly from customers and associates.
Answers about how you can do more with Walmart!"
Computational Approaches to Melodic Analysis of Indian Art Music
1. Computational Approaches to Melodic
Analysis of Indian Art Music
Indian Institute of Sciences, Bengaluru, India 2016
Sankalp Gulati
Music Technology Group, Universitat Pompeu Fabra, Barcelona, Spain
4. Tonic Identification
time (s)
Frequency(Hz)
0 1 2 3 4 5 6 7 8
0
1000
2000
3000
4000
5000
100 150 200 250 300
0
0.2
0.4
0.6
0.8
1
Frequency (bins), 1bin=10 cents, Ref=55 Hz
Normalizedsalience
f2
f3
f4
f
5f6
Tonic
Signal processing Learning
q Tanpura / drone background sound
q Extent of gamakas on Sa and Pa svara
q Vadi, sam-vadi svara of the rāga
S. Gulati, A. Bellur, J. Salamon, H. Ranjani, V. Ishwar, H.A. Murthy, and X. Serra. Automatic tonic identification in Indian art music: approaches
and evaluation. Journal of New Music Research, 43(01):55–73, 2014.
Salamon, J., Gulati, S., & Serra, X. (2012). A multipitch approach to tonic identification in Indian classical music. In Proc. of Int. Conf. on Music
Information Retrieval (ISMIR) (pp. 499–504), Porto, Portugal.
Bellur, A., Ishwar, V., Serra, X., & Murthy, H. (2012). A knowledge based signal processing approach to tonic identification in Indian classical music. In 2nd
CompMusic Workshop (pp. 113–118) Istanbul, Turkey.
Ranjani, H. G., Arthi, S., & Sreenivas, T. V. (2011). Carnatic music analysis: Shadja, swara identification and raga verification in Alapana using stochastic
models. Applications of Signal Processing to Audio and Acoustics (WASPAA), IEEE Workshop , 29–32, New Paltz, NY.
Accuracy : ~90% !!!
8. q Pitch (Fundamental frequency-F0) of the lead
artist
q Pitch estimation
§ Melodic contour characteristics
§ Dual melodic lines in Indian art music
Signal processing
Learning
Salamon, Justin, and Emilia Gómez. "Melody extraction from polyphonic music signals using pitch contour characteristics." Audio, Speech, and Language
Processing, IEEE Transactions on 20.6 (2012): 1759-1770.
Rao, Vishweshwara, and Preeti Rao. "Vocal melody extraction in the presence of pitched accompaniment in polyphonic music." Audio, Speech, and
Language Processing, IEEE Transactions on 18.8 (2010): 2145-2154.
De Cheveigné, A., & Kawahara, H. (2002). YIN, a fundamental frequency estimator for speech and music. The Journal of the Acoustical Society of America,
111, 1917.
16. Melody Histogram Computation
50 100 150 200 250 300
0
0.2
0.4
0.6
0.8
1
Frequency (bins), 1 bin = 10 Cents, Ref = 55 Hz
Normalizedsalience
Lower Sa
Tonic
middle Sa
Higher Sa
Frequency (cents), fref = tonic frequency
0 120-120
time (s)
10 30
17. Intonation Analysis
Mohana - G Begada - G
• Koduri, Gopala Krishna, et al. "Intonation Analysis of Rāgas in Carnatic Music." Journal of New Music Research
43.1 (2014): 72-93.
• Koduri, Gopala K., Serrà Joan, and Xavier Serra. "Characterization of Intonation in Carnatic Music by
Parametrizing Pitch Histograms." (2012): 199-204.
24. Melodic Pattern Discovery
-Predominant pitch
estimation
-Downsampling
-Hz to Cents
-Tonic normalization
-Brute-force
segmentation
-Segment filtering
-Uniform Time-scaling
Flat Non-flat
Data processing Intra-recording
discovery
Inter-recording
search
Rank-refinement
q S. Gulati, J. Serrà, V. Ishwar, and X. Serra, “Mining melodic patterns in large audio collections
of Indian art music,” in Int. Conf. on Signal Image Technology & Internet Based Systems -
MIRA, Marrakesh, Morocco, 2014, pp. 264–271.
25. Data Preprocessing
-Predominant pitch
estimation
-Downsampling
-Hz to Cents
-Tonic normalization
-Brute-force
segmentation
-Segment filtering
-Uniform Time-scaling
Flat Non-flat
q S. Gulati, J. Serrà, V. Ishwar, and X. Serra, “Mining melodic patterns in large audio collections
of Indian art music,” in Int. Conf. on Signal Image Technology & Internet Based Systems -
MIRA, Marrakesh, Morocco, 2014, pp. 264–271.
26. Melodic Similarity
q S. Gulati, J. Serrà and X. Serra, "An Evaluation of Methodologies for Melodic Similarity in
Audio Recordings of Indian Art Music", in Proceedings of IEEE Int. Conf. on Acoustics,
Speech, and Signal Processing (ICASSP), Brisbane, Australia 2015
28. Computational Complexity
q Lower bounding techniques (DTW)
Rakthanmanon, T., Campana, B., Mueen, A., Batista, G., Westover, B., Zhu, Q., ... & Keogh, E. (2012,
August). Searching and mining trillions of time series subsequences under dynamic time warping. In
Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data
mining (pp. 262-270). ACM.
Image taken from: http://www.cs.ucr.edu/~eamonn/LB_Keogh.htm
29. Melodic Similarity Improvements
q S. Gulati, J. Serrà and X. Serra, "Improving Melodic Similarity in Indian Art Music Using
Culture-specific Melodic Characteristics", in International Society for Music Information
Retrieval Conference (ISMIR) , pp. 680-686, Spain, 2015
31. Melodic Pattern Network
q M. EJ Newman, “The structure and function of complex networks,” Society for Industrial and
Applied Mathematics (SIAM) review, vol. 45, no. 2, pp. 167–256, 2003.
Undirectional
33. Similarity Threshold Estimation
q M. EJ Newman, “The structure and function of complex networks,” Society for Industrial and Applied
Mathematics (SIAM) review, vol. 45, no. 2, pp. 167–256, 2003.
q S. Maslov and K. Sneppen, “Specificity and stability in topology of protein networks,” Science, vol. 296, no.
5569, pp. 910– 913, 2002.
Ts
*
34. Melodic Pattern Characterization
V. D. Blondel, J. L. Guillaume, R. Lambiotte, and E. Lefebvre, “Fast unfolding of communities in large networks,”
Journal of Statistical Mechanics: Theory and Experiment, vol. 2008, no. 10, pp. P10008, 2008.
35. Melodic Pattern Characterization
V. D. Blondel, J. L. Guillaume, R. Lambiotte, and E. Lefebvre, “Fast unfolding of communities in large networks,”
Journal of Statistical Mechanics: Theory and Experiment, vol. 2008, no. 10, pp. P10008, 2008.
36. Melodic Pattern Characterization
V. D. Blondel, J. L. Guillaume, R. Lambiotte, and E. Lefebvre, “Fast unfolding of communities in large networks,”
Journal of Statistical Mechanics: Theory and Experiment, vol. 2008, no. 10, pp. P10008, 2008.
38. Melodic Pattern Characterization
q S. Gulati, J. Serrà, V. Ishwar, S. Şentürk and X. Serra, "Discovering Rāga Motifs by
characterizing Communities in Networks of Melodic Patterns", in IEEE Int. Conf. on
Acoustics, Speech, and Signal Processing (ICASSP), pp. 286-290, Shanghai, China, 2016.
43. Rāga Characterization: Svaras
50 100 150 200 250 300
0
0.2
0.4
0.6
0.8
1
Frequency (bins), 1 bin = 10 Cents, Ref = 55 Hz
Normalizedsalience
Lower Sa
Tonic
middle Sa
Higher Sa
Frequency (cents), fref = tonic frequency
0 120-120
q P. Chordia and S. Şentürk, “Joint recognition of raag and tonic in North Indian music,” Computer Music Journal, vol.
37, no. 3, pp. 82–98, 2013.
q G. K. Koduri, S. Gulati, P. Rao, and X. Serra, “Rāga recognition based on pitch distribution methods,” Journal of New
Music Research, vol. 41, no. 4, pp. 337–350, 2012.
time (s)
10 30
44. Rāga Characterization: Intonation
50 100 150 200 250 300
0
0.2
0.4
0.6
0.8
1
Frequency (bins), 1 bin = 10 Cents, Ref = 55 Hz
Normalizedsalience
Lower Sa
Tonic
middle Sa
Higher Sa
Frequency (cents), fref = tonic frequency
0 120-120
time (s)
10 30
45. Rāga Characterization: Intonation
50 100 150 200 250 300
0
0.2
0.4
0.6
0.8
1
Frequency (bins), 1 bin = 10 Cents, Ref = 55 Hz
Normalizedsalience
Lower Sa
Tonic
middle Sa
Higher Sa
Frequency (cents), fref = tonic frequency
0 120-120
q G.K.Koduri,V.Ishwar,J.Serrà,andX.Serra,“Intonation analysis of rāgas in Carnatic music,” Journal of New Music
Research, vol. 43, no. 1, pp. 72–93, 2014.
q H. G. Ranjani, S. Arthi, and T. V. Sreenivas, “Carnatic music analysis: Shadja, swara identification and raga
verification in alapana using stochastic models,” in IEEE WASPAA, 2011, pp. 29–32.
time (s)
10 30
47. time (s)
10 30
Rāga Characterization: Ārōh-Avrōh
q S. Shetty and K. K. Achary, “Raga mining of indian music by extracting arohana-avarohana pattern,” Int.
Journal of Recent Trends in Engineering, vol. 1, no. 1, pp. 362–366, 2009.
q V. Kumar, H Pandya, and C. V. Jawahar, “Identifying ragas in indian music,” in 22nd Int. Conf. on Pattern
Recognition (ICPR), 2014, pp. 767–772.
q P. V. Rajkumar, K. P. Saishankar, and M. John, “Identification of Carnatic raagas using hidden markov
models,” in IEEE 9th Int. Symposium on Applied Machine Intelligence and Informatics (SAMI), 2011, pp.
107–110.
Melodic Progression Templates
N-gram Distribution
Hidden Markov Model
49. time (s)
10 30
Rāga Characterization: Melodic motifs
q R. Sridhar and T. V. Geetha, “Raga identification of carnatic music for music information retrieval,”
International Journal of Recent Trends in Engineering, vol. 1, no. 1, pp. 571–574, 2009.
q S. Dutta, S. PV Krishnaraj, and H. A. Murthy, “Raga verification in carnatic music using longest common
segment set,” in Int. Soc. for Music Information Retrieval Conf. (ISMIR), pp. 605-611,2015
Rāga A Rāga B Rāga C
50. Goal
q Automatic rāga recognition
Training corpus
Rāga recognition
System
Rāga label
Yaman
Shankarabharnam
Todī
Darbari
Kalyan
Bageśrī
Kambhojī
Hamsadhwani
Des
Harikambhoji
Kirvani
Atana
Behag
Kapi
Begada
51. Goal
q Automatic rāga recognition
Training corpus
Rāga recognition
System
Rāga label
time (s)
10
time (s)
10
time (s)
30
time (s)
10
time (s)
10
time (s)
30
time (s
10
time (s)
10
time (s)
30
time
10
time (s
10
time (s)
30
tim
10
time (
10
time (s)
0 30
tim
10
time
10
time (s)
10 30
t
10
tim
10
time (s)
10 30
10
ti
10
time (s)
10 30
1010
time (s)
10 30
1010
time (s)
10 30
1010
time (s)
10 30
1010
time (s)
10 30
1010
time (s)
10 30
1010
time (s)
10 30
1010
time (s)
10 30
Yaman
Shankarabharnam
Todī
Darbari
Kalyan
Bageśrī
Kambhojī
Hamsadhwani
Des
Harikambhoji
Kirvani
Atana
Behag
Kapi
Begada
time (s)
10 30
time (s)
10 30
30
58. Classification methodology
q Experimental setup
§ Stratified 12-fold cross validation (balanced)
§ Repeat experiment 20 times
§ Evaluation measure: mean classification accuracy
q Classifiers
§ Multinomial, Gaussian and Bernoulli naive Bayes
(NBM, NBG and NBB)
§ SVM with a linear and RBF-kernel, and with a
SGD learning (SVML, SVMR and SGD)
§ logistic regression (LR) and random forest (RF)
59. Results
phrase
e of a
of oc-
as, we
es that
ile for
re vec-
ument
(2)
(3)
ere the
rdings.
db Mtd Ftr NBM NBB LR SVML 1NN
DB10r¯aga
M
F1 90.6 74 84.1 81.2 -
F2 91.7 73.8 84.8 81.2 -
F3 90.5 74.5 84.3 80.7 -
S1
PCD120 - - - - 82.2
PCDfull - - - - 89.5
S2 PDparam 37.9 11.2 70.1 65.7 -
DB40r¯aga
M
F1 69.6 61.3 55.9 54.6 -
F2 69.6 61.7 55.7 54.3 -
F3 69.5 61.5 55.9 54.5 -
S1
PCD120 - - - - 66.4
PCDfull - - - - 74.1
S2 PDparam 20.8 2.6 51.4 44.2 -
Table 1. Accuracy (in percentage) of different methods (Mtd) for
two datasets (db) using different classifiers and features (Ftr).
the evaluation measure. In order to assess if the difference in the
S. GulaD, J. Serrà, V. Ishwar and X. Serra, "Phrase-based Rāga RecogniDon Using Vector
Space Modelling", in IEEE Int. Conf. on AcousDcs, Speech, and Signal Processing (ICASSP),
pp. 66-70, Shanghai, China, 2016.