Music Information Retrieval (MIR) is an interdisciplinary field that retrieves information from music. MIR aims to make the world's music accessible by using techniques from computer science, information retrieval, audio processing, musicology and more. MIR applications include music document retrieval, recommender systems, emotion detection and more. Music document retrieval identifies music through metadata like Table of Contents or content-based fingerprinting like Shazam. Emotion detection in music aims to classify emotions in music but faces challenges due to subjective human emotion and requires multi-disciplinary techniques.
The document discusses music information retrieval (MIR) and content-based approaches. It describes how MIR deals with intrinsic characteristics of music like pitch, intensity, and timbre. Key concepts covered include music formats, dimensions of music language, and different types of music users. The document also summarizes Shazam's process for identifying songs from short audio clips, which involves fingerprinting audio files and matching fingerprints to identify songs.
CSTalks - Music Information Retrieval - 23 Febcstalks
The document discusses similarity measures used in music information retrieval systems. It defines music information retrieval as searching for music objects using musical queries. Some applications of MIR discussed are music search and recommendation. The document outlines different methods for calculating musical similarity, including text-based, audio feature-based, semantic concept-based, and multimodal fusion approaches. It concludes by noting future directions for similarity measures in MIR.
Query By Humming - Music Retrieval TechniqueShital Kat
This seminar report summarizes query by humming technology. The basic architecture involves extracting melodic information from a hummed input, transcribing it, and comparing it to melodic contours in a database. Challenges include imperfect user queries and accurately capturing pitches from hums. Popular query by humming applications include Shazam, SoundHound, and Midomi. The report also discusses file formats like WAV and MIDI, and the Parsons code algorithm for representing melodies.
slides presented at a three-hour local AI music course in Taiwan in Oct 2021; part 1: a brief introduction to music information retrieval (+analysis, +generation)
인공지능의 음악 인지 모델 - 65차 한국음악지각인지학회 기조강연 (최근우 박사)Keunwoo Choi
The document discusses artificial intelligence models for music perception. It summarizes the talk that analyzes and classifies music AI into analysis, creation, signal generation, and signal processing. Specifically, the analysis part is discussed in detail by dividing it into timbre, notes, and lyrics recognition. Through this, we can understand what music AI researchers aim for, assume, develop, neglect, and misunderstand.
Audio descriptive analysis of singer and musical instrument identification in...eSAT Journals
Abstract Music information retrieval (MIR) has reached to a reasonably stable state after advancement in the Low Level audio Descriptors (LLDs) and feature extraction techniques. The analysis of sound has now become simple by the continuous efforts and research of MIR community in the field of signal processing from last two decades. In north Indian classical music, a singer is accompanied by some instruments such as harmonium, violin or flute. These instruments are tuned in the same musical scale (pitch range) in which the singer is signing. Separate researches have been made in recent past to identify a musical instrument and a singer. In this paper, we have analyzed the low level audio descriptors, for singing voice and musical instrument sound together, that appears to human ear as similar with respect to ‘timbre’, to see if we could treat them same and use identification/ classification routines to classify them into their classes. We have used Hybrid Selection algorithm from wrapper technique(the one that uses classifier also in feature selection process) to identify and extract the features and K-Means and K nearest neighbor classifiers to classify and cross verify the accuracy of classification. The accuracy of classification achieved was 91.1% which clearly proves that musical instruments and singing voice that sounds similar in timbral aspect can be grouped together and classification is possible with mixed database of instruments and singing voices. Keywords: Music Information Retrieval (MIR), Timbre, Singing Voice, Low level Descriptors (LLD, North Indian Classical music. MIRTOOL BOX
When recording sound electronically, there will likely be analogue and digital distortions. Mono recordings use one microphone or speaker while stereo uses more than one. Sound is recorded in digital audio files like .mp3, .wav, etc. Analogue sound uses less bandwidth and is more accurate but digital sound is more reliable, flexible, and compatible with other digital systems. CDs and DVDs can store music, photos, or videos for playback. Television and mobile phones use sound to convey emotion and allow communication. Pitch describes the ordering of sounds on the frequency scale as higher or lower. Decibels indicate sound intensity while loudspeakers generate sound waves. Frequency is the rate of sound wave movement measured in Hz.
The document discusses music information retrieval (MIR) and content-based approaches. It describes how MIR deals with intrinsic characteristics of music like pitch, intensity, and timbre. Key concepts covered include music formats, dimensions of music language, and different types of music users. The document also summarizes Shazam's process for identifying songs from short audio clips, which involves fingerprinting audio files and matching fingerprints to identify songs.
CSTalks - Music Information Retrieval - 23 Febcstalks
The document discusses similarity measures used in music information retrieval systems. It defines music information retrieval as searching for music objects using musical queries. Some applications of MIR discussed are music search and recommendation. The document outlines different methods for calculating musical similarity, including text-based, audio feature-based, semantic concept-based, and multimodal fusion approaches. It concludes by noting future directions for similarity measures in MIR.
Query By Humming - Music Retrieval TechniqueShital Kat
This seminar report summarizes query by humming technology. The basic architecture involves extracting melodic information from a hummed input, transcribing it, and comparing it to melodic contours in a database. Challenges include imperfect user queries and accurately capturing pitches from hums. Popular query by humming applications include Shazam, SoundHound, and Midomi. The report also discusses file formats like WAV and MIDI, and the Parsons code algorithm for representing melodies.
slides presented at a three-hour local AI music course in Taiwan in Oct 2021; part 1: a brief introduction to music information retrieval (+analysis, +generation)
인공지능의 음악 인지 모델 - 65차 한국음악지각인지학회 기조강연 (최근우 박사)Keunwoo Choi
The document discusses artificial intelligence models for music perception. It summarizes the talk that analyzes and classifies music AI into analysis, creation, signal generation, and signal processing. Specifically, the analysis part is discussed in detail by dividing it into timbre, notes, and lyrics recognition. Through this, we can understand what music AI researchers aim for, assume, develop, neglect, and misunderstand.
Audio descriptive analysis of singer and musical instrument identification in...eSAT Journals
Abstract Music information retrieval (MIR) has reached to a reasonably stable state after advancement in the Low Level audio Descriptors (LLDs) and feature extraction techniques. The analysis of sound has now become simple by the continuous efforts and research of MIR community in the field of signal processing from last two decades. In north Indian classical music, a singer is accompanied by some instruments such as harmonium, violin or flute. These instruments are tuned in the same musical scale (pitch range) in which the singer is signing. Separate researches have been made in recent past to identify a musical instrument and a singer. In this paper, we have analyzed the low level audio descriptors, for singing voice and musical instrument sound together, that appears to human ear as similar with respect to ‘timbre’, to see if we could treat them same and use identification/ classification routines to classify them into their classes. We have used Hybrid Selection algorithm from wrapper technique(the one that uses classifier also in feature selection process) to identify and extract the features and K-Means and K nearest neighbor classifiers to classify and cross verify the accuracy of classification. The accuracy of classification achieved was 91.1% which clearly proves that musical instruments and singing voice that sounds similar in timbral aspect can be grouped together and classification is possible with mixed database of instruments and singing voices. Keywords: Music Information Retrieval (MIR), Timbre, Singing Voice, Low level Descriptors (LLD, North Indian Classical music. MIRTOOL BOX
When recording sound electronically, there will likely be analogue and digital distortions. Mono recordings use one microphone or speaker while stereo uses more than one. Sound is recorded in digital audio files like .mp3, .wav, etc. Analogue sound uses less bandwidth and is more accurate but digital sound is more reliable, flexible, and compatible with other digital systems. CDs and DVDs can store music, photos, or videos for playback. Television and mobile phones use sound to convey emotion and allow communication. Pitch describes the ordering of sounds on the frequency scale as higher or lower. Decibels indicate sound intensity while loudspeakers generate sound waves. Frequency is the rate of sound wave movement measured in Hz.
Anat Gilboa's thesis presentation to the University of Virginia School of Engineering and Applied Science.
Over the course of her final semester, Anat and Dr. Qi, PhD, tested various methods to determine similarity between songs using features extracted from metadata in the Million Song Dataset.
This document defines the basic elements of music and their functions. It discusses that music is organized sound with elements like pitch, duration, timbre, harmony, texture and dynamics. It then explains the various functions of music including aesthetics, emotional appeal, nationalism, entertainment, and marketing. It concludes by describing the key musical elements of rhythm, melody, harmony, texture, timbre, dynamics, form, and musical instruments.
GS6887A: Applying Musical Expectation- Perception and Interpretation (Individ...Stefan
This document discusses musical expectation from a knowledge management perspective. It describes how listeners implicitly gain tacit musical knowledge through experiences that integrate explicit data and theories. Musical expectations involve perceiving patterns in pitch, meter, tempo, density and being surprised by deviations from these patterns. The document also discusses how musical prosody, or expressive nuances in performance, can enhance or violate expectations and elicit emotional responses like chills or tears. It provides examples of how Stravinsky's Rite of Spring defied expectations by using unusual harmonies and structures.
Yi-Hsuan Yang is an Associate Research Fellow with Academia Sinica. He received his Ph.D. degree in Communication Engineering from National Taiwan University in 2010, and became an Assistant Research Fellow in Academia Sinica in 2011. He is also an Adjunct Associate Professor with the National Tsing Hua University, Taiwan. His research interests include music information retrieval, machine learning and affective computing. Dr. Yang was a recipient of the 2011 IEEE Signal Processing Society (SPS) Young Author Best Paper Award, the 2012 ACM Multimedia Grand Challenge First Prize, and the 2014 Ta-You Wu Memorial Research Award of the Ministry of Science and Technology, Taiwan. He is an author of the book Music Emotion Recognition (CRC Press 2011) and a tutorial speaker on music affect recognition in the International Society for Music Information Retrieval Conference (ISMIR 2012). In 2014, he served as a Technical Program Co-chair of ISMIR, and a Guest Editor of the IEEE Transactions on Affective Computing and the ACM Transactions on Intelligent Systems and Technology.
This document defines and explains the common elements of music, including rhythm, melody, harmony, texture, tempo, and form. It describes rhythm as the variation of accentuated sounds over time, and includes aspects like pulse, meter, and melodic rhythm. Melody is defined as a series of tones in succession organized by pitch. Elements of melody include pitch, dynamics, and timbre. Harmony refers to how melodies interact vertically. Texture describes the number of musical lines and their relationship. Tempo is the speed of a piece, while form refers to a composition's overall structure and layout in sections.
The document outlines Olmo Cornelis' research on opportunities for symbiosis between Western and non-Western musical idioms from 2008-2014. It discusses his background, previous work digitizing an ethnomusicological archive, challenges in accurately describing non-Western music, and a proposed methodology using music information retrieval techniques to objectively analyze ethnic music and inform new compositions blending musical elements from different cultures.
This document provides an overview of sound and audio concepts. It begins with the basic physics of sound, discussing how sound is formed through vibration of air molecules. It then covers types of sound including voice, sound effects, and music used in film. Key audio concepts like intensity, pitch, attack/sustain/decay are explained. The document also discusses modes of listening, sound art, recorded sound, and includes examples of early sound artists. Microphone basics and considerations for achieving realism in recorded sound are also covered.
This document discusses symbolic melodic similarity and methods for representing and comparing melodies symbolically. It presents several approaches for representing melodies, such as sequences of text, graphs, n-grams, and curves. It focuses on an approach where melodies are represented as curves interpolated from note points, and similarity is measured by aligning the curves and comparing the areas between their derivatives. Evaluation results from music information retrieval competitions show this curve alignment approach performs well. The document concludes with a demonstration of melody comparison software implementing these techniques.
This document discusses conceptual metaphors in music through the lens of the Study and Research Group on Musical Metaphors (GERMM). It provides examples of conceptual metaphors that understand musical ideas in terms of other domains, such as architecture, language, and the body. The group aims to study the links between metaphorical language, conceptual metaphors, and our physical and sensory experiences of music through techniques like questionnaires, interviews, motion capture, and neuroimaging. Their research could provide insights into how metaphor, expectation, and embodiment relate to musical meaning and response.
MUSIC OF LOWLAND LUZON-ELEMENTS OF MUSIC.pptxRhiaLopez3
This document discusses the musical elements and characteristics of folk songs from the lowlands of Luzon in the Philippines. It describes the typical melody, rhythm, texture, harmony, dynamics, tempo, and timbre of these folk songs. It also discusses the types of folk songs based on function, including ballads, lullabies, love songs, and work songs. The document provides examples of two specific folk songs, Atin Cu Pung Singsing and Pamulinawen, summarizing their musical elements, form, and function.
Rhythm, tempo, melody, dynamics, harmony, timbre, and form are elements of music discussed in the document. It provides explanations of various musical terms including accent and syncopation in rhythm, harmony as chords that follow each other, and tempo as the basic pace of music. Musical form is described as the organization of musical elements in time that provides structure. The document also discusses techniques used in musical form such as repetition, contrast, and variation.
This document discusses music representation and notation. It begins by motivating the importance of representation for content-based music information retrieval. It then examines different levels of music representation from abstract semantics to concrete notation. The document analyzes the complexities of music notation, particularly western conventional music notation, and how it both enables and challenges computational interpretation. It concludes by emphasizing the need for high-level intelligence to fully interpret music notation and represent its semantics.
Introduction to LC Faceted Vocabularies for Music Resources (August 2018)ALATechSource
This document discusses new Library of Congress vocabularies for describing music resources, including the Library of Congress Genre/Form Terms for Library and Archival Materials (LCGFT), the Library of Congress Medium of Performance Thesaurus (LCMPT), and the Library of Congress Demographic Group Terms (LCDGT). It provides examples of how to search, browse, and apply these terms in bibliographic records to classify genre, medium of performance, audience, and creator/contributor characteristics in a more granular and standardized way than was possible with LC Subject Headings alone. Best practices for using these new terms are still developing.
This lecture provides an overview of musicology and how it relates to the analysis of popular music. It discusses the history of musicology focusing originally on art music, and how it has expanded to include popular music. The goals of the module are introduced as analyzing popular music forms, developing critical skills, and improving writing and presentation abilities. An overview of the course schedule is given along with details on assessments, which include a group presentation and written essay. Various analytical tools and layers involved in analyzing songs, arrangements, and recorded tracks are defined and examples are discussed.
1. The study analyzed bass melodies from Red Hot Chili Peppers albums before and after 1999 to investigate changes in the playing style of bassist Flea over time. Features related to pitch, duration, and note counts were extracted from MIDI data and used to accurately classify melodies as pre- or post-1999.
2. A second study investigated the relationship between verbal impressions of equalized vocal tones (e.g. warm, bright) and parametric equalizer settings. Participants evaluated modified vocal recordings and their responses were used to map impressions to frequency boosts and cuts.
3. The results of both studies could help develop music analysis and generation systems that incorporate stylistic changes over time or allow intuitive equalizer
This document describes an algorithm for automatically generating optimized music playlists that minimize jarring transitions between songs. The algorithm uses a directed graph to represent songs as nodes and possible transitions as weighted edges. It calculates pair-wise transition scores between songs based on the distance between them across 5 dimensions: mood, energy, volume, danceability, and beats per minute. The algorithm performs a depth-first search of the graph to visit as many songs as possible while choosing low-weight edge transitions, generating a playlist with smoother acoustic transitions between songs. It then evaluates the generated playlist by comparing its average transition score to that of a randomly shuffled version.
This document describes an algorithm for automatically generating optimal music playlists that minimize jarring transitions between songs. The algorithm uses a directed graph to represent songs as nodes and possible transitions as weighted edges. It calculates pair-wise transition scores between songs based on the distance between them across 5 dimensions: mood, energy, volume, danceability, and beats per minute. The algorithm performs a depth-first search of the graph to visit as many nodes as possible while using low-weight edges, representing smoother transitions, to generate a playlist. It then evaluates the generated playlist by comparing its average transition score to that of a randomly shuffled version.
Research Skills Musicology Final Session Prior To Easter BreakPaul Carr
This document provides guidance for a 2000-word musicology essay final assignment. Students can analyze a single piece of music or compare two pieces using discussed methodologies, such as Philip Tagg's approach. Alternatively, students can present a contextual analysis of an artist, discussing factors like authenticity, local/global influences, and how the artist relates to other musicians and styles. The essay should include transcriptions, recordings, and references. Assessment will consider the analysis detail, use of examples, and cross-referencing of academic texts. The deadline is May 7th, 2010.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
Anat Gilboa's thesis presentation to the University of Virginia School of Engineering and Applied Science.
Over the course of her final semester, Anat and Dr. Qi, PhD, tested various methods to determine similarity between songs using features extracted from metadata in the Million Song Dataset.
This document defines the basic elements of music and their functions. It discusses that music is organized sound with elements like pitch, duration, timbre, harmony, texture and dynamics. It then explains the various functions of music including aesthetics, emotional appeal, nationalism, entertainment, and marketing. It concludes by describing the key musical elements of rhythm, melody, harmony, texture, timbre, dynamics, form, and musical instruments.
GS6887A: Applying Musical Expectation- Perception and Interpretation (Individ...Stefan
This document discusses musical expectation from a knowledge management perspective. It describes how listeners implicitly gain tacit musical knowledge through experiences that integrate explicit data and theories. Musical expectations involve perceiving patterns in pitch, meter, tempo, density and being surprised by deviations from these patterns. The document also discusses how musical prosody, or expressive nuances in performance, can enhance or violate expectations and elicit emotional responses like chills or tears. It provides examples of how Stravinsky's Rite of Spring defied expectations by using unusual harmonies and structures.
Yi-Hsuan Yang is an Associate Research Fellow with Academia Sinica. He received his Ph.D. degree in Communication Engineering from National Taiwan University in 2010, and became an Assistant Research Fellow in Academia Sinica in 2011. He is also an Adjunct Associate Professor with the National Tsing Hua University, Taiwan. His research interests include music information retrieval, machine learning and affective computing. Dr. Yang was a recipient of the 2011 IEEE Signal Processing Society (SPS) Young Author Best Paper Award, the 2012 ACM Multimedia Grand Challenge First Prize, and the 2014 Ta-You Wu Memorial Research Award of the Ministry of Science and Technology, Taiwan. He is an author of the book Music Emotion Recognition (CRC Press 2011) and a tutorial speaker on music affect recognition in the International Society for Music Information Retrieval Conference (ISMIR 2012). In 2014, he served as a Technical Program Co-chair of ISMIR, and a Guest Editor of the IEEE Transactions on Affective Computing and the ACM Transactions on Intelligent Systems and Technology.
This document defines and explains the common elements of music, including rhythm, melody, harmony, texture, tempo, and form. It describes rhythm as the variation of accentuated sounds over time, and includes aspects like pulse, meter, and melodic rhythm. Melody is defined as a series of tones in succession organized by pitch. Elements of melody include pitch, dynamics, and timbre. Harmony refers to how melodies interact vertically. Texture describes the number of musical lines and their relationship. Tempo is the speed of a piece, while form refers to a composition's overall structure and layout in sections.
The document outlines Olmo Cornelis' research on opportunities for symbiosis between Western and non-Western musical idioms from 2008-2014. It discusses his background, previous work digitizing an ethnomusicological archive, challenges in accurately describing non-Western music, and a proposed methodology using music information retrieval techniques to objectively analyze ethnic music and inform new compositions blending musical elements from different cultures.
This document provides an overview of sound and audio concepts. It begins with the basic physics of sound, discussing how sound is formed through vibration of air molecules. It then covers types of sound including voice, sound effects, and music used in film. Key audio concepts like intensity, pitch, attack/sustain/decay are explained. The document also discusses modes of listening, sound art, recorded sound, and includes examples of early sound artists. Microphone basics and considerations for achieving realism in recorded sound are also covered.
This document discusses symbolic melodic similarity and methods for representing and comparing melodies symbolically. It presents several approaches for representing melodies, such as sequences of text, graphs, n-grams, and curves. It focuses on an approach where melodies are represented as curves interpolated from note points, and similarity is measured by aligning the curves and comparing the areas between their derivatives. Evaluation results from music information retrieval competitions show this curve alignment approach performs well. The document concludes with a demonstration of melody comparison software implementing these techniques.
This document discusses conceptual metaphors in music through the lens of the Study and Research Group on Musical Metaphors (GERMM). It provides examples of conceptual metaphors that understand musical ideas in terms of other domains, such as architecture, language, and the body. The group aims to study the links between metaphorical language, conceptual metaphors, and our physical and sensory experiences of music through techniques like questionnaires, interviews, motion capture, and neuroimaging. Their research could provide insights into how metaphor, expectation, and embodiment relate to musical meaning and response.
MUSIC OF LOWLAND LUZON-ELEMENTS OF MUSIC.pptxRhiaLopez3
This document discusses the musical elements and characteristics of folk songs from the lowlands of Luzon in the Philippines. It describes the typical melody, rhythm, texture, harmony, dynamics, tempo, and timbre of these folk songs. It also discusses the types of folk songs based on function, including ballads, lullabies, love songs, and work songs. The document provides examples of two specific folk songs, Atin Cu Pung Singsing and Pamulinawen, summarizing their musical elements, form, and function.
Rhythm, tempo, melody, dynamics, harmony, timbre, and form are elements of music discussed in the document. It provides explanations of various musical terms including accent and syncopation in rhythm, harmony as chords that follow each other, and tempo as the basic pace of music. Musical form is described as the organization of musical elements in time that provides structure. The document also discusses techniques used in musical form such as repetition, contrast, and variation.
This document discusses music representation and notation. It begins by motivating the importance of representation for content-based music information retrieval. It then examines different levels of music representation from abstract semantics to concrete notation. The document analyzes the complexities of music notation, particularly western conventional music notation, and how it both enables and challenges computational interpretation. It concludes by emphasizing the need for high-level intelligence to fully interpret music notation and represent its semantics.
Introduction to LC Faceted Vocabularies for Music Resources (August 2018)ALATechSource
This document discusses new Library of Congress vocabularies for describing music resources, including the Library of Congress Genre/Form Terms for Library and Archival Materials (LCGFT), the Library of Congress Medium of Performance Thesaurus (LCMPT), and the Library of Congress Demographic Group Terms (LCDGT). It provides examples of how to search, browse, and apply these terms in bibliographic records to classify genre, medium of performance, audience, and creator/contributor characteristics in a more granular and standardized way than was possible with LC Subject Headings alone. Best practices for using these new terms are still developing.
This lecture provides an overview of musicology and how it relates to the analysis of popular music. It discusses the history of musicology focusing originally on art music, and how it has expanded to include popular music. The goals of the module are introduced as analyzing popular music forms, developing critical skills, and improving writing and presentation abilities. An overview of the course schedule is given along with details on assessments, which include a group presentation and written essay. Various analytical tools and layers involved in analyzing songs, arrangements, and recorded tracks are defined and examples are discussed.
1. The study analyzed bass melodies from Red Hot Chili Peppers albums before and after 1999 to investigate changes in the playing style of bassist Flea over time. Features related to pitch, duration, and note counts were extracted from MIDI data and used to accurately classify melodies as pre- or post-1999.
2. A second study investigated the relationship between verbal impressions of equalized vocal tones (e.g. warm, bright) and parametric equalizer settings. Participants evaluated modified vocal recordings and their responses were used to map impressions to frequency boosts and cuts.
3. The results of both studies could help develop music analysis and generation systems that incorporate stylistic changes over time or allow intuitive equalizer
This document describes an algorithm for automatically generating optimized music playlists that minimize jarring transitions between songs. The algorithm uses a directed graph to represent songs as nodes and possible transitions as weighted edges. It calculates pair-wise transition scores between songs based on the distance between them across 5 dimensions: mood, energy, volume, danceability, and beats per minute. The algorithm performs a depth-first search of the graph to visit as many songs as possible while choosing low-weight edge transitions, generating a playlist with smoother acoustic transitions between songs. It then evaluates the generated playlist by comparing its average transition score to that of a randomly shuffled version.
This document describes an algorithm for automatically generating optimal music playlists that minimize jarring transitions between songs. The algorithm uses a directed graph to represent songs as nodes and possible transitions as weighted edges. It calculates pair-wise transition scores between songs based on the distance between them across 5 dimensions: mood, energy, volume, danceability, and beats per minute. The algorithm performs a depth-first search of the graph to visit as many nodes as possible while using low-weight edges, representing smoother transitions, to generate a playlist. It then evaluates the generated playlist by comparing its average transition score to that of a randomly shuffled version.
Research Skills Musicology Final Session Prior To Easter BreakPaul Carr
This document provides guidance for a 2000-word musicology essay final assignment. Students can analyze a single piece of music or compare two pieces using discussed methodologies, such as Philip Tagg's approach. Alternatively, students can present a contextual analysis of an artist, discussing factors like authenticity, local/global influences, and how the artist relates to other musicians and styles. The essay should include transcriptions, recordings, and references. Assessment will consider the analysis detail, use of examples, and cross-referencing of academic texts. The deadline is May 7th, 2010.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
20 Comprehensive Checklist of Designing and Developing a WebsitePixlogix Infotech
Dive into the world of Website Designing and Developing with Pixlogix! Looking to create a stunning online presence? Look no further! Our comprehensive checklist covers everything you need to know to craft a website that stands out. From user-friendly design to seamless functionality, we've got you covered. Don't miss out on this invaluable resource! Check out our checklist now at Pixlogix and start your journey towards a captivating online presence today.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
4. What is MIR?
• Music Information Retrieval (MIR): the
interdisciplinary science of retrieving information
from music. MIR is a small but growing field of
research with many real-world applications.
• Objective: make the world’s vast store of music
accessible to all.
• The contributing disciplines: computer science,
information retrieval, audio engineering, digital
sound processing, musicology, library science,
cognitive science, psychology, philosophy and law.
6. Music Terms - Pitch & Melody
• Pitch is a particular frequency of sound
• E.g., 440 Hz
• Note is a named pitch by us humans.
• E.g., Western music generally refers to the
440 Hz pitch as A, specifically A4
• Melody is A pattern of pitches
• Only a sound produced electronically can have
only one pitch; all other sounds consist of
multiple pitches.
• The mix of frequencies in a sound results in the
Timbre
7. Music Terms - Timbre
• In music
– The characteristic quality of sound produced by
a particular instrument or voice; tone color.
• In acoustics and phonetics
– The characteristic quality of a sound,
independent of pitch and loudness
– Depends on the relative strengths
of its component frequencies;
– E.g, A4 on a guitar a sound
composed of the following Freq:
440 Hz, 880 Hz, 1320 Hz, 1760 Hz,
etc
10. MDR - Music Identification
• Metadata-based Approach:
– Music identification relies on information about
the content rather than the content itself.
– Ex. TOC
• Content-based Approach:
– Ex. Shazam Service
11. MDR - Music Identification - TOC
• TOC (Table Of Contents): a representation of the
start positions and lengths of the tracks on the disc.
• This feature is highly specific, because it is extremely
rare for different albums to share the same lengths
of tracks in the same order.
• But, slight differences in the generation of CDs, even
from the same source audio material, can produce
different TOCs, which will then fail to match each
other.
• Ex. freedb
12. MDR - Music Identification - Shazam
• Shazam:
a mobile app that recognizes music and TV around
you. (it lets you record up to 15 seconds of the song
you are hearing and then it will tell you everything
you want to know about that song: the artist, the
name of the song, the album, offer you links to
YouTube or to buy the song on iTunes)
13. MDR - Music Identification - Shazam
The Initial Spectrogram
14. MDR - Music Identification - Shazam
• They will store only the intense sounds in the song, the time
when they appear in the song and at which frequency.
The Simplified Spectrogram
15. MDR - Music Identification - Shazam
• To store this in the database in a way in which is efficient to search for a
match (easy to index), they choose some of the points from within the
simplified spectrogram (called “anchor points”) and zones in the vicinity of
them (called “target zone”)
Pairing the anchor point with points in a target zone
16. MDR - Music Identification - Shazam
• For each point in the target zone, they will create a hash that
will be the aggregation of the following:
– F1: the frequency at which the anchor point is located
– F2: the frequency at which the point in the target zone is
located
– T2 - T1: the time difference between the time when the
point in the target zone is located in the song (t2) and the
time when the anchor point is located in the song (t1)
• 64-bit struct, 32 bits for the hash and 32 bits for the time
offset and track ID.
17. MDR - Music Identification - Shazam
How do they find the song based on the recorded sample ?
• Repeat the same fingerprinting to the recorded sample.
• Each hash generated from the sample sound, will be searched
for a match in the database.
• If a match is found you will have:
– The time of the hash from the sample (th1)
– The time of the hash from the song in the database (th2)
• Draw a new graph called scatter graph.
– The horizontal axis (X): th2
– The vertical axis (Y): th1
– The point of intersection of the two occurrence times (th1 and th2)
will be marked with a small circle.
18. MDR - Music Identification - Shazam
• If the graph will contain a lot of pairs of th1‘s and th2‘s from the same
song, a diagonal line will form.
Scatter graph of a matching run
19. MDR - Music Identification - Shazam
• Calculate a difference between th2 and th1 (dth) and they will plot it in
a histogram.
• If there is a match in the graph plotted, then there will be a lot
of dths with the same value.
Histogram of a matching run
20. MDR – Similarity Search
• The concept of similarity is less specific than identity.
• There are many different types of musical similarity.
– Two different performances played from the same
notation
– Same composer
– Same function, for example dances
– Same genre
– Same culture
25. QBH – Ranking evaluation measures
A. Mean Reciprocal Rank (MRR):
MRR = (1/3 + 1/2 + 1)/3 = 11/18 or about 0.61
26. QBH – Ranking evaluation measures
B. Top-X Hit Rate
• The position r of the correct result of the
search is in the first X positions or not.
• Mathematically: r(Qi) ≤ X.
29. Emotions?
• Music is language of emotion.
• Users often want to listen to music that is in a certain category
of emotions or they want to listen to music that brings them
in a certain mood.
• What affect the mood of the song?
– Harmony
– Timbre
– Interpretation
– lyrics
30. Challenging Problem !!
• Ambiguous
– Due to the ambiguities of human emotions.
– Different mood interpretation & perception between
individuals
• Cross disciplinary endeavor
– Signal processing
– Machine learning
– Understanding of auditory perception, psychology, and
music theory.
• Mood may change over its durations
31. Different Methods
Contextual
text
information
• websites
• tags
• lyrics
Content-
based
approaches
• audios
• images
• videos
combining
multiple
feature
domains
• Audio & Lyrics
• Audio & Tags
• Audio & Images
(album covers, artist photos, etc.)
32. Contextual text information
• Web-Documents
– Artist biographies, album reviews, and song
reviews are rich sources of information about
music.
– Collect from the Internet by
• querying search engines
• monitoring MP3 blogs
• crawling a music website
– Can be noisy
33. Mood Representation
Categorical psychometrics
• A set of emotional descriptors (tags)
Scalar/dimensional psychometrics
• Mood can be scaled and measured by a
continuum of descriptors or simple
multidimensional metrics.
• Most noted: two dimensional Valence-Arousal
(V-A) space
36. Valence-Arousal (V-A) space
• Simple, powerful way of thinking about the spectrum
of human emotions.
• Both valence and arousal can be defined as
subjective experiences (Russell, 1989).
– Valiance describes whether the emotion is positive or
negative
– Arousal describes the level of alertness or energy involved
in the emotion.
37. Emotion Recognition Problem
• Multiclass multi label classification or regression
problem
• A music piece
– an entire song
– a section of a song (e.g., chorus, verse)
– a fixed-length clip (e.g., 30-second song snipet)
– a short-term segment (e.g., 1 second )
39. Mood representation - vectors
a single multi-dimensional vector
• Each dimension represents
• a single emotion (e.g., angry).
• or a bi-polar pair of emotions
(e.g., positive/negative).
a time-series of vectors over a
semantic space of emotions
• Track changes in emotional content over the
duration of a piece
40. Mood Representation- Vector Values
• a binary label
– The presence or absence of the emotion
• a real-valued score
– e.g., Likert scale value
– Probability estimate
• A Likert scale is a psychometric scale commonly involved in
research that employs questionnaires. It is the most widely used
approach to scaling responses in survey research
45. Timbre Features
• Musical instruments usually produce sound waves with
frequencies
• The lowest frequency is
– The fundamental frequency f0
– Close relation with pitch
• The second and higher frequencies are
– Called overtones
Editor's Notes
Two well-known examples of music-recommendation systems are Pandora Radio, which is a content-based system, and Last.fm, thought as a metadata-based system.
For example, playing A4 on a guitar will actually result in a sound composed of the following frequencies: 440 Hz, 880 Hz, 1320 Hz, 1760 Hz, etc. The particular strength, or amplitude, of the frequencies results in the timbre.
When a user produces a query Q related to a certain tune A, the QBH system returns a rank of a certain length N in which the tune A is located at position r.
The particular reciprocal rank for that A query is defined as 1/r.
Mean Reciprocal Rank: the mean value of the reciprocal ranks obtained when the system is evaluated with n queries.
By doing this, we can obtain the average of how many times the QBH system retrieves the correct result among the first X positions.