MUSIC RECOGNITION USING DEEP LEARNING
dc.contributor.author | KHANAL, PRADEEP KUMAR | |
dc.contributor.author | KAYASTHA, ACHYUT | |
dc.contributor.author | KHATAKHO, ASHISH | |
dc.date.accessioned | 2023-07-31T06:12:32Z | |
dc.date.available | 2023-07-31T06:12:32Z | |
dc.date.issued | 2022-04 | |
dc.description | In our daily lives, we often listen to songs that we like and enjoy. However, there may be instances when we are in transit or at venues such as clubs or restaurants, and we hear a song playing in the background that catches our attention. We may desire to listen to this song again at a later time, but unfortunately, we may not be aware of the title of the song. | en_US |
dc.description.abstract | In our daily lives, we often listen to songs that we like and enjoy. However, there may be instances when we are in transit or at venues such as clubs or restaurants, and we hear a song playing in the background that catches our attention. We may desire to listen to this song again at a later time, but unfortunately, we may not be aware of the title of the song. As a result, we are unable to locate and listen to it again. Our project, titled “Music Recognition Using Deep Learning,” aims to provide a convenient solution for identifying songs that are heard in various locations. While there are several existing popular applications, such as Shazam, SoundHound, and Google Sound Search, which offer music recognition services, we conducted an in-depth study of papers related to these apps to identify appropriate technologies and algorithms for our project. Our project employs a deep neural network that leverages a contrastive learning approach for the purpose of song recognition. Initially, a large collection of songs is gathered and subjected to signal processing techniques, including Short Time Fourier Transform (STFT), mel filter bank, and decibel scale to generate log mel-spectrograms. These log mel-spectrograms are then fed into the neural network, which is trained to generate a fingerprint for each song at the segment level. These fingerprints are stored in a database. | en_US |
dc.identifier.uri | https://hdl.handle.net/20.500.14540/18844 | |
dc.language.iso | en | en_US |
dc.publisher | I.O.E. Pulchowk Campus | en_US |
dc.subject | music recognition, | en_US |
dc.subject | deep learning in music, | en_US |
dc.subject | contrastive learning | en_US |
dc.title | MUSIC RECOGNITION USING DEEP LEARNING | en_US |
dc.type | Report | en_US |
local.academic.level | Bachelor | en_US |
local.affiliatedinstitute.title | Pulchowk Campus | en_US |
local.institute.title | Institute of Engineering | en_US |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- pradeep k khnal et al. be report electronics apr 2022.pdf
- Size:
- 1.84 MB
- Format:
- Adobe Portable Document Format
- Description:
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 1.71 KB
- Format:
- Item-specific license agreed upon to submission
- Description: