Search or add a thesis

Advanced Search (Beta)
Home > Efficient Music Fingerprinting for Music Speech Segregation

Efficient Music Fingerprinting for Music Speech Segregation

Thesis Info

Access Option

External Link

Author

Qazi, Khurram Ashfaq

Program

PhD

Institute

University of Engineering and Technology

City

Taxila

Province

Punjab

Country

Pakistan

Thesis Completing Year

2018

Thesis Completion Status

Completed

Subject

Software Engineering

Language

English

Link

http://prr.hec.gov.pk/jspui/bitstream/123456789/10077/1/Khurram%20Ashfaq%20Qazi_Software%20Engg_2018_PRR.pdf

Added

2021-02-17 19:49:13

Modified

2024-03-24 20:25:49

ARI ID

1676727745461

Similar


Since decade, a lot of music fingerprinting and speech segregation algorithms have exhaled. Music speech segregation includes music identification and followed by speech segregation. This becomes challenging in the presences of the noisy environment and noisy sample case. A rapid development has taken place in the field of multimedia content analysis. Music information retrieval applications increased the emphases on the development of music fingerprinting algorithms. Noise affects the efficiency and accuracy of the audio information retrieval algorithms. This research thesis presents a deep analysis of music fingerprinting and speech segregation algorithms. A novel algorithm is presented for music fingerprinting which is used for efficient speech segregation in which music fingerprinting is performed over a noisy audio sample. This research work proposes a system that performs music fingerprinting in-depth evolving the speech segregation processes in presence of background noise. Noise is removed from the audio signal using layered separation model of the recurrent neural network. Music fingerprinting is performed on the basis of pitch based acoustic features classified using distributed dictionary based features learning model. The classified music is processed for speech segregation after noise removal using layered separation model. Speech is segregated using vocal based acoustic features. Features are classified using improved dictionary based fisher algorithm. Structured based classes are used for the classification process. The systematic evaluation of the proposed system for music fingerprinting and speech segregation produces competitive results for three datasets (i.e. TIMIT, MIR-1K, and MusicBrainz), and the results indicate the strength of the proposed system. The proposed system produces significantly better results when the qualitative and quantitative analysis is carried out over the standard datasets showing the better efficiency of our proposed system from the past systems.
Loading...
Loading...

Similar News

Loading...

Similar Articles

Loading...

Similar Article Headings

Loading...