Université Blida 1

DEEP CO-TRAINING FRAMEWORK FOR SEMI-SUPERVISED AUDIO TAGGING

Afficher la notice abrégée

dc.contributor.author Cheifa, Ikram
dc.contributor.author Yakhlef, Hadjer ( Promotrice)
dc.contributor.author Diffallah, Zhor ( promotrice)
dc.date.accessioned 2022-12-13T12:14:12Z
dc.date.available 2022-12-13T12:14:12Z
dc.date.issued 2022
dc.identifier.uri https://di.univ-blida.dz/jspui/handle/123456789/20463
dc.description ill., Bibliogr. Cote: ma-004-879 fr_FR
dc.description.abstract Audio tagging, also known as Sound Event Recognition, is concerned with the development of systems that are able to recognize sound events. A sound event is perceived as a separate individual entity that we can name and recognize, such as helicopter, glass breaking, baby crying, speech, etc. Considerable attention has been geared towards audio tagging for various applications, such as information retrieval, music tagging, and acoustic monitoring. The general framework for audio tagging usually involves two major steps: feature extraction and classification. Clearly, obtaining well-annotated, strongly labeled data is an expensive and time-consuming process. Therefore, a large portion of recent development has been devoted to effectively using weakly labeled data extracted from websites like Youtube, Freesound, or Flickr. Various semi-supervised learning approaches have been proposed in the literature. We can cite Mean Teacher, Pseudo Labeling, Mix Match, and most recently, Deep Co-training. The purpose of this project consists of devising an audio tagging system within the semi-supervised learning paradigm, specifically the Deep Co-training framework. Such systems essentially use both labeled and unlabeled audio data. In addition, our system is trained on two different datasets :Urban8k and Environmental Sound Classification, based on a deep residual neural network (ResNet) and a wide residual neural network (WideResNet). We supported our analysis and discussion with numerous statistical tests to analyze and compare our results. We have investigated the impact of differentiating the supervised ratio on the system’s performance and have tested the impact of various variants of DCT systems based on different adversarial attacks. The results demonstrate the efficacy of the Deep Co-training SSL strategy that significantly boosts the overall performance. Keywords: Audio Tagging, Semi-supervised learning, Deep Co-training, Feature Extraction, Statistical Tests. fr_FR
dc.language.iso en fr_FR
dc.publisher Université Blida 1 fr_FR
dc.subject Audio Tagging fr_FR
dc.subject Semi-Supervised Learning fr_FR
dc.subject Deep Co-training fr_FR
dc.subject Feature Extraction fr_FR
dc.subject Statistical Tests fr_FR
dc.title DEEP CO-TRAINING FRAMEWORK FOR SEMI-SUPERVISED AUDIO TAGGING fr_FR
dc.type Thesis fr_FR


Fichier(s) constituant ce document

Ce document figure dans la(les) collection(s) suivante(s)

Afficher la notice abrégée

Chercher dans le dépôt


Recherche avancée

Parcourir

Mon compte