Veuillez utiliser cette adresse pour citer ce document : https://di.univ-blida.dz/jspui/handle/123456789/20463
Affichage complet
Élément Dublin CoreValeurLangue
dc.contributor.authorCheifa, Ikram-
dc.contributor.authorYakhlef, Hadjer ( Promotrice)-
dc.contributor.authorDiffallah, Zhor ( promotrice)-
dc.date.accessioned2022-12-13T12:14:12Z-
dc.date.available2022-12-13T12:14:12Z-
dc.date.issued2022-
dc.identifier.urihttps://di.univ-blida.dz/jspui/handle/123456789/20463-
dc.descriptionill., Bibliogr. Cote: ma-004-879fr_FR
dc.description.abstractAudio tagging, also known as Sound Event Recognition, is concerned with the development of systems that are able to recognize sound events. A sound event is perceived as a separate individual entity that we can name and recognize, such as helicopter, glass breaking, baby crying, speech, etc. Considerable attention has been geared towards audio tagging for various applications, such as information retrieval, music tagging, and acoustic monitoring. The general framework for audio tagging usually involves two major steps: feature extraction and classification. Clearly, obtaining well-annotated, strongly labeled data is an expensive and time-consuming process. Therefore, a large portion of recent development has been devoted to effectively using weakly labeled data extracted from websites like Youtube, Freesound, or Flickr. Various semi-supervised learning approaches have been proposed in the literature. We can cite Mean Teacher, Pseudo Labeling, Mix Match, and most recently, Deep Co-training. The purpose of this project consists of devising an audio tagging system within the semi-supervised learning paradigm, specifically the Deep Co-training framework. Such systems essentially use both labeled and unlabeled audio data. In addition, our system is trained on two different datasets :Urban8k and Environmental Sound Classification, based on a deep residual neural network (ResNet) and a wide residual neural network (WideResNet). We supported our analysis and discussion with numerous statistical tests to analyze and compare our results. We have investigated the impact of differentiating the supervised ratio on the system’s performance and have tested the impact of various variants of DCT systems based on different adversarial attacks. The results demonstrate the efficacy of the Deep Co-training SSL strategy that significantly boosts the overall performance. Keywords: Audio Tagging, Semi-supervised learning, Deep Co-training, Feature Extraction, Statistical Tests.fr_FR
dc.language.isoenfr_FR
dc.publisherUniversité Blida 1fr_FR
dc.subjectAudio Taggingfr_FR
dc.subjectSemi-Supervised Learningfr_FR
dc.subjectDeep Co-trainingfr_FR
dc.subjectFeature Extractionfr_FR
dc.subjectStatistical Testsfr_FR
dc.titleDEEP CO-TRAINING FRAMEWORK FOR SEMI-SUPERVISED AUDIO TAGGINGfr_FR
dc.typeThesisfr_FR
Collection(s) :Mémoires de Master

Fichier(s) constituant ce document :
Fichier Description TailleFormat 
Cheifa Ikram.pdf5,58 MBAdobe PDFVoir/Ouvrir


Tous les documents dans DSpace sont protégés par copyright, avec tous droits réservés.