Afficher la notice abrégée
dc.contributor.author |
Bouchelaram, Ishrak |
|
dc.contributor.author |
Chita, Ramzi |
|
dc.contributor.author |
Kameche, A. (Promoteur) |
|
dc.date.accessioned |
2022-11-07T12:50:08Z |
|
dc.date.available |
2022-11-07T12:50:08Z |
|
dc.date.issued |
2022-09-25 |
|
dc.identifier.uri |
https://di.univ-blida.dz/jspui/handle/123456789/19960 |
|
dc.description |
ill., Bibliogr. Cote: ma-004-869 |
fr_FR |
dc.description.abstract |
The main purpose of this project is to design an environmental
general audio content description using text, where a system accepts as an
input an audio signal and outputs the textual description of that signal.
This task has drawn lots of attention during the past several years
as a result of quick devolvement of different methods that can provide
captions for a general audio recording. To accomplish the automatic audio
captioning task, we have performed multiple experiments using a Clotho
dataset. Two deep neural networks have been employed in the
construction of our systems Recurrent Neural Network and Gated
Recurrent Unit, along with encoder-decoder architecture and a
combination of feature representations based on audio processing
techniques like Mel Spectrogram and text processing techniques used in
text decoding from word embeddings like one-hot-encoding and BERT.
Keywords: Audio Captioning, Machine Learning, Encoder Decoder Models, Signal Processing, Natural Language Processing. |
fr_FR |
dc.language.iso |
en |
fr_FR |
dc.publisher |
Université Blida 1 |
fr_FR |
dc.subject |
Audio Captioning |
fr_FR |
dc.subject |
Machine Learning |
fr_FR |
dc.subject |
Encoder Decoder Models |
fr_FR |
dc.subject |
Signal Processing |
fr_FR |
dc.subject |
Natural Language Processing |
fr_FR |
dc.title |
ENCODER-DECODER NEURAL NETWORK ARCHITECTURES FOR AUTOMATIC AUDIO CAPTIONING |
fr_FR |
dc.type |
Thesis |
fr_FR |
Fichier(s) constituant ce document
Ce document figure dans la(les) collection(s) suivante(s)
Afficher la notice abrégée