Diversity and bias in audio captioning datasets

Tutkimustuotos: KonferenssiartikkeliScientificvertaisarvioitu

2 Lataukset (Pure)


Describing soundscapes in sentences allows better understanding of the acoustic scene than a single label indicating the acoustic scene class or a set of audio tags indicating the sound events active in the audio clip. In addition, the richness of natural language allows a range of possible descriptions for the same acoustic scene. In this work, we address the diversity obtained when collecting descriptions of soundscapes using crowdsourcing. We study how much the collection of audio captions can be guided by the instructions given in the annotation task, by analysing the possible bias introduced by auxiliary information provided in the annotation process. Our study shows that even when given hints on the audio content, different annotators describe the same soundscape using different vocabulary. In automatic captioning, hints provided as audio tags represent grounding textual information that facilitates guiding the captioning output towards specific concepts. We also release a new dataset of audio captions and audio tags produced by multiple annotators for a subset of the TAU Urban Acoustic Scenes 2018 dataset, suitable for studying guided captioning.
OtsikkoProceedings of the 6th Workshop on Detection and Classication of Acoustic Scenes and Events (DCASE 2021)
ToimittajatFrederic Font, Annamaria Mesaros, Daniel P.W. Ellis, Eduardo Fonseca, Magdalena Fuentes, Benjamin Elizalde
ISBN (elektroninen)978-84-09-36072-7
DOI - pysyväislinkit
TilaJulkaistu - 15 marrask. 2021
OKM-julkaisutyyppiA4 Artikkeli konferenssijulkaisussa
TapahtumaDetection and Classication of Acoustic Scenes and Events - , Espanja
Kesto: 15 marrask. 202119 marrask. 2021


ConferenceDetection and Classication of Acoustic Scenes and Events


  • Jufo-taso 0


Sukella tutkimusaiheisiin 'Diversity and bias in audio captioning datasets'. Ne muodostavat yhdessä ainutlaatuisen sormenjäljen.

Siteeraa tätä