Diversity and bias in audio captioning datasets
Martin Morato, Irene; Mesaros, Annamaria (2021-11-15)
Martin Morato, Irene
Mesaros, Annamaria
Teoksen toimittaja(t)
Font, Frederic
Mesaros, Annamaria
P.W. Ellis, Daniel
Fonseca, Eduardo
Fuentes, Magdalena
Elizalde, Benjamin
DCASE
15.11.2021
Julkaisun pysyvä osoite on
https://urn.fi/URN:NBN:fi:tuni-202202282218
https://urn.fi/URN:NBN:fi:tuni-202202282218
Kuvaus
Peer reviewed
Tiivistelmä
Describing soundscapes in sentences allows better understanding of the acoustic scene than a single label indicating the acoustic scene class or a set of audio tags indicating the sound events active in the audio clip. In addition, the richness of natural language allows a range of possible descriptions for the same acoustic scene. In this work, we address the diversity obtained when collecting descriptions of soundscapes using crowdsourcing. We study how much the collection of audio captions can be guided by the instructions given in the annotation task, by analysing the possible bias introduced by auxiliary information provided in the annotation process. Our study shows that even when given hints on the audio content, different annotators describe the same soundscape using different vocabulary. In automatic captioning, hints provided as audio tags represent grounding textual information that facilitates guiding the captioning output towards specific concepts. We also release a new dataset of audio captions and audio tags produced by multiple annotators for a subset of the TAU Urban Acoustic Scenes 2018 dataset, suitable for studying guided captioning.
Kokoelmat
- TUNICRIS-julkaisut [18531]