TY - GEN
T1 - Crowdsourcing Multi-label Audio Annotation Tasks with Citizen Scientists
AU - Cartwright, Mark
AU - Dove, Graham
AU - Méndez, Ana Elisa Méndez
AU - Bello, Juan P.
AU - Nov, Oded
N1 - Publisher Copyright:
© 2019 Copyright held by the owner/author(s).
PY - 2019/5/2
Y1 - 2019/5/2
N2 - Annotating rich audio data is an essential aspect of training and evaluating machine listening systems. We approach this task in the context of temporally-complex urban soundscapes, which require multiple labels to identify overlapping sound sources. Typically this work is crowdsourced, and previous studies have shown that workers can quickly label audio with binary annotation for single classes. However, this approach can be difcult to scale when multiple passes with diferent focus classes are required to annotate data with multiple labels. In citizen science, where tasks are often image-based, annotation eforts typically label multiple classes simultaneously in a single pass. This paper describes our data collection on the Zooniverse citizen science platform, comparing the efciencies of diferent audio annotation strategies. We compared multiple-pass binary annotation, single-pass multi-label annotation, and a hybrid approach: hierarchical multi-pass multi-label annotation. We discuss our fndings, which support using multi-label annotation, with reference to volunteer citizen scientists’ motivations.
AB - Annotating rich audio data is an essential aspect of training and evaluating machine listening systems. We approach this task in the context of temporally-complex urban soundscapes, which require multiple labels to identify overlapping sound sources. Typically this work is crowdsourced, and previous studies have shown that workers can quickly label audio with binary annotation for single classes. However, this approach can be difcult to scale when multiple passes with diferent focus classes are required to annotate data with multiple labels. In citizen science, where tasks are often image-based, annotation eforts typically label multiple classes simultaneously in a single pass. This paper describes our data collection on the Zooniverse citizen science platform, comparing the efciencies of diferent audio annotation strategies. We compared multiple-pass binary annotation, single-pass multi-label annotation, and a hybrid approach: hierarchical multi-pass multi-label annotation. We discuss our fndings, which support using multi-label annotation, with reference to volunteer citizen scientists’ motivations.
KW - Audio annotation
KW - Citizen science
KW - Crowdsourcing
UR - http://www.scopus.com/inward/record.url?scp=85064262497&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85064262497&partnerID=8YFLogxK
U2 - 10.1145/3290605.3300522
DO - 10.1145/3290605.3300522
M3 - Conference contribution
AN - SCOPUS:85064262497
T3 - Conference on Human Factors in Computing Systems - Proceedings
BT - CHI 2019 - Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems
PB - Association for Computing Machinery
T2 - 2019 CHI Conference on Human Factors in Computing Systems, CHI 2019
Y2 - 4 May 2019 through 9 May 2019
ER -