Better quantifying inter-annotator variability: A step towards citizen science in underwater passive acoustics

G. Dubus, Maëlle Torterotot, P. V. N. Duc, Julie Béesau, D. Cazau, Olivier Adam

2023 Oceans Cited 5 times

Abstract

Deployments of underwater passive acoustic recorders have been widely used to study marine biodiversity, especially to detect vocal cetaceans. To process the huge amount of data collected, automatic detection and classification methods are necessary. Recently the development of such methods, which includes training and then testing the models, is mainly based on so-called ground-truth labels, obtained by manual annotation of audio files.However, manual annotation is a difficult and time-consuming process because of the large size of the datasets, the large diversity of the sounds, their unfamiliar representation, the variant quality of the acoustic recordings and the variability in human appreciation.These different factors induce non-negligible differences from one annotator to another, and better quantifying and understanding such differences is capital to make progress in machine learning applications.On this topic, the inter-annotator variability is investigated on three multi-annotator annotation campaigns performed on different marine bioacoustics datasets. Each of them gathered more than 10 annotators with different profiles, from novices to field experts, covering different annotation tasks, different geographical areas and varieties of sound classes. From this multi-annotation, this work enhances the understanding of the inter-annotator variability through the kappa-metrics. In a second part, from a grouping method of annotation based on a majority vote, a drastic reduction of the potential errors in the annotation from novice annotators is observed. This last observation enlightens the possibility of using citizen sciences to overcome the lack of annotation, while maintaining a quality of annotation expected by an expert.

BibTeX
@inproceedings{Dubus2023BetterQuantifying,
  author = {Dubus, Gabriel and Torterotot, Maëlle and Duc, Paul Nguyen Hong and Beesau, Julie and Cazau, Dorian and Adam, Olivier},
  booktitle = {OCEANS 2023 - Limerick},
  title = {Better quantifying inter-annotator variability: A step towards citizen science in underwater passive acoustics},
  year = {2023},
  volume = {},
  number = {},
  pages = {1-8},
  abstract = {Deployments of underwater passive acoustic recorders have been widely used to study marine biodiversity, especially to detect vocal cetaceans. To process the huge amount of data collected, automatic detection and classification methods are necessary. Recently the development of such methods, which includes training and then testing the models, is mainly based on so-called ground-truth labels, obtained by manual annotation of audio files.However, manual annotation is a difficult and time-consuming process because of the large size of the datasets, the large diversity of the sounds, their unfamiliar representation, the variant quality of the acoustic recordings and the variability in human appreciation.These different factors induce non-negligible differences from one annotator to another, and better quantifying and understanding such differences is capital to make progress in machine learning applications.On this topic, the inter-annotator variability is investigated on three multi-annotator annotation campaigns performed on different marine bioacoustics datasets. Each of them gathered more than 10 annotators with different profiles, from novices to field experts, covering different annotation tasks, different geographical areas and varieties of sound classes. From this multi-annotation, this work enhances the understanding of the inter-annotator variability through the kappa-metrics. In a second part, from a grouping method of annotation based on a majority vote, a drastic reduction of the potential errors in the annotation from novice annotators is observed. This last observation enlightens the possibility of using citizen sciences to overcome the lack of annotation, while maintaining a quality of annotation expected by an expert.},
  keywords = {Training;Systematics;Annotations;Biological system modeling;Manuals;Acoustics;Recording;marine bioacoustics;passive acoustic monitoring;citizen sciences;inter-annotator agreement;manual annotation},
  doi = {10.1109/OCEANSLimerick52467.2023.10244502},
  issn = {},
  month = {June},
}