Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
General:  hacktoberfest Type:  dataset Data Domain:  audio
51fcad7fb0
Added Dataset
2 years ago
8956f8472f
Data description added
2 years ago
51fcad7fb0
Added Dataset
2 years ago
51fcad7fb0
Added Dataset
2 years ago
51fcad7fb0
Added Dataset
2 years ago
8956f8472f
Data description added
2 years ago
8956f8472f
Data description added
2 years ago
8956f8472f
Data description added
2 years ago
Storage Buckets
Data Pipeline
Legend
DVC Managed File
Git Managed File
Metric
Stage File
External File

README.md

You have to be logged in to leave a comment. Sign In

Bird-Audio-Detection-challenge

image

Detecting bird sounds in audio is an important task for automatic wildlife monitoring, as well as in citizen science and audio library management. The current generation of software tools require manual work from the user: to choose the algorithm, to set the settings, and to post-process the results. This is holding bioacoustics back in embracing its “big data” era: let’s make this better!

In collaboration with the IEEE Signal Processing Society we propose a research data challenge for you to create a robust and scalable bird detection algorithm. We offer new datasets collected in real live bioacoustics monitoring projects, and an objective, standardised evaluation framework – and prizes for the strongest submissions.

The subset of dataset (freefield1010) is uploaded to DagsHub, enabling you to preview the dataset before downloading it.

Main links for the original Bird Audio Detection challenge (2016/2017):

  1. View the results online
  2. Download the data
  3. Read the academic papers:

There was also a discussion forum used during the challenge.

Task description and datasets

The task is to design a system that, given a short audio recording, returns a binary decision for the presence/absence of bird sound (bird sound of any kind). The output can be just “0” or “1”, but we encourage weighted/probability outputs in the continuous range [0,1] for the purposes of evaluation. For the main assessment we will use the well-known “Area Under the ROC Curve” (AUC) measure for classification performance.

**Dataset 1: Field recordings (freefield1010) ** [Dataset is available at DAGsHub]

Our first dataset comes from freefield1010 – a collection of over 7,000 excerpts from field recordings around the world, gathered by the FreeSound project, and then standardised for research. This collection is very diverse in location and environment, and for the BAD Challenge we have newly annotated it for the presence/absence of birds.

Dataset 2: Crowdsourced dataset (Warblr)

Our second dataset comes from a UK bird-sound crowdsourcing research spinout called Warblr. From this initiative we have 10,000 ten-second smartphone audio recordings from around the UK. The audio totals around 44 hours duration. The audio will be published by Warblr under a Creative Commons licence. The audio covers a wide distribution of UK locations and environments, and includes weather noise, traffic noise, human speech and even human bird imitations. It is directly representative of the data that is collected from a mobile crowdsourcing initiative.

Dataset 3: Remote monitoring dataset (Chernobyl)

Our third dataset comes from the Natural Environment Research Council (NERC)-funded TREE (Transfer-Exposure-Effects) research project, which is deploying unattended remote monitoring equipment in the Chernobyl Exclusion Zone (CEZ). This academic investigation into the long-term effects of the Chernobyl accident on local ecology is led by Dr Wood. The project has captured approximately 10,000 hours of audio to date (since June 2015). The audio covers a range of birds and includes weather, large mammal and insect noise sampled across various CEZ environments, including abandoned village, grassland and forest areas.

The task specification gives technical details of how participants will submit their analyses.

Please note: we wish to encourage the development of generalisable and “tuning-free” methods by ensuring that testing data is recorded under different conditions than the publicly-available data. This helps ensure that the challenge addresses the needs identified in recent research for methods that can operate in unseen conditions. The training data will come from freefield1010 and Warblr, and the testing data from a mixture of data, predominantly from the Chernobyl dataset.

Data downloads

Download the development data:

and the testing data:

(Thanks to Figshare and archive.org for hosting data downloads.)

Directory

The dataset is small (6.33 GB) and simple to navigate as it has only single folder audio files. The ff1010bird_labels.csv file contains label [0,1] against audio files. There are 7690 excerpts from field recordings around the world, gathered by the FreeSound project, and then standardised for research. This collection is very diverse in location and environment, and for the BAD Challenge we have newly annotated it for the presence/absence of birds.

Note: We have only uploaded freefield1010 to DVC server, to download entire dataset check the links in above sections.

<root directory>
    |
    .- README.md
    |
    .- LICENSE.txt
    |
    .- ff1010bird_labels.csv
    |
    .- Data/
          |
          .- 55.wav
          |
          .- 87.wav
          |
          .- 99.wav
          |
          ...

How the challenge will work

The interface for submitting results files will be a Kaggle-like interface designed for academic users. The main criterion for evaluating submissions will be the AUC, but we also aim to encourage computationally-efficient methods by recording the computation time taken.

As well as the public online challenge, we are also organising:

  • A conference special session on the topic, with contributions from international experts.
  • A journal special issue on the topic.

Prizes: There will be one prize of £500 and one prize of €500, to be awarded to the strongest submissions. One prize will go to the submission with the best overall performance on the testing data (NB: to be eligible we will also require the teams to provide documentation / source code for their method – since, after all, the aim is to advance the state of the art). One prize will be the judges’ award, which will go to the submission which the judges feel has made a notable contribution to the topic. The organisers reserve the right to make the final decision on prize allocations, but please ask if you have any queries.

RESULTS

In revalidating the testing set, we examined those items with the strongest mismatch between manual and automatic detection, to determine which was in error: 500 presumed negative and 1,243 presumed positive items. This showed inter-rater disagreement in 16.6% of such cases predominantly, the most ambiguous cases with barely audible bird sounds with amplitude close to the noise threshold. Note that this percentage is not representative of disagreement across the whole dataset, but only on the “controversial” cases. We also observed that a strong mismatch according to the automatic detectors did not necessarily imply human mislabelling: some perceptually obvious data items could be consistently misjudged by algorithms. We will discuss algorithm errors further below. Through revalidation, the inter-rater reliability, measured via the AUC, was measured as 96.7%. This value provides an approximate upper limit for machine performance since it reflects the extent of ambiguity in the data according to human listeners’ perception.

The two baseline classifiers gave relatively good performance on the development data, the strongest at over 85% AUC in matched conditions, but generalised poorly. The simpler GMM-based baseline classifier showed consistently lower results than the more advanced classifier, as expected. It also showed strong resistance to overfitting in the sense that its performance on its training set was a very good predictor of its performance on a matched-conditions testing set. However, this was not sufficient to allow it to generalise to mismatched conditions, in which its performance degraded dramatically (Figure 1). The more advanced baseline classifier also degraded when tested in mismatched conditions, though to a lesser extent, attaining 74.8% AUC in the main evaluation.

image

Figure 1

Open in figure viewerPowerPoint

Final scores attained by the highest performing submission for each team. Error bars are estimated by bootstrap sampling. “skfl” and “smacpy” shown near the start are the baseline systems

ACKNOWLEDGEMENTS

We thank Nick Beresford and Sergey Gashchak for their help with capturing the Chernobyl soundscape recordings, Paul Kendrick for preparation and annotation of Chernobyl data, Luciana Barçada for annotation of Chernobyl data, and Julien Ricard for programming and administering the challenge submission website. We also thank the many challenge participants for their enthusiastic effort. The Chernobyl data collection was undertaken within the TREE project (www.ceh.ac.uk/TREE), which is funded by the Natural Environment Research Council, the Environment Agency and Radioactive Waste Management Ltd.

AUTHORS’ CONTRIBUTIONS

D.S., M.D.W., Y.S., and H.G. designed the data challenge study and its analysis; D.S., M.D.W., and H.P. provided datasets; H.G. led the creation of the challenge submission website; D.S. and H.P. performed tests of machine learning systems; D.S., H.G., and H.P. analysed the results; D.S. led the writing of the manuscript, with some sections by H.P. All authors contributed critically to the drafts and gave final approval for publication.

DATA ACCESSIBILITY

Organisers

  • Dr Dan Stowell, Queen Mary University of London, London, UK.
  • Pr. Hervé Glotin, Scaled Acoustic BioDiversity Dept, Univ. of Toulon & Inst. Univ. de France.
  • Pr. Yannis Stylianou, Computer Science Dept, University of Crete.
  • Dr Mike Wood, University of Salford, Greater Manchester, UK.

Queen_Mary,_University_of_London_logo Univ Toulon Logo Univ Crete Logo Univ Salford LogoMadics logo


Original Dataset: [data labels] • [audio files ]

Photo by Zdeněk Macháček on Unsplash


This open source contribution is part of DagsHub x Hacktoberfest

Tip!

Press p or to see the previous file or, n or to see the next file

About

Detecting bird sounds in audio is an important task for automatic wildlife monitoring, as well as in citizen science and audio library management.

http://machine-listening.eecs.qmul.ac.uk/bird-audio-detection-challenge/
Collaborators 2

Comments

Loading...