A Deep Neural Network Approach to the LifeCLEF 2014 bird task

Hendrik Vincent Koops, Jan Van Balen, Frans Wiering

    Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademic

    Abstract

    This paper describes the methods that are used in our submission
    to the LifeCLEF 2014 Bird task. A segmentation algorithm is created
    that is capable of segmenting the audio files of the Bird task dataset.
    These segments are used to select relevant Mel-Frequency Cepstral Coefficients
    (MFCC) frames from the MFCC dataset. Three datasets are
    created, 48: containing only the mean MFCC per segment, 96: containing
    the mean and variance of the MFCCs in a segment, and 240: containing
    the mean, variance and the mean of three sections. These dataset are
    shuffled and split in a test and train set to train Deep Neural Networks
    with several topologies, which are capable to classify the segments of the
    datasets. It was found that the best network was capable of correctly
    classifying 73% of the segments. The results of a run from our system
    placed us 6th in the list of 10 participating teams. In a follow-up research
    it is found that shuffling the data before splitting introduces overfitting,
    which can be reduced by not shuffling the datasets prior to splitting, and
    using dropout networks.
    Original languageEnglish
    Title of host publicationCLEF2014 Working Notes
    Subtitle of host publicationSheffield, UK, September 15-18, 2014
    EditorsLinda Cappellato , Nicola Ferro, Martin Halvey , Wessel Kraaij
    Pages634-642
    Publication statusPublished - 2014

    Publication series

    NameCEUR Workshop Proceedings
    Volume1180
    ISSN (Electronic)1613-0073

    Keywords

    • Deep Learning
    • Neural Networks
    • Feature Learning
    • Birdsong Recognition
    • Bioacoustics

    Fingerprint

    Dive into the research topics of 'A Deep Neural Network Approach to the LifeCLEF 2014 bird task'. Together they form a unique fingerprint.

    Cite this