Usage
  • 101 views
  • 816 downloads

Pre-processing spectrogram parameters improve the accuracy of bioacoustic classification using convolutional neural networks

  • Author(s) / Creator(s)
  • A variety of automated classification approaches have been developed to extract species detection information from large bioacoustic datasets. Convolutional neural networks (CNNs) are an image classification technique that can be operated on the spectrogram of an audio recording. Using CNNs for bioacoustic classification negates the need for sophisticated feature extraction techniques; however, CNNs may be sensitive to the parameters used to create spectrograms. We used AlexNet to classify spectrograms of audio clips from 19 species of birdsong. We trained and tested AlexNet with the spectrograms and observed that mean classification accuracy ranged from 88.9% to 96.9% depending on the parameters used to create the spectrogram. Classification accuracy was highest when we used a composite of four spectrograms with different combinations of scales for frequency and amplitude. Classification accuracy also varied depending on the FFT window size of the spectrogram. Overall, our results suggest that optimal spectrogram parameters for CNN classification may differ from those used for human visualization or other classification approaches. We suggest that if spectrogram parameters are appropriately selected, classification accuracy similar to current state-of-the-art methods can be achieved using off-the-shelf software and without the need to extract domain-specific features.

  • Date created
    2021-06-01
  • Subjects / Keywords
  • Type of Item
    Article (Published)
  • DOI
    https://doi.org/10.7939/r3-x76p-kh75
  • License
    Attribution-NonCommercial 4.0 International
  • Language
  • Citation for previous publication
    • Knight, E. C., Hernandez, S. P., Bayne, E. M., Bulitko, V., & Tucker, B. V. (2020). Pre-processing spectrogram parameters improve the accuracy of bioacoustic classification using convolutional neural networks. Bioacoustics, 29(3), 337–355. https://doi.org/10.1080/09524622.2019.1606734