• Title/Summary/Keyword: Bird Sound Classification

Search Result 5, Processing Time 0.02 seconds

Animal Sounds Classification Scheme Based on Multi-Feature Network with Mixed Datasets

  • Kim, Chung-Il;Cho, Yongjang;Jung, Seungwon;Rew, Jehyeok;Hwang, Eenjun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.8
    • /
    • pp.3384-3398
    • /
    • 2020
  • In recent years, as the environment has become an important issue in dealing with food, energy, and urban development, diverse environment-related applications such as environmental monitoring and ecosystem management have emerged. In such applications, automatic classification of animals using video or sound is very useful in terms of cost and convenience. So far, many works have been done for animal sounds classification using artificial intelligence techniques such as a convolutional neural network. However, most of them have dealt only with the sound of a specific class of animals such as bird sounds or insect sounds. Due to this, they are not suitable for classifying various types of animal sounds. In this paper, we propose a sound classification scheme based on a multi-feature network for classifying sounds of multiple species of animals. To do that, we first collected multiple animal sound datasets and grouped them into classes. Then, we extracted their audio features by generating mixed records and used those features for training. To evaluate the effectiveness of our scheme, we constructed an animal sound classification model and performed various experiments. We report some of the results.

Wild Bird Sound Classification Scheme using Focal Loss and Ensemble Learning (Focal Loss와 앙상블 학습을 이용한 야생조류 소리 분류 기법)

  • Jaeseung Lee;Jehyeok Rew
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.29 no.2
    • /
    • pp.15-25
    • /
    • 2024
  • For effective analysis of animal ecosystems, technology that can automatically identify the current status of animal habitats is crucial. Specifically, animal sound classification, which identifies species based on their sounds, is gaining great attention where video-based discrimination is impractical. Traditional studies have relied on a single deep learning model to classify animal sounds. However, sounds collected in outdoor settings often include substantial background noise, complicating the task for a single model. In addition, data imbalance among species may lead to biased model training. To address these challenges, in this paper, we propose an animal sound classification scheme that combines predictions from multiple models using Focal Loss, which adjusts penalties based on class data volume. Experiments on public datasets have demonstrated that our scheme can improve recall by up to 22.6% compared to an average of single models.

Temporal attention based animal sound classification (시간 축 주의집중 기반 동물 울음소리 분류)

  • Kim, Jungmin;Lee, Younglo;Kim, Donghyeon;Ko, Hanseok
    • The Journal of the Acoustical Society of Korea
    • /
    • v.39 no.5
    • /
    • pp.406-413
    • /
    • 2020
  • In this paper, to improve the classification accuracy of bird and amphibian acoustic sound, we utilize GLU (Gated Linear Unit) and Self-attention that encourages the network to extract important features from data and discriminate relevant important frames from all the input sequences for further performance improvement. To utilize acoustic data, we convert 1-D acoustic data to a log-Mel spectrogram. Subsequently, undesirable component such as background noise in the log-Mel spectrogram is reduced by GLU. Then, we employ the proposed temporal self-attention to improve classification accuracy. The data consist of 6-species of birds, 8-species of amphibians including endangered species in the natural environment. As a result, our proposed method is shown to achieve an accuracy of 91 % with bird data and 93 % with amphibian data. Overall, an improvement of about 6 % ~ 7 % accuracy in performance is achieved compared to the existing algorithms.

Bird sounds classification by combining PNCC and robust Mel-log filter bank features (PNCC와 robust Mel-log filter bank 특징을 결합한 조류 울음소리 분류)

  • Badi, Alzahra;Ko, Kyungdeuk;Ko, Hanseok
    • The Journal of the Acoustical Society of Korea
    • /
    • v.38 no.1
    • /
    • pp.39-46
    • /
    • 2019
  • In this paper, combining features is proposed as a way to enhance the classification accuracy of sounds under noisy environments using the CNN (Convolutional Neural Network) structure. A robust log Mel-filter bank using Wiener filter and PNCCs (Power Normalized Cepstral Coefficients) are extracted to form a 2-dimensional feature that is used as input to the CNN structure. An ebird database is used to classify 43 types of bird species in their natural environment. To evaluate the performance of the combined features under noisy environments, the database is augmented with 3 types of noise under 4 different SNRs (Signal to Noise Ratios) (20 dB, 10 dB, 5 dB, 0 dB). The combined feature is compared to the log Mel-filter bank with and without incorporating the Wiener filter and the PNCCs. The combined feature is shown to outperform the other mentioned features under clean environments with a 1.34 % increase in overall average accuracy. Additionally, the accuracy under noisy environments at the 4 SNR levels is increased by 1.06 % and 0.65 % for shop and schoolyard noise backgrounds, respectively.

Environmental Factors Affecting the Start of the Dawn Chorus of Wild Birds and the Differences between Each Species (야생조류 새벽 지저귐 개시에 영향을 주는 환경요인과 종간 차이)

  • Ki, Kyong-Seok;Hong, Suk-Hwan;Gim, Ji-Youn
    • Korean Journal of Environment and Ecology
    • /
    • v.29 no.3
    • /
    • pp.344-352
    • /
    • 2015
  • This study aims to identify the difference in the starting time of the dawn chorus of different species of birds and related environmental conditions. For this study, the bird songs were recorded every 2 hours a day starting from 2 hours before sunrise for a whole spring season (from April to June, 2014) in Korea. The recorded sounds were analyzed to find the starting time of each species' song by sonogram pattern using a sound analyzing program (Adobe Audition CC). To analyze the relationship between the environmental conditions and the time of the birdsong of each species, we collected the data of environmental conditions: daily precipitation, Julian date, time of sunrise and moonrise, temperature data and twilight time from the Korea Meteorological Office. As a result of this study, it was found that there are several statistical correlation between the starting time of the birdsong and environmental conditions (every environmental condition except daily precipitation and time of moonrise). Also, the difference of the starting time among each species were identified. The species were divided into 3 groups (early, mid and late) by their wake-up time. Ring-necked pheasant and yellow-throated bunting were categorized as 'Early' group and all woodpeckers were categorized as'Late' group. This study is significant as it first presented a classification of wild birds in Korea by the starting time of their dawn chorus. However, further studies are required to identify why each species to do their dawn singing at different times.