Show simple item record

Files in this item

Thumbnail

Item metadata

dc.contributor.authorMadhusudhana, Shyam
dc.contributor.authorShiu, Yu
dc.contributor.authorKlinck, Holger
dc.contributor.authorFleishman, Erica
dc.contributor.authorLiu, Xiaobai
dc.contributor.authorNosal, Eva-Marie
dc.contributor.authorHelble, Tyler
dc.contributor.authorCholewiak, Danielle
dc.contributor.authorGillespie, Douglas
dc.contributor.authorŠirović, Ana
dc.contributor.authorRoch, Marie A
dc.date.accessioned2021-07-28T09:30:08Z
dc.date.available2021-07-28T09:30:08Z
dc.date.issued2021-07
dc.identifier275214157
dc.identifierc9f003f8-148c-47fb-acb7-f60b29de85d4
dc.identifier000676307900001
dc.identifier85111982081
dc.identifier.citationMadhusudhana , S , Shiu , Y , Klinck , H , Fleishman , E , Liu , X , Nosal , E-M , Helble , T , Cholewiak , D , Gillespie , D , Širović , A & Roch , M A 2021 , ' Improve automatic detection of animal call sequences with temporal context ' , Journal of the Royal Society Interface , vol. 18 , no. 180 , 20210297 . https://doi.org/10.1098/rsif.2021.0297en
dc.identifier.issn1742-5662
dc.identifier.otherRIS: urn:749B5F2AA55940CFE3965339CA37091E
dc.identifier.otherORCID: /0000-0001-9628-157X/work/97884665
dc.identifier.urihttps://hdl.handle.net/10023/23659
dc.descriptionFunding: This work was supported by the US Office of Naval Research (grant no. N00014-17-1-2867).en
dc.description.abstractMany animals rely on long-form communication, in the form of songs, for vital functions such as mate attraction and territorial defence. We explored the prospect of improving automatic recognition performance by using the temporal context inherent in song. The ability to accurately detect sequences of calls has implications for conservation and biological studies. We show that the performance of a convolutional neural network (CNN), designed to detect song notes (calls) in short-duration audio segments, can be improved by combining it with a recurrent network designed to process sequences of learned representations from the CNN on a longer time scale. The combined system of independently trained CNN and long short-term memory (LSTM) network models exploits the temporal patterns between song notes. We demonstrate the technique using recordings of fin whale (Balaenoptera physalus) songs, which comprise patterned sequences of characteristic notes. We evaluated several variants of the CNN + LSTM network. Relative to the baseline CNN model, the CNN + LSTM models reduced performance variance, offering a 9-17% increase in area under the precision-recall curve and a 9-18% increase in peak F1-scores. These results show that the inclusion of temporal information may offer a valuable pathway for improving the automatic recognition and transcription of wildlife recordings.
dc.format.extent13
dc.format.extent3249808
dc.language.isoeng
dc.relation.ispartofJournal of the Royal Society Interfaceen
dc.subjectBioacousticsen
dc.subjectImproved performanceen
dc.subjectMachine learningen
dc.subjectPassive acoustic monitoringen
dc.subjectRobust automatic recognitionen
dc.subjectTemporal contexten
dc.subjectQA76 Computer softwareen
dc.subjectQH301 Biologyen
dc.subjectDASen
dc.subject.lccQA76en
dc.subject.lccQH301en
dc.titleImprove automatic detection of animal call sequences with temporal contexten
dc.typeJournal articleen
dc.contributor.institutionUniversity of St Andrews. School of Biologyen
dc.contributor.institutionUniversity of St Andrews. Sea Mammal Research Uniten
dc.contributor.institutionUniversity of St Andrews. Scottish Oceans Instituteen
dc.contributor.institutionUniversity of St Andrews. Sound Tags Groupen
dc.contributor.institutionUniversity of St Andrews. Bioacoustics groupen
dc.contributor.institutionUniversity of St Andrews. Marine Alliance for Science & Technology Scotlanden
dc.identifier.doihttps://doi.org/10.1098/rsif.2021.0297
dc.description.statusPeer revieweden


This item appears in the following Collection(s)

Show simple item record