Please use this identifier to cite or link to this item: https://hdl.handle.net/1959.11/58355
Full metadata record
DC FieldValueLanguage
dc.contributor.authorShahinfar, Salehen
dc.contributor.authorMeek, Paulen
dc.contributor.authorFalzon, Gregoryen
dc.date.accessioned2024-04-15T05:28:04Z-
dc.date.available2024-04-15T05:28:04Z-
dc.date.issued2020-
dc.identifier.citationEcological Informatics, v.57, p. 1-16en
dc.identifier.issn1878-0512en
dc.identifier.issn1574-9541en
dc.identifier.urihttps://hdl.handle.net/1959.11/58355-
dc.description.abstract<p>Deep learning (DL) algorithms are the state of the art in automated classification of wildlife camera trap images. The challenge is that the ecologist cannot know in advance how many images per species they need to collect for model training in order to achieve their desired classification accuracy. In fact there is limited empirical evidence in the context of camera trapping to demonstrate that increasing sample size will lead to improved accuracy.</p> <p>In this study we explore in depth the issues of deep learning model performance for progressively increasing per class (species) sample sizes. We also provide ecologists with an approximation formula to estimate how many images per animal species they need for certain accuracy level a priori. This will help ecologists for optimal allocation of resources, work and efficient study design.</p> <p>In order to investigate the effect of number of training images" seven training sets with 10, 20, 50, 150, 500, 1000 images per class were designed. Six deep learning architectures namely ResNet-18, ResNet-50, ResNet-152, DnsNet-121, DnsNet-161, and DnsNet-201 were trained and tested on a common exclusive testing set of 250 images per class. The whole experiment was repeated on three similar datasets from Australia, Africa and North America and the results were compared. Simple regression equations for use by practitioners to approximate model performance metrics are provided. Generalizes additive models (GAM) are shown to be effective in modelling DL performance metrics based on the number of training images per class, tuning scheme and dataset.</p> <p>Overall, our trained models classified images with 0.94 accuracy (ACC), 0.73 precision (PRC), 0.72 true positive rate (TPR), and 0.03 false positive rate (FPR). Variation in model performance metrics among datasets, species and deep learning architectures exist and are shown distinctively in the discussion section. The ordinary least squares regression models explained 57%, 54%, 52%, and 34% of expected variation of ACC, PRC, TPR, and FPR according to number of images available for training. Generalised additive models explained 77%, 69%, 70%, and 53% of deviance for ACC, PRC, TPR, and FPR respectively.</p> <p>Predictive models were developed linking number of training images per class, model, dataset to performance metrics. The ordinary least squares regression and Generalised additive models developed provides a practical toolbox to estimate model performance with respect to different numbers of training images.</p>en
dc.languageenen
dc.publisherElsevier BVen
dc.relation.ispartofEcological Informaticsen
dc.title“How many images do I need?” Understanding how sample size per class affects deep learning model performance metrics for balanced designs in autonomous wildlife monitoringen
dc.typeJournal Articleen
dc.identifier.doi10.1016/j.ecoinf.2020.101085en
local.contributor.firstnameSalehen
local.contributor.firstnamePaulen
local.contributor.firstnameGregoryen
local.profile.schoolSchool of Science and Technologyen
local.profile.schoolSchool of Environmental and Rural Scienceen
local.profile.schoolSchool of Science and Technologyen
local.profile.emailsshahinf@une.edu.auen
local.profile.emailpmeek5@une.edu.auen
local.profile.emailgfalzon2@une.edu.auen
local.output.categoryC1en
local.record.placeauen
local.record.institutionUniversity of New Englanden
local.publisher.placeThe Netherlandsen
local.identifier.runningnumber101085en
local.format.startpage1en
local.format.endpage16en
local.peerreviewedYesen
local.identifier.volume57en
local.contributor.lastnameShahinfaren
local.contributor.lastnameMeeken
local.contributor.lastnameFalzonen
dc.identifier.staffune-id:sshahinfen
dc.identifier.staffune-id:pmeek5en
dc.identifier.staffune-id:gfalzon2en
local.profile.orcid0000-0002-1989-9357en
local.profile.roleauthoren
local.profile.roleauthoren
local.profile.roleauthoren
local.identifier.unepublicationidune:1959.11/58355en
dc.identifier.academiclevelAcademicen
dc.identifier.academiclevelAcademicen
dc.identifier.academiclevelAcademicen
local.title.maintitle“How many images do I need?” Understanding how sample size per class affects deep learning model performance metrics for balanced designs in autonomous wildlife monitoringen
local.relation.fundingsourcenoteFunding for this project was provided by the Australian Government Department of Agriculture and Water Resources through the eTechnology Hub – Utilising Technology to Improve Pest Management Effectiveness and Enhance Welfare Outcomes project.en
local.output.categorydescriptionC1 Refereed Article in a Scholarly Journalen
local.search.authorShahinfar, Salehen
local.search.authorMeek, Paulen
local.search.authorFalzon, Gregoryen
local.uneassociationYesen
local.atsiresearchNoen
local.sensitive.culturalNoen
local.year.published2020en
local.fileurl.closedpublishedhttps://rune.une.edu.au/web/retrieve/6072b8a8-881a-4262-99d1-9a506f2d9e96en
local.subject.for20203003 Animal production not elsewhere classifieden
local.subject.seo2020tbden
local.profile.affiliationtypeUNE Affiliationen
local.profile.affiliationtypeUNE Affiliationen
local.profile.affiliationtypeUNE Affiliationen
local.date.moved2024-04-15en
Appears in Collections:Journal Article
School of Environmental and Rural Science
School of Science and Technology
Files in This Item:
1 files
File SizeFormat 
Show simple item record

SCOPUSTM   
Citations

101
checked on Jul 6, 2024
Google Media

Google ScholarTM

Check

Altmetric


Items in Research UNE are protected by copyright, with all rights reserved, unless otherwise indicated.