4.2 Article

Feasibility of using convolutional neural networks for individual-identification of wild Asian elephants

期刊

MAMMALIAN BIOLOGY
卷 102, 期 3, 页码 909-919

出版社

SPRINGER HEIDELBERG
DOI: 10.1007/s42991-021-00206-2

关键词

Artificial Intelligence; Computer Vision; Image Classification; Individual-Identification; Machine Learning; Pattern Recognition

类别

资金

  1. U.S. Fish and Wildlife Asian Elephant Conservation Grant program

向作者/读者索取更多资源

Individual identification is crucial for research in behavior, ecology, and conservation. Deep-learning methods, particularly Convolutional Neural Networks, show promise in automating this task. Different CNN models were evaluated for identifying Asian elephants through photographs, with VGG16 performing best on body and face datasets, and Xception excelling on ear dataset.
Individual identification is a basic requirement for research in behavior, ecology and conservation. Photographic records are commonly used in situations where individuals are visually distinct. However, keeping track of identities becomes challenging with increasing population sizes and corresponding datasets. There is growing interest in the potential of deep-learning methods for computer vision to assist with automating this task. Here we apply Convolutional Neural Networks, a popular architecture for Artificial Neural Networks used in image classification, to the problem of identifying individual Asian elephants through photographs. We evaluate the performance of five different types of CNN models used in facial recognition (VGG16, ResNet50, InceptionV3, Xception, and Alexnet), on datasets representing three different feature regions (the full body, face, and ears), trained with two techniques (transfer learning vs. training from scratch) for n = 56 elephants. We tested accuracy in matching the top candidate as well as top five candidates. We found that VGG16 trained with the transfer-learning technique outperformed other models on the body and face datasets with accuracies of 21.34% and 42.35%, respectively, in matching the top candidate. Nevertheless, the best performance was achieved by an Xception model trained from the scratch on the ear dataset, with an accuracy of 89.02% for matching the top candidate and 99.27% for including the correct individual among the top five. However, this impressive level of accuracy was obtained with a dataset of 3816 labeled training images of 56 elephants. There are more than 1000 wild elephants in the population under observation, requiring extensive human effort and skill to initially annotate the images used as training data. Therefore, we consider this approach impractical for monitoring large wild populations. Nevertheless this it could be very useful in record keeping and fraud prevention for large captive elephant populations, as well as monitoring animals that have been rehabilitated and released or moved for management purposes.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.2
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据