Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Bhattacharyya Parameter of Monomial Codes for the Binary Erasure Channel: From Pointwise to Average Reliability
Next Article in Special Issue
An Evaluation of the Factors Affecting ‘Poacher’ Detection with Drones and the Efficacy of Machine-Learning for Detection
Previous Article in Journal
Managing the Cloud Continuum: Lessons Learnt from a Real Fog-to-Cloud Deployment
Previous Article in Special Issue
Behaviour Classification on Giraffes (Giraffa camelopardalis) Using Machine Learning Algorithms on Triaxial Acceleration Data of Two Commonly Used GPS Devices and Its Possible Application for Their Management and Conservation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Wildlife Monitoring on the Edge: A Performance Evaluation of Embedded Neural Networks on Microcontrollers for Animal Behavior Classification

by
Juan P. Dominguez-Morales
1,2,3,4,*,
Lourdes Duran-Lopez
1,2,3,4,
Daniel Gutierrez-Galan
1,2,3,4,
Antonio Rios-Navarro
1,2,3,4,
Alejandro Linares-Barranco
1,2,3,4 and
Angel Jimenez-Fernandez
1,2,3,4
1
Robotics and Tech. of Computers Lab, Universidad de Sevilla, 41012 Seville, Spain
2
Escuela Técnica Superior de Ingeniería Informática (ETSII), Universidad de Sevilla, 41012 Seville, Spain
3
Escuela Politécnica Superior, Universidad de Sevilla, 41012 Seville, Spain
4
Smart Computer Systems Research and Engineering Lab (SCORE), Research Institute of Computer Engineering (I3US), Universidad de Sevilla, 41012 Seville, Spain
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(9), 2975; https://doi.org/10.3390/s21092975
Submission received: 11 March 2021 / Revised: 16 April 2021 / Accepted: 22 April 2021 / Published: 23 April 2021
(This article belongs to the Special Issue Sensors and Artificial Intelligence for Wildlife Conservation)

Abstract

:
Monitoring animals’ behavior living in wild or semi-wild environments is a very interesting subject for biologists who work with them. The difficulty and cost of implanting electronic devices in this kind of animals suggest that these devices must be robust and have low power consumption to increase their battery life as much as possible. Designing a custom smart device that can detect multiple animal behaviors and that meets the mentioned restrictions presents a major challenge that is addressed in this work. We propose an edge-computing solution, which embeds an ANN in a microcontroller that collects data from an IMU sensor to detect three different horse gaits. All the computation is performed in the microcontroller to reduce the amount of data transmitted via wireless radio, since sending information is one of the most power-consuming tasks in this type of devices. Multiples ANNs were implemented and deployed in different microcontroller architectures in order to find the best balance between energy consumption and computing performance. The results show that the embedded networks obtain up to 97.96% ± 1.42% accuracy, achieving an energy efficiency of 450 Mops/s/watt.

1. Introduction

Monitoring wild or semi-wild animals is a useful tool for biologists and ethologists to understand their behavior, habits, possible diseases, interaction with other animals and the environment, and movement patterns. To learn and understand such details regarding wildlife, it is necessary to obtain information about movement patterns, behavior, and other biometric data such as the body temperature and heart rate of the animal during long periods of time. Thus, several studies have focused on the design and implementation of animal monitoring systems.
One of the most complex and expensive activities of animal monitoring is capturing the animal itself in order to place the tracking device onto it and ensure the device is fixed to the animal and operational for the required time. The complexity of this task requires these devices to meet a series of requirements, such as being lightweight, fault-tolerant, and low-power consuming, which make their design and implementation a technological challenge. Thus, providing biologists with a lightweight and long-battery-life device which can process a large amount of information and access it online is not an easy task.
Traditionally, animal monitoring has been a task that experts have carried out through direct observation. However, due to the presence of humans in the same territory, the observed animals do not behave as they would naturally, and data collection is limited to daytime hours [1]. In order to increase the observation time and reduce human presence, different strategies have been implemented to observe animal behavior and their interaction with the environment. One of such strategies is camera traps, which consists in the installation of cameras in the environment where the animals under study usually go to feed or rest. These cameras start recording when movement is detected on the scene and store the information in the camera. Then, the recorded video has to be recovered on site and the information is post-processed using different algorithms [2,3] to classify the animal species and obtain extra information. This strategy is minimally invasive, since it is not necessary to capture the animal in order to install any device onto it. However, it does not allow the monitoring of each animal individually.
To increase the monitoring time, it is necessary to attach a device to the animal in order to collect data that can be processed and obtain information about the animal’s behavior. This device can host some sensors, such as Global Positioning System (GPS) and Inertial Measurement Unit (IMU), among others, which generate a vast amount of raw data that can be processed by machine learning algorithms; this allows classifying some patterns which help experts to understand the behavior of the studied animal [4,5]. In the case that several animals, which can interact with each other and move in the same environment, need to be monitored, there exist other strategies based on Wireless Sensor Network (WSN). These devices can be attached to the animals and share information between them. The information is sent to the coordinator node, which transmits it to another system that is usually in the cloud [6,7]. This kind of network normally uses two types of wireless connections: one for short range, such as Bluetooth or ZigBee, and another one for large range or Internet connection, such as WiFi, GPRS, or LoRA [8]. Recent studies [9,10,11,12,13,14] have proposed collar devices consisting of a GPS module and an IMU (or an accelerometer only) for animal behavior recognition. In [9], the authors present a device which collects information from a 3-axis accelerometer and a GPS. The information obtained from 5 different behaviors (foraging, ruminating, traveling, resting, and “other”), which was stored on a card, was then used to train and evaluate a decision tree offline, achieving an 85.5% accuracy. In [10], a device, which hosts an IMU, was used to collect information and perform 5-class behavior classification of cows (grazing, walking, ruminating, resting, and “other”). A set of binary time series classifiers were trained and validated offline, achieving an 85% accuracy. In [11], they used 3-axis neck-mounter accelerometers for dairy cattle behavior classification using a decision tree. Three different behaviors were considered (eating, rumination, and lying), for which the authors reported 99.2% accuracy after training and validating the classifier in Matlab. In [12], the authors presented a system based on combining leg-mounted and collar-mounted accelerometers. The system classifies between four different behaviors (walking, feeding, lying, and standing) using a Random Forest algorithm with an 94.3% accuracy in an offline manner. In [13], the authors compared two commercial devices that provide tri-axial acceleration data and GPS information. These devices are used to obtain data related to different behaviors from giraffes, which are then used to train and evaluate a machine learning algorithm (Random Forest), achieving a mean accuracy of 88.4% over six different behaviors when using cross validation. Finally, in [14], different offline machine learning algorithms weree compared over data corresponding to four behaviors obtained from accelerometers mounted on sheep. Among the different algorithms, the best result was achieved with a Support Vector Machine (76.9% accuracy).
All these aforementioned devices store the information generated by the sensors and, after collecting it, it is processed in a local computer, where the classifier is trained and tested. In order to perform a real-time classification of the acquired data, two possible approaches could be followed. The first way consists in sending the raw information, through a radio transmitter device, to another system for post-processing; and the second one, called edge-computing, consists in processing the information in the device itself and sending the new resulting information [15,16]. The processing can be done using statistical models like [15,16] or using Embedded Neural Network (ENN) algorithms like [17]. This computing paradigm, which calls for processing the data at the edge of the device [18], has several advantages, such as real-time analysis, low latency, security, and reduced data delivery, which can be beneficial in the field of animal monitoring. One of the most expensive tasks in terms of energy consumption is the transmission of information through the radio of the monitoring device. Thus, if the rate of data to be sent is reduced, the battery life of the device is increased, which would prolong the time that the animal could be monitored without changing the battery. To reduce the sending rate, the data should be processed in the device itself. As previously commented, some devices use Artificial Neural Networks (ANNs) to process the information obtained by the sensors of the device, but this type of algorithm is power-hungry and requires exhaustive computation. To reduce the complexity of these algorithms, their parameters are quantified without significantly reducing their accuracy [19,20]. This technique allows the use of these algorithms in embedded systems using an affordable energy consumption for their purpose.
This paper presents an analysis and study of the execution of an ENN, taking as input the data generated by an IMU in different Cortex-M architectures for use in an embedded system. Different comparisons were performed between the different studied microcontrollers in terms of power consumption, execution time and performance of different ENNs implemented in them.

2. Materials and Methods

2.1. Wearable Data Acquisition Platform for Wildlife

In this work, a full custom wearable device was used and validated in real-test scenarios (Figure 1a). The aim of this device, also called collar, was to collect precise information from the animal on which it was attached by using multiple devices. A MinIMU-9V2 IMU, which consists of a LSM303DLHC 3-axis accelerometer, a L3GD20 3-axis gyroscope, and a 3-axis magnetometer, was used to measure the animal’s movement dynamics. By means of an I2C interface, it was possible to read nine independent rotation, acceleration, and magnetic measurements with 12-bit resolution for each sample and with a sample rate of 33 samples/s.
A GPS module was used along with the IMU module in order to tag each sample with the location and the time in which the sample was obtained. The periodical measures of each sensor were carried out using a low power microcontroller (STM32L152 (STMicroelectronics, Geneva, Switzerland)), which was powered using a four AAA battery pack (1.5 V, 1155 mAh each). The wearable prototype also includes an XBee module (XBee PRO S2B) for transmitting the acquired data through a wireless communication to a computer in order to be stored in a database. To ensure the correct dataset acquisition, and due to the fact that capturing a semi-wild animal is very expensive and difficult, a microSD card was added to the device in order to store the read samples and avoid data loss due to a bad wireless connection. Two different horses from completely different breeds, places were considered in this work for acquiring the data, which was done during two different seasons of the year.

2.2. Dataset

The IMU sensors presented in Section 2.1 were used to collect experimental data from three different horse gaits: motionless, walking, and trotting. A total of 6841 samples were obtained for the motionless behavior, 6842 for walking, and 11,512 for trotting. These data were obtained after placing the collar device in a vertical position at the right part of the horse’s head (Figure 1b), following the recommendations of experts biologists that supported us throughout this study [21].
For this study, two different datasets were considered to train and validate the Neural Network (NN). On the one hand, the first dataset consisted in the raw data obtained from the accelerometer sensor, since it was demonstrated that it is the most relevant IMU sensor for classifying between motionless, walking, and trotting [21]. On the other hand, the raw data, collected from the accelerometer, gyroscope, and magnetometer sensors, were pre-processed using a Kalman filter [22] to obtain pitch, roll, and yaw values. These pre-processed samples were used as the second dataset. In [21], the authors proved that the Kalman algorithm improves the accuracy between these three classes when training a simple NN (one hidden layer with less than 30 neurons). However, no comparison between the two kinds of data was performed for more complex networks.

2.3. Neural Network Models

In this work, several ANN architectures were trained and tested for each dataset (see Section 2.2). Three neurons were set for the input layer. For the hidden layers (HLs), different experiments were carried out, with models having between one and three layers, each of them with 10, 20, 30, 50, 100, 200, 300, and 500 neurons. Sigmoid activation functions and batch normalization was used after each of the aforementioned layers. Finally, a Softmax decision layer with three neurons receives the output from the last fully-connected layer and predicts the corresponding class.
K-fold stratified cross-validation was performed to measure the generalization ability of the models. This technique consisted in dividing the datasets presented in Section 2.2 into 5 sets ( K = 5 ), each of them with a proportional number of samples per class. For each fold, the networks were trained using four of the five sets (80% of the dataset) for 50,000 epochs, and were validated using the remaining set (20% of the dataset). This way, for each experiment, the networks were trained and validated a total of five times with different data. Therefore, summing up all the different possible combinations, the number of models that were trained and tested were: 2 datasets × 3 HLs × 8 different numbers of neurons in the HL× 5 cross-validation sets, which produced a total of 240 models. All of these models were trained and tested using TensorFlow (https://www.tensorflow.org, accessed on 23 April 2021) [23] and Keras (https://keras.io, accessed on 23 April 2021) [24]. Adam was used as optimizer, with a learning rate of 0.01 and a batch size of 64 samples. The class imbalance was taken into account using Keras’ class_weight parameter, which automatically weighs each class based on the number of training samples used. The final results for a specific experiment were presented as the mean accuracy and their standard deviation calculated over the five cross-validation folds.

2.4. Full Integer Quantization

TensorFlow Lite (https://www.tensorflow.org/lite, accessed on 23 April 2021) is an open-source deep learning framework for on-device inference using pre-trained models. It provides different tools for converting and running TensorFlow and Keras models on embedded devices and microcontrollers. Some of these tools are dedicated to the quantization of pre-trained models by reducing the precision of the weights of the NN and its operations, leading to better performance and lighter model size, sacrificing a small loss of precision and accuracy of the classification.
Different post-training quantization techniques are already implemented in TensorFlow Lite, such as dynamic range quantization, float16 quantization, and full integer quantization (FIQ) (https://www.tensorflow.org/lite/performance/post_training_quantization, accessed on 23 April 2021) [25]. In this work, FIQ is studied, since it is optimized for deploying models on microcontrollers. This method consists in reducing all model math to 8-bit integer values, reducing the peak memory usage and improving the latency of the classification on the microcontroller. Following the documentation provided in TensorFlow Lite, this method achieves 4× smaller models with 2 × 3 × speedup.

2.5. Integrated Development Environment

STM32CubeIDE (https://www.st.com/en/development-tools/stm32cubeide.html, accessed on 23 April 2021) is a multi-OS Integrated Development Environment (IDE) for C/C++ development, with peripheral configuration, code generation, code compilation, and debug features for STM32 microcontrollers. Hundreds of plugins can be integrated within STM32CubeIDE in order to extend the IDE’s functionalities, including STM32CubeMX, which is a standalone GUI-based software tool for initializing and configuring a project for a specific microcontroller and platform.
Within this set of tools, STM32Cube.AI (https://www.st.com/content/st_com/en/stm32-ann.html, accessed on 23 April 2021) can be found. STM32Cube.AI is an extension pack that is fully integrated into STM32CubeMX and allows mapping and running pre-trained ANNs on the broad STM32 microcontroller family. It allows for a fast and automatic conversion of pre-trained ANNs into optimized code that can run on a microcontroller. This tool also reports feedback on the performance of the converted NN in the microcontroller, comparing it with the original model in terms of accuracy, error, and other useful metrics.

2.6. Cortex-M Microcontrollers

To evaluate the ENN performance from different points of view, two 32-bit Cortex-M processors were used. Firstly, the NN was deployed in a low-power Cortex-M4 processor, specifically a STM32L475 (https://www.st.com/en/microcontrollers-microprocessors/stm32l475vg.html, accessed on 23 April 2021), which works at 80 MHz. Then, an advanced Cortex-M7 (STM32F746 (https://www.st.com/content/st_com/en/products/microcontrollers-microprocessors/stm32-32-bit-arm-cortex-mcus/stm32-high-performance-mcus/stm32f7-series/stm32f7x6/stm32f746vg.html, accessed on 23 April 2021)) with a clock of 216 MHz and a superscalar architecture was used. STM32L475 and STM32F476 have 1 Mbytes of flash memory, and 128 and 320 Kbytes of RAM memory, respectively. These data are summarized in Table 1. The aim of choosing both microcontrollers is to offer a criterion in terms of power consumption and execution time for the same ANNs between the two architectures.
Since the executed code in both microcontrollers is the same, there is no difference in the NN complexity, accuracy, or even in the confusion matrix. Therefore, with similar inputs, the same output is obtained. However, there is a considerable difference in terms of power consumption and execution time.

3. Results

3.1. Cross-Validation Results

The results obtained from each ANN model for the two datasets after training and validating them using TensorFlow and Keras are presented in Figure 2 and Figure 3. These results are the average of the 5 cross-validation folds, with their corresponding standard deviation. The values from which the figure was generated are also presented in Table A1. The results for each cross-validation fold for all the experiments that were carried out are detailed in Table A2. The aforementioned table also includes the mean precision, recall, F1-score, and balanced accuracy per fold. As can be seen, for the classification of the three studied horse gaits, the accelerometer data are more representative than the data obtained after processing the output of the IMU with the Kalman filter. However, to achieve better results with the accelerometer data, a higher number of HLs in the model and neurons per HL have to be used, which directly affects the power consumption of the hardware platform in which the model is deployed, as well as the processing time needed for every input sample.
For each dataset and number of HLs, the best models in terms of average accuracy across the number of neurons in the HL were selected for further implementation and analysis in different hardware platforms (see Section 2.6). These models are highlighted in bold in Table A1 and correspond to 1 HL with 30 neurons (73.57% ± 0.83% accuracy), 2 HLs with 200 neurons each (87.36% ± 8.57% accuracy), and 3 HLs with 500 neurons each (97.96% ± 1.42% accuracy) for the accelerometer dataset. For the Kalman dataset, the models selected are: 1 HL with 100 neurons (83.76% ± 5.37% accuracy), 2 HLs with 200 neurons each (90.73% ± 0.71% accuracy), and 3 HLs with 300 neurons each (91.92% ± 0.45% accuracy).
The already-trained selected models obtained with Keras (.h5) were then converted to TensorFlow Lite (.tflite) models in order to integrate them in the different microcontrollers and hardware platforms presented in Section 2.6. Moreover, a full integer quantization was also performed to the converted models, which, as presented in Section 2.4, allows achieving smaller and faster versions of these in terms of memory size and performance, making them more appropriate for being deployed in a microcontroller.

3.2. Hardware Integration Results

The selected models mentioned in the previous Section 3.1 were deployed on two different platforms (Section 2.6) using STM32Cube.AI (Section 2.5). Different metrics were considered in order to compare original models trained in TensorFlow and Keras with the TensorFlow Lite and full integer quantized ones deployed on the microcontrollers. Two of the selected models could not be deployed on the different hardware platforms due to their limitations on Flash memory. These correspond to TensorFlow Lite versions of the selected models with 3 HLs.
Figure 4 presents the X-cross accuracy and the accuracy obtained from the different selected models after deploying them in the two microcontrollers considered in this work. These two parameters represent the difference in terms of accuracy with the model trained in TensorFlow and Keras, that is, before converting them to the TensorFlow Lite format and quantizing them, as well as implementing them on the hardware platforms. Therefore, an X-cross accuracy of 100% means that the model has the same accuracy as the original model. As was previously described, both microcontrollers provide the same X-cross accuracy, since the code executed is the same. Therefore, these results are reported for the deployed ANN, and not for its execution on a specific processor. Figure 4 shows that all the different TensorFlow Lite models achieve the same results as their software counterparts. However, when performing the full integer quantization, the X-cross accuracy is lower, and thus the accuracy decreases. According to the data presented in Figure 4, X-cross accuracy is 100% in TFL conversions, whereas in FIQ conversions, this value ranges between 94% and 84%. As a benefit of FIQ ANNs having lower accuracy, they have better memory requirements. The improvement in execution time depends on network complexity, as we discuss next. According to specific applications, a balance between TFL and FIQ has to be achieved when looking for better performance in terms of classification, execution time, resources, and power demand.
Figure 5 shows the execution time for each of the proposed models, comparing between the two hardware platforms in which they were deployed. From the reported results, it can clearly be seen that the quantized model allows a faster execution of the ANN, since the complexity of the model was reduced in this process. As expected, the greater the number of HLs in the ANN, the longer the time it requires to perform a prediction. On average, the STM32F746 microcontroller needs 3.6 times less execution time to perform a classification with respect to STM32L475. If we compare the execution time between TFL models with FIQ models, different results can be found depending on the ANN size in both microcontrollers. For 1 HL networks, FIQ models are 70% slower than TFL models, but for 2 HL, FIQ models are 30% faster than TFL models.
Figure 6 compares the different models considered in this approach in terms of Mops/s/Watt ( 10 6 operations per second, per watt). This value is widely used when comparing the performance of embedded devices and hardware platforms. The plot shows that, in terms of performance, STM32L475 achieves better results (more than 2 times better than those of STM32F746), being able to execute more operations per demanded watt. Again, differences between 1 HL and 2 HL models can be found when comparing TFL and FIQ. For 1 HL, TFL models are 75% more efficient than FIQ models, whereas for 2 HL models, FIQ are 25% more efficient than TFL.
Figure 7 shows the memory usage in terms of flash and RAM memories, which are the same for both microcontrollers. Regarding the flash memory demand, TFL models need, on average, 3.4 times more flash memory than FIQ. This is due to the fact that TFL models store the ANN weights as 4-byte floating point numbers, while FIQ uses 1 byte per weight, and some overhead is added due to the network structure, which must be data-representation independent. Regarding RAM memory, FIQ demands 1.8 times more memory than TFL on average. This disparity could be related to the internal memory organization and data packaging.
Finally, the battery consumption was estimated for each microcontroller in order to estimate battery life. The results obtained are presented in Table 2. The battery considered in this study is a LiSOCL2, with a capacity of 3400 mAh. It was suggested by STM32CubeMX taking the smallest compatible battery based on the maximum continuous current demanded by the microcontroller. This battery has a weight of 50 grams and, therefore, it is suitable for being placed on horses. Our experiment consisted in the estimation of battery life for the best case (1 HL NN), midpoint case (2 HL NN) and worst case (3 HL NN) when considering an execution rate of 10 predictions per second. With low complex NNs, the STM32L475 can be powered for more than 1 year, but, in the case of the STM32F746, this number is reduced to 3 months. With medium complexity NNs, this estimations decays approximately to 65% and 35% of the original time, respectively. Finally, for higher complexity networks, the estimated battery life is 3 months for the STM32L475 microcontroller, and 15 days for the STM32F746.
Table A3 contains all the numerical results used to obtain the figures that are shown in this section.

4. Discussion

When comparing Kalman and raw-accelerometer datasets, the former achieves better results in lower complex networks (smaller number of HLs and neurons). On the other hand, if we look for the best results in terms of accuracy, these are obtained with the accelerometer data when using a NN with a greater number of layers. Even though having less neurons and layers could also mean having a lower power consumption and a longer battery life, the power consumption of calculating the Kalman filter per input data was not considered in this work, since we only focused on the NN.
Regarding the different hardware platforms considered in this study, the Cortex-M7 takes 3.6 times less execution time to perform a prediction. On the other hand, the low power Cortex-M4 architecture needs half of the power consumption to perform the same classification, with a battery life six times bigger on average. Making a decision between them strongly depends on the target application. In our specific case, since the system is designed for wildlife monitoring, the low-power solution of the STM32L475 is more suitable for this kind of application, which is also applicable to general wearable devices.
An interesting comparison can also be found between TFL and FIQ models. Generally, TFL models achieve better accuracies, although with larger networks. Contrary to TFL, FIQ models demand less resources and are faster. As a counterpart, FIQ models are less efficient in terms of energy consumption per prediction. Summarizing, in the case of wildlife monitoring, choosing between TFL and FIQ models is not clear. Researchers should look for the balance between accuracy, battery life, and microcontroller’s capabilities in order to find the most appropriate approach depending on the task.
In this work we analyzed and tested a prototype particularly developed for semi-wild horses, since it is a simpler and more convenient case study than others that include smaller and wilder animals in terms of capturing them and inducing them to perform a set of gaits. Testing the prototype system on horses was the recommended scenario and approach by the expert biologists that supported this project. If this approach is to be used for other semi-wild or wild animals, their habits should be studied and those of interest should be recorded and used to train a neural network in a supervised manner. A recent study [26] proposed a novel approach for using data from captive individuals to infer wildlife behavior, which could be taken as an inspiration for future works.
In previous state-of-the-art solutions, such as [9,10,11,12,13,14], authors evaluated the classifier in a local computer without providing on-device or cloud-based real-time animal behavior classification. This fact restricts their implementation in a real scenario, since predicting behavioral patterns of the animals after removing the monitoring device from them and processing the stored information is not the optimal use case. Our proposed system is capable of performing real-time online classification, allowing expert biologists to access the data as soon as it is predicted by the collar devices. In terms of battery life and power consumption, the aforementioned studies report a battery life of 12–14 days [9], 14 days [10], 7 days [12], and 17 days [13], respectively. When using the lower power consumption microcontroller from the two options that were considered in this work, we obtain an estimated battery life of 88 days in the worst case, which is 5 times more battery life than [13], taking into consideration that, in our work, we also include the processing of the data and the prediction from the ENN in this estimation. However, if a midpoint case is considered, the estimated battery life is more than 15 times higher than the one from [13].
Even though the proposed system was evaluated using a 5-fold cross-validation scheme, which demonstrates a good generalization of the model, using only two horses could be seen as a limitation of this study. Future works will consider gathering a bigger dataset with different horses in order to have more variation in the obtained data.

5. Conclusions

In this work, different NNs for classifying between three different horse gait patterns were trained and tested using two different datasets (Kalman and raw accelerometer data) obtained from an IMU sensor placed on horses. A total of 24 different ANN models per dataset were trained using a 5-fold cross-validation scheme in order to compare them in terms of complexity and accuracy. The best models were then converted to microcontroller-friendly TensorFlow Lite (TFL) models and quantized (FIQ) in order to reduce the complexity of the network. These two versions of the aforementioned models were deployed on two different microcontrollers (STM32L475 and STM32F746) for further analysis in terms of performance, power consumption, energy efficiency, and battery life. The proposed system improves the battery life of current state-of-the-art solutions by more than five times in the worst case scenario. Furthermore, recent works present animal behavior classification systems where the inference is performed outside of the collar device, either by sending the raw sensor information to a computer or by recovering the SD card from the collar after some days. On the other hand, the edge-computing system presented in this work is able to perform on-device real-time animal behavior classification, reducing data transmissions and, thus, reducing power consumption, together with providing biologists access to the data anytime.
The results show that raw accelerometer data achieve better accuracy results for larger ANNs (around 98% accuracy for three different horse gaits). When embedded on the microcontroller, FIQ models obtain lower classification results and higher RAM memory usage, while benefiting from lower prediction time and flash memory usage. The comparison between the two microcontrollers and the different parameters considered in this study was analyzed. The choice of which microcontroller, model, number of layers, neurons per layer, and dataset to use depends on the particular application, and researchers must select the best one according to their restrictions (e.g., maximizing either battery life or accuracy).

Author Contributions

Conceptualization, J.P.D.-M., L.D.-L., D.G.-G., A.R.-N., and A.J.-F.; methodology, J.P.D.-M., L.D.-L., and A.J.-F.; software, J.P.D.-M. and L.D.-L.; validation, D.G.-G., A.R.-N., and A.J.-F.; formal analysis, J.P.D.-M., L.D.-L., D.G.-G., A.R.-N., and A.J.-F.; investigation, J.P.D.-M., L.D.-L., and A.J.-F.; resources, A.J.-F. and A.L.-B.; data curation, J.P.D.-M., L.D.-L., D.G.-G., and A.R.-N.; writing—original draft preparation, J.P.D.-M., L.D.-L., D.G.-G., A.R.-N., and A.J.-F.; writing—review and editing, J.P.D.-M., L.D.-L., D.G.-G., A.R.-N., A.J.-F., and A.L.-B.; visualization, J.P.D.-M., L.D.-L., D.G.-G., A.R.-N., A.J.-F., and A.L.-B.; supervision, A.J.-F. and A.L.-B.; project administration, A.L.-B.; funding acquisition, A.L.-B. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by Spanish Agencia Estatal de Investigación (AEI) project MINDROB: “Percepción y Cognición Neuromórfica para Actuación Robótica de Alta Velocidad (PID2019-105556GB-C33/AEI/10.13039/501100011033)”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The authors would like to thank the Spanish Agencia Estatal de Investigación (AEI) for supporting this work. The work of Daniel Gutierrez-Galan was supported by a Formación de Personal Investigador Scholarship from the Spanish Ministry of Education, Culture and Sport.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANNArtificial Neural Network
ENNEmbedded Neural Network
FIQFull Integer Quantization
GPSGlobal Positioning System
HLHidden Layer
IDEIntegrated Development Environment
IMUInertial Measurement Unit
NNNeural Network
TFLTensorFlow Lite
WSNWireless Sensor Network

Appendix A

Table A1. Average of the results obtained over the different cross-validation folds for each of the studied configurations. The best results for each dataset and number of layers are highlighted in bold.
Table A1. Average of the results obtained over the different cross-validation folds for each of the studied configurations. The best results for each dataset and number of layers are highlighted in bold.
DatasetNo.
HL
No. Neurons
10203050100200300500
ModelAccel173.13 ± 0.4872.81 ± 1.1173.57 ± 0.8372.96 ± 1.2972.49 ± 1.9770.61 ± 3.0270.43 ± 2.9669.82 ± 3.61
272.93 ± 1.7773.98 ± 0.3774.53 ± 0.5474.96 ± 0.6184.01 ± 10.0687.36 ± 8.5782.89 ± 1.0784.43 ± 1.64
373.99 ± 0.3574.32 ± 0.4175.07 ± 0.7977.24 ± 1.5987.67 ± 5.1097.88 ± 0.6397.34 ± 1.9297.96 ± 1.42
Kalman173.59 ± 0.6575.60 ± 3.3480.35 ± 4.3280.37 ± 3.7383.76 ± 5.3777.24 ± 5.6574.76 ± 4.0569.64 ± 3.50
280.21 ± 1.9284.18 ± 1.1885.31 ± 1.8488.60 ± 0.6890.53 ± 0.5090.73 ± 0.7190.07 ± 0.4489.11 ± 1.02
383.36 ± 0.6685.40 ± 1.0587.88 ± 0.6490.91 ± 0.6391.51 ± 0.6091.81 ± 0.1891.92 ± 0.4591.55 ± 0.69
Table A2. Results obtained for the different cross-validation folds for each of the studied configurations.
Table A2. Results obtained for the different cross-validation folds for each of the studied configurations.
Accuracy per Fold (%)Metrics (Mean) (%)
DataHLNeurons12345AccuracyPrecisionRecallF1-ScoreBalanced
Accuracy
Accel.11072.5273.4073.1573.8872.7473.1473.3673.1473.0575.24
2072.0873.9170.9873.6873.4372.8272.9872.8272.6974.96
3072.0273.9574.5373.6173.7773.5873.7573.5873.4875.63
5073.4674.0072.9973.8870.4872.9673.1472.9672.8675.02
10071.7668.9874.2374.2373.2772.4972.6872.4972.3174.73
20066.8174.2074.1869.0868.7870.6170.8170.6170.2573.25
30067.1474.2068.5273.8268.5170.4470.8370.4470.1173.11
50066.7674.4374.0767.0366.8569.8370.0569.8369.4572.55
21073.2874.5174.2573.0869.5672.9473.3072.9472.8775.03
2074.3174.3474.2073.6173.4573.9874.2873.9873.9675.90
3074.3975.0375.1373.6174.5374.5474.8374.5474.5276.41
5074.8975.0175.9373.9875.0274.9775.2074.9774.9376.81
10096.1876.4876.6474.3596.4384.0184.0584.0183.9985.02
20080.2898.3480.4280.4297.3887.3787.3787.3787.3588.11
30083.0883.7182.4581.0884.1782.9082.9182.9082.8883.76
50082.3287.0283.7883.5585.5184.4484.4684.4484.4385.51
31073.5174.3974.1273.6574.3073.9974.3473.9973.9775.94
2074.4174.9974.2073.7074.3174.3274.6574.3274.3076.24
3074.6676.1275.6573.8475.1175.0875.3775.0875.0576.91
5077.7578.0378.2374.0778.1277.2477.4377.2477.2278.79
10097.4282.9184.3787.0286.6687.6887.7287.6887.6788.35
20098.6096.9297.8898.5397.5197.8997.8997.8997.8998.03
30098.5393.6797.3699.1598.0097.3497.3597.3497.3597.55
50095.1498.8798.7498.8298.2397.9697.9797.9697.9698.11
Kalman11074.5773.5673.7973.5472.5173.5973.5373.5973.0874.77
2077.7476.4475.1779.2069.4775.6075.4675.6075.1576.85
3083.475.9883.1884.8874.3180.3580.2480.3580.1581.61
5073.7080.7080.6985.2381.5580.3880.2980.3880.1681.52
10088.2288.4074.9787.2180.0383.7783.7683.7783.6084.82
20076.2088.2775.2973.8472.6077.2477.2377.2476.8078.38
30072.8280.1670.7579.1170.9974.7674.9074.7674.2575.77
50073.3070.3263.0769.8171.7369.6570.3069.6568.2671.16
21081.4779.4581.5076.7681.9080.2280.1180.2280.0681.50
2084.8385.4482.5682.9585.1484.1884.1284.1884.1085.34
3082.8888.0683.7686.4085.4885.3285.2885.3285.2386.37
5089.3587.9788.5087.7689.4288.6088.6088.6088.5689.43
10090.1991.5190.4390.4790.0990.5490.5390.5490.5091.28
20091.1490.9890.6689.4191.4990.7490.7290.7490.7091.47
30090.2689.3789.9790.7390.0290.0790.0790.0790.0390.83
50088.2689.7687.5789.6590.3189.1189.0989.1189.0689.89
31084.0883.8583.0783.6082.2283.3783.3483.3783.2784.53
2084.5487.0784.0585.8785.4885.4085.4085.4085.3286.55
3087.2587.9487.2387.9989.0087.8887.8987.8887.8288.84
5091.8391.1090.0891.2690.3290.9290.9290.9290.8891.70
10091.7191.1990.7191.4292.5391.5191.5291.5191.4892.20
20091.5391.9091.7692.0991.8191.8291.8391.8291.7992.50
30092.5791.1992.1691.7991.9091.9291.9391.9291.9092.56
50090.9391.3792.2490.7392.4891.5591.5491.5591.5292.26
Table A3. Results obtained for each of the implemented models on the two microcontrollers considered. The complexity of the models in terms of number of operations is also presented.
Table A3. Results obtained for each of the implemented models on the two microcontrollers considered. The complexity of the models in terms of number of operations is also presented.
MicrocontrollerClock Freq.
(MHz)
DataHLModelFlash
(KB)
RAM
(KB)
Complexity
(ops)
Current
(mA)
Power
(mW)
Energy
(mJ)
Processing time
(ms)
CPU cyclescycles/MACAcc
(%)
l2rX-Cross acc
(%)
Mops/s/Watt
STM32L47580Accel1TFLite1.140.15234463010.732.15.93850.1851479423.4874.530.732473100106.087
Quant.0.4023440.33007835110.732.111.26710.3512808480.0167.000.80861689.1931.152
2TFLite164.141.594569010.732.1185.76275.78746296510.1398.340.172436100245.958
Quant.42.622.154248110.732.1145.25254.5253619888.5279.250.62425486.03292.463
3TFLite-------------
Quant.500.045.3850759110.732.11125.811235.07228057595.5380.470.6056288.15450.858
Kalman1TFLite3.60.425781189010.732.114.63760.4563644219.2887.210.478239100129.119
Quant.1.211.0898110.732.123.52930.7335868059.8268.570.77191783.8241.692
2TFLite164.141.594569010.732.1186.01955.79546357910.1593.860.316368100245.619
Quant.42.622.154248110.732.1148.9444.643711768.7487.880.46295494.27285.214
3TFLite-------------
Quant.182.863.2318458110.732.1491.964615.32612260696.6488.400.43929290.34375.191
STM32F746216Accel1TFLite1.140.153846630130.1390.321.85680.0561213319.2674.530.73427310028.823
Quant.0.4023440.330078351130.1390.338.24940.0982111560.1667.000.80861689.199.176
2TFLite164.141.5945690130.1390.3588.57241.5083258147.1398.340.17243610077.628
Quant.42.622.1542481130.1390.3439.08751.1252429195.7279.250.62425486.0396.748
3TFLite-------------
Quant.500.045.38507581130.1390.32998.67497.68316595653.2780.470.6056288.15169.268
Kalman1TFLite3.60.4257811890130.1390.352.69050.1352912915.4187.210.47823910035.869
Quant.1.211.08981130.1390.380.79210.2074465645.5268.570.77191783.8212.142
2TFLite164.141.5945690130.1390.3590.52391.5133268347.1593.860.31636810077.371
Quant.42.622.1542481130.1390.3435.57481.1162411285.6887.880.46295494.2797.528
3TFLite-------------
Quant.182.863.23184581130.1390.31360.58583.4867530584.0888.40.43929290.34135.662

References

  1. Gaynor, K.M.; Hojnowski, C.E.; Carter, N.H.; Brashares, J.S. The influence of human disturbance on wildlife nocturnality. Science 2018, 360, 1232–1235. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Verma, G.K.; Gupta, P. Wild Animal Detection Using Deep Convolutional Neural Network. In Proceedings of the 2nd International Conference on Computer Vision & Image Processing; Chaudhuri, B.B., Kankanhalli, M.S., Raman, B., Eds.; Springer: Singapore, 2018; pp. 327–338. [Google Scholar]
  3. Nguyen, H.; Maclagan, S.J.; Nguyen, T.D.; Nguyen, T.; Flemons, P.; Andrews, K.; Ritchie, E.G.; Phung, D. Animal Recognition and Identification with Deep Convolutional Neural Networks for Automated Wildlife Monitoring. In Proceedings of the 2017 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Tokyo, Japan, 19–21 October 2017; pp. 40–49. [Google Scholar] [CrossRef]
  4. Hammond, T.T.; Springthorpe, D.; Walsh, R.E.; Berg-Kirkpatrick, T. Using accelerometers to remotely and automatically characterize behavior in small animals. J. Exp. Biol. 2016, 219, 1618–1624. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Leoni, J.; Tanelli, M.; Strada, S.C.; Berger-Wolf, T. Ethogram-based automatic wild animal monitoring through inertial sensors and GPS data. Ecol. Inform. 2020, 59, 101112. [Google Scholar] [CrossRef]
  6. Ayele, E.D.; Das, K.; Meratnia, N.; Havinga, P.J.M. Leveraging BLE and LoRa in IoT network for wildlife monitoring system (WMS). In Proceedings of the 2018 IEEE 4th World Forum on Internet of Things (WF-IoT), Singapore, 5–8 February 2018; pp. 342–348. [Google Scholar] [CrossRef]
  7. Ayele, E.D.; Meratnia, N.; Havinga, P.J.M. Towards a New Opportunistic IoT Network Architecture for Wildlife Monitoring System. In Proceedings of the 2018 9th IFIP International Conference on New Technologies, Mobility and Security (NTMS), Paris, France, 26–28 February 2018; pp. 1–5. [Google Scholar] [CrossRef] [Green Version]
  8. Sornin, N.; Luis, M.; Eirich, T.; Kramp, T.; Hersent, O. LoRaWAN Specifications; LoRa Alliance: San Ramon, CA, USA, 2015. [Google Scholar]
  9. González, L.A.; Bishop-Hurley, G.J.; Handcock, R.N.; Crossman, C. Behavioral classification of data from collars containing motion sensors in grazing cattle. Comput. Electron. Agric. 2015, 110, 91–102. [Google Scholar] [CrossRef]
  10. Smith, D.; Rahman, A.; Bishop-Hurley, G.J.; Hills, J.; Shahriar, S.; Henry, D.; Rawnsley, R. Behavior classification of cows fitted with motion collars: Decomposing multi-class classification into a set of binary problems. Comput. Electron. Agric. 2016, 131, 40–50. [Google Scholar] [CrossRef]
  11. Tamura, T.; Okubo, Y.; Deguchi, Y.; Koshikawa, S.; Takahashi, M.; Chida, Y.; Okada, K. Dairy cattle behavior classifications based on decision tree learning using 3-axis neck-mounted accelerometers. Anim. Sci. J. 2019, 90, 589–596. [Google Scholar] [CrossRef] [PubMed]
  12. Tran, D.-N.; Nguyen, T.; Khanh, P.C.P.; Trana, D.-T. An IoT-based Design Using Accelerometers in Animal Behavior Recognition Systems. IEEE Sens. J. 2021. [Google Scholar] [CrossRef]
  13. Brandes, S.; Sicks, F.; Berger, A. Behaviour Classification on Giraffes (Giraffa camelopardalis) Using Machine Learning Algorithms on Triaxial Acceleration Data of Two Commonly Used GPS Devices and Its Possible Application for Their Management and Conservation. Sensors 2021, 21, 2229. [Google Scholar] [CrossRef] [PubMed]
  14. Fogarty, E.; Swain, D.; Cronin, G.; Moraes, L.; Trotter, M. Behaviour classification of extensively grazed sheep using machine learning. Comput. Electron. Agric. 2020, 169, 105175. [Google Scholar] [CrossRef]
  15. Debauche, O.; Mahmoudi, S.; Mahmoudi, S.A.; Manneback, P.; Bindelle, J.; Lebeau, F. Edge Computing for Cattle Behavior Analysis. In Proceedings of the 2020 Second International Conference on Embedded Distributed Systems (EDiS), Oran, Algeria, 3 November 2020; pp. 52–57. [Google Scholar] [CrossRef]
  16. Caria, M.; Schudrowitz, J.; Jukan, A.; Kemper, N. Smart farm computing systems for animal welfare monitoring. In Proceedings of the 2017 40th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 22–26 May 2017; pp. 152–157. [Google Scholar] [CrossRef]
  17. Arablouei, R.; Currie, L.; Kusy, B.; Ingham, A.; Greenwood, P.L.; Bishop-Hurley, G. In-situ classification of cattle behavior using accelerometry data. Comput. Electron. Agric. 2021, 183, 106045. [Google Scholar] [CrossRef]
  18. Shi, W.; Cao, J.; Zhang, Q.; Li, Y.; Xu, L. Edge Computing: Vision and Challenges. IEEE Internet Things J. 2016, 3, 637–646. [Google Scholar] [CrossRef]
  19. Vanhoucke, V.; Senior, A.; Mao, M.Z. Improving the speed of neural networks on CPUs. In Proceedings of the Deep Learning and Unsupervised Feature Learning Workshop, NIPS 2011, Granada, Spain, 16–17 December 2011. [Google Scholar]
  20. Krishnamurthy, A.K.; Ahalt, S.C.; Melton, D.E.; Chen, P. Neural networks for vector quantization of speech and images. IEEE J. Sel. Areas Commun. 1990, 8, 1449–1457. [Google Scholar] [CrossRef]
  21. Gutierrez-Galan, D.; Dominguez-Morales, J.P.; Cerezuela-Escudero, E.; Rios-Navarro, A.; Tapiador-Morales, R.; Rivas-Perez, M.; Dominguez-Morales, M.; Jimenez-Fernandez, A.; Linares-Barranco, A. Embedded neural network for real-time animal behavior classification. Neurocomputing 2018, 272, 17–26. [Google Scholar] [CrossRef]
  22. Welch, G.; Bishop, G. An Introduction to the Kalman Filter. 1995. Available online: https://perso.crans.org/club-krobot/doc/kalman.pdf (accessed on 23 April 2021).
  23. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  24. Gulli, A.; Pal, S. Deep Learning with Keras; Packt Publishing Ltd.: Birmingham, UK, 2017. [Google Scholar]
  25. Dokic, K.; Martinovic, M.; Mandusic, D. Inference speed and quantisation of neural networks with TensorFlow Lite for Microcontrollers framework. In Proceedings of the 2020 5th South-East Europe Design Automation, Computer Engineering, Computer Networks and Social Media Conference (SEEDA-CECNSM), Corfu, Greece, 25–27 September 2020; pp. 1–6. [Google Scholar]
  26. Rast, W.; Kimmig, S.E.; Giese, L.; Berger, A. Machine learning goes wild: Using data from captive individuals to infer wildlife behaviours. PLoS ONE 2020, 15, e0227317. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (a) From left to right: Antenova Asper 2.4 G/GNSS Antenna, and PCB front view, showing the microcontroller, the XBee module, and the GPS module. The IMU is under the XBee radio module, between the PCB and the module. (b) Horse with the collar device prototype attached.
Figure 1. (a) From left to right: Antenova Asper 2.4 G/GNSS Antenna, and PCB front view, showing the microcontroller, the XBee module, and the GPS module. The IMU is under the XBee radio module, between the PCB and the module. (b) Horse with the collar device prototype attached.
Sensors 21 02975 g001
Figure 2. Average of the results obtained with the accelerometer dataset over the different cross-validation folds for each of the analyzed configurations.
Figure 2. Average of the results obtained with the accelerometer dataset over the different cross-validation folds for each of the analyzed configurations.
Sensors 21 02975 g002
Figure 3. Average of the results obtained with the Kalman dataset over the different cross-validation folds for each of the analyzed configurations.
Figure 3. Average of the results obtained with the Kalman dataset over the different cross-validation folds for each of the analyzed configurations.
Sensors 21 02975 g003
Figure 4. X-cross accuracy and accuracy of the different ANN models deployed on the different hardware platforms presented in Section 2.6. Since these results are hardware-independent, no distinction between them is made. TFL stands for TensorFlow Lite, while FIQ refers to Full Integer Quantization.
Figure 4. X-cross accuracy and accuracy of the different ANN models deployed on the different hardware platforms presented in Section 2.6. Since these results are hardware-independent, no distinction between them is made. TFL stands for TensorFlow Lite, while FIQ refers to Full Integer Quantization.
Sensors 21 02975 g004
Figure 5. Execution time required for the ANN to perform a prediction for each of the two hardware platforms presented in Section 2.6.
Figure 5. Execution time required for the ANN to perform a prediction for each of the two hardware platforms presented in Section 2.6.
Sensors 21 02975 g005
Figure 6. Mops/s per Watt for each of the studied models and hardware platforms.
Figure 6. Mops/s per Watt for each of the studied models and hardware platforms.
Sensors 21 02975 g006
Figure 7. Flash memory usage for each of the studied models and hardware platforms.
Figure 7. Flash memory usage for each of the studied models and hardware platforms.
Sensors 21 02975 g007
Table 1. Main features of STM32L475 and STM32F476 microcontrollers.
Table 1. Main features of STM32L475 and STM32F476 microcontrollers.
DeviceArchitectureFeaturesClock (MHz)RAM (kBytes)Flash (Mbytes)Current (mA)
STM32L475Cortex-M4Ultra-low-power
Single precision FPU
80128110.7
STM32F746Corterx-M7High-performance
DSP with FPU
2163201130.1
Table 2. Estimated battery life for a Li-SOCL2(A3400) battery when performing 10 classifications per second with the studied microcontrollers and FIQ network architectures.
Table 2. Estimated battery life for a Li-SOCL2(A3400) battery when performing 10 classifications per second with the studied microcontrollers and FIQ network architectures.
DeviceBest Case (1 HL)Midpoint Case (2 HL)Worst Case (3 HL)
STM32L475404 days263 days88 days
STM32F746101 days35 days15 days
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dominguez-Morales, J.P.; Duran-Lopez, L.; Gutierrez-Galan, D.; Rios-Navarro, A.; Linares-Barranco, A.; Jimenez-Fernandez, A. Wildlife Monitoring on the Edge: A Performance Evaluation of Embedded Neural Networks on Microcontrollers for Animal Behavior Classification. Sensors 2021, 21, 2975. https://doi.org/10.3390/s21092975

AMA Style

Dominguez-Morales JP, Duran-Lopez L, Gutierrez-Galan D, Rios-Navarro A, Linares-Barranco A, Jimenez-Fernandez A. Wildlife Monitoring on the Edge: A Performance Evaluation of Embedded Neural Networks on Microcontrollers for Animal Behavior Classification. Sensors. 2021; 21(9):2975. https://doi.org/10.3390/s21092975

Chicago/Turabian Style

Dominguez-Morales, Juan P., Lourdes Duran-Lopez, Daniel Gutierrez-Galan, Antonio Rios-Navarro, Alejandro Linares-Barranco, and Angel Jimenez-Fernandez. 2021. "Wildlife Monitoring on the Edge: A Performance Evaluation of Embedded Neural Networks on Microcontrollers for Animal Behavior Classification" Sensors 21, no. 9: 2975. https://doi.org/10.3390/s21092975

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop