Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
New Energy Efficient Multi-Hop Routing Techniques for Wireless Sensor Networks: Static and Dynamic Techniques
Previous Article in Journal
Two-Dimensional DOA Estimation for Three-Parallel Nested Subarrays via Sparse Representation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Acoustic–Seismic Mixed Feature Extraction Based on Wavelet Transform for Vehicle Classification in Wireless Sensor Networks

College of Artificial Intelligence, National University of Defense Technology, Changsha, Hunan 410073, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(6), 1862; https://doi.org/10.3390/s18061862
Submission received: 28 April 2018 / Revised: 31 May 2018 / Accepted: 5 June 2018 / Published: 7 June 2018
(This article belongs to the Section Sensor Networks)

Abstract

:
An acoustic–seismic mixed feature extraction method based on the wavelet coefficient energy ratio (WCER) of the target signal is proposed in this study for classifying vehicle targets in wireless sensor networks. The signal was decomposed into a set of wavelet coefficients using the à trous algorithm, which is a concise method used to implement the wavelet transform of a discrete signal sequence. After the wavelet coefficients of the target acoustic and seismic signals were obtained, the energy ratio of each layer coefficient was calculated as the feature vector of the target signals. Subsequently, the acoustic and seismic features were merged into an acoustic–seismic mixed feature to improve the target classification accuracy after the acoustic and seismic WCER features of the target signal were simplified using the hierarchical clustering method. We selected the support vector machine method for classification and utilized the data acquired from a real-world experiment to validate the proposed method. The calculated results show that the WCER feature extraction method can effectively extract the target features from target signals. Feature simplification can reduce the time consumption of feature extraction and classification, with no effect on the target classification accuracy. The use of acoustic–seismic mixed features effectively improved target classification accuracy by approximately 12% compared with either acoustic signal or seismic signal alone.

1. Introduction

Wireless sensor networks (WSNs) consist of nodes capable of sensing, signal processing, and communicating. In addition to object detection, also of interest are the inherent properties of control and activation in WSNs [1]. One application of WSNs is the classification of moving vehicles in the interested area, which is a precondition for further decision-making in military operations. Moreover, vehicle classification is an important signal processing task and has found widespread civilian applications, such as intelligent transportation systems [2] and real-time traffic surveillance [3].
Signal feature extraction is the basis of vehicle classification and significantly affects the results of classification. Feature extraction methods of target signals applied to WSNs can be classified into two categories: Data fusion algorithms and algorithms based on a special sensor signal.
The feature extraction algorithm, based on data fusion, requires at least two signal sequences of the vehicle. The Haar discrete wavelet transform (DWT)-based method [4] is a vehicle feature extraction method based on data fusion. This method was proposed after combining the multi-sensor data fusion method with the wavelet transform (WT) method. This method employs a maximum entropy measure to determine significant wavelet coefficients. Features are obtained by calculating the energy of coefficients grouped around the competing clusters. Li et al. [5] proposed a multiple sensor information fusion method for vehicle type recognition based on the wireless geomagnetic sensor network. A maximum likelihood estimation method is used to perform vehicle feature waveform information fusion to improve vehicle type recognition accuracy. The feature in, feature out [6] process is another novel fusion method for achieving a joint feature vector estimate based on multiple sensors. The output of this process is fused feature vectors based on the single-source feature vector of an observed object provided by each sensor in the distributed embedded system.
The proposed algorithms based on the data fusion method [7,8] utilize target signals acquired from multi-sensors in WSNs to enhance the accuracy of target classification, but they require greater communication between nodes in a group or in a cluster and consume more energy. However, as hardware and software resources of nodes in WSNs are limited, the complexity and amount of resources required by the algorithms should be considered first. Despite the satisfactory performance demonstrated by a classification-algorithm-based fusion method [9], its high resource requirement has prompted researchers to consider feature extraction methods based on a special sensor signal.
Feature extraction algorithms based on a special sensor signal often use target features in the frequency domain or wavelet domain of their acoustic, seismic, or magnetic signals. The improved time-encoded signal processing algorithm [10] is a feature extraction method based on the acoustic signal of a special sensor. This method was proposed to ameliorate the time-encoded signal processing algorithm, which is effective for feature extraction of speech signals, but not suitable for vehicle acoustic signals. This algorithm designs an extensional symbol table with 40 characters according to the characteristic features of the vehicle acoustic signal and, thereafter, uses a support vector machine (SVM) [11] as the classifier to recognize different vehicle types. A fast Fourier transform (FFT)-based feature extraction method [12] was introduced to extract vehicle features in a real-life vehicle tracking sensor network. The extracted features, which were based on the frequency spectrum of the acoustic signals of the target, were used to test three classification algorithms. There have also been studies attempting to use the time-domain features of the target signal to classify vehicles. For example, a time domain harmonics amplitude method [13], which extracts features from the energy of the target signal by estimating the strongest harmonic frequency and the harmonics’ amplitude from a template of the acoustic signal, was developed for vehicle classification in a sensor network.
The algorithms mentioned above utilize the acoustic signal, seismic signal, or magnetic signal of the target acquired by a single sensor node in WSNs to classify the target type. In this paper, we propose a different approach to extract the signal feature based on the acoustic and seismic signals acquired from a single sensor node for vehicle classification. Instead of using a traditional spectral-domain or time-domain method, we use a WT-based method to extract features from acoustic and seismic signals. First, the signal sequence was decomposed into a set of wavelet coefficients by using the à trous algorithm. After the wavelet coefficients of the target acoustic and seismic signals were obtained, the energy ratio of each layer coefficient was calculated and utilized as the feature vector of the target signals. Subsequently, the acoustic and seismic features were merged into an acoustic–seismic mixed feature to improve target classification accuracy.
The rest of the paper is organized as follows. The wavelet coefficient energy ratio (WCER) method, based on wavelet decomposition using the à trous algorithm, is introduced in Section 2. Section 3 introduces the feature vector, simplified using the hierarchical cluster method, and Section 4 introduces the acoustic–seismic mixed feature of the target signal. After briefly reviewing the data source and the DARPA SensIT project [12], and utilizing the subset data provided by this experiment to validate the proposed method, the performance of the method is discussed in Section 5. Section 6 concludes the paper.

2. WCER Feature Extraction Method

Wavelet transform is a signal processing technique that represents a transient or non-stationary signal in terms of time and scale distribution, and is an excellent tool for data compression [14], analysis, and denoising [15]. After wavelet composition of the signal, the wavelet coefficients represent the signal components in different frequency bands, making it feasible to differentiate signals of different targets.

2.1. Wavelet Decomposition Based on à Trous Algorithm

The wavelet decomposition method and reconstruction method are based on WT. WT decomposes a signal into wavelet coefficients by using a set of base functions [16]. A set of functions gained following basic wavelet function Ψ(t) were flexed in scale and displaced in the time domain.
W T x ( a , b ) = < x ( t ) , Ψ a , b ( t ) > = 1 | a | + x ( t ) Ψ ( t b a ) d t , a , b R , a 0
where Ψ(t) is the wavelet “prototype”, which can be considered a bandpass function [17], a is the contraction–expansion or scale factor, and b is the displacement factor.
Instead of continuously varying the parameters, as in the case of the continuous wavelet transform [18], which contains a large amount of redundant information, we can analyze the signal with a smaller number of scales with a varying number of translations at each scale. This is the DWT process [19]. The à trous algorithm [20] is one of the most concise methods to implement the fast DWT. The algorithm utilizes the equivalent translocation property of the z-transform and performs a convolution operation of orthogonal filter coefficients and signal to calculate the wavelet coefficients of the signal sequence after inserting a certain number of zeros between the coefficients of the orthogonal filter.
Figure 1 shows the scheme of the three-level wavelet decomposition based on the à trous algorithm. x(k) is the signal sequence and hj and gj are the orthogonal filter bank coefficients of the (j − 1) level in the wavelet decomposition process of the signal. cAj(k) indicates the wavelet approximation coefficients of the jth level and represents the low-frequency components of the current wavelet decomposition of the signal. cDj(k) indicates the wavelet detail coefficients of the jth level and represents the high-frequency components of the current wavelet decomposition of the signal. hj represents h0 up-sampled by 2j, while gj represents g0 up-sampled by 2j. When using the à trous algorithm to decompose the target signal, the orthogonal filter bank coefficients of the upper level are interpolated to obtain the filter coefficients of the current level first and, thereafter, the input sequence of the current level is convolved with the orthogonal filter banks to calculate the wavelet coefficients. The approximation coefficient obtained at the previous level is utilized as the input signal of the next level. The formulas for the calculation of cAj(k) and cDj(k) are as follows:
c A j ( k ) = c A j 1 ( k ) h j = l = 1 k h j ( l ) c A j 1 ( k l + 1 ) , h j ( l ) = 0   w h i l e l > L j
c D j ( k ) = c D j 1 ( k ) g j = l = 1 k g j ( l ) c D j 1 ( k l + 1 ) , g j ( l ) = 0   w h i l e l > L j
where represents a convolution operation; k = 1, 2, …, K and K is the length of the signal sequence; and j = 1, 2, …, J and J is the desired wavelet composition depth. The lengths of the orthogonal filter bank coefficients, gi and hi, at each decomposition level are different. Lj represents the length of the orthogonal filter bank coefficients at the jth level. cA0(k) is equal to the initial signal sequence, x(k), when calculating cA1(k) and cD1(k) using Equations (2) and (3), respectively. The length of the result sequence after the convolution operation is K + Lj − 1, which indicates that the sequence length after convolution at different levels of wavelet decomposition is not the same. To simplify the calculation, we intercept a sequence of length K from the middle of the convolution result sequence at each wavelet decomposition level.
Considering the wavelet decomposition of the first layer as an example, the convolution result of the signal, x(k), and decomposition high-pass filters, g0, is the wavelet detail coefficient, cD1(k), and the wavelet approximation coefficient, cA1(k), is obtained using the convolved signal, x(k), and the decomposition low-pass filters, h0. cA1(k), is, thereafter, utilized as the input signal of the second level. After this wavelet decomposition algorithm is completed, the signal, x(k), is decomposed into a set of approximation coefficients and detail coefficients.

2.2. Wavelet Coefficients Energy Ratio (WCER) Method

As the wavelet coefficients of the signal correspond to different frequency bands of the target signal with different signal spectrum distribution, the wavelet coefficients of the target signals have different characteristics and can be utilized to extract spectrum features for classification purposes.
The difference on spectrums of different target signals make the energy of the signal component in different frequency bands discrepant, hence, the wavelet coefficients of acoustic and seismic signals of different targets vary. WCER, which can indicate the difference between energies of coefficients, can intuitively reflect the energy difference of signals in different frequency bands. WCER can be expressed as the wavelet detail coefficient energy ratio (WDcer) and wavelet approximation coefficient energy ratio (WAcer), which are defined as follows:
W D c e r ( j ) = P D ( j ) j = 1 J P D ( j ) , W A c e r ( j ) = P A ( j ) j = 1 J P A ( j )
where J is the desired wavelet transform depth, PD(j) represents the jth wavelet detail coefficient energy, and PA(j) represents the jth wavelet approximation coefficient energy. The formulas for calculating PD(j) and PA(j) are as follows:
P D ( j ) = k = 1 K ( c D j ( k ) ) 2 , P A ( j ) = k = 1 K ( c A j ( k ) ) 2
where N is the length of the wavelet coefficient, which is also the length of the signal sequence. After WDcer and WAcer are calculated, the feature of the target signal based on WCER can be obtained. The feature vector of the target signal is denoted as f.
f = [ W D c e r ( 1 ) , , W D c e r ( J ) ; W A c e r ( 1 ) , , W A c e r ( J ) ]

2.3. Signal Feature Extraction

The feature vector should be extracted after a specified length of the signal sequence is acquired. Assuming the feature of the target signal sequence is extracted every K points, the target signal that is acquired is divided into N segments. The process of feature extraction based on WCER is shown in Figure 2.
Figure 2 shows the process of feature extraction based on WCER. J is the desired wavelet decomposition depth and K is the length of each segment of the signal. cDj(k) (k = 1, 2, …, K) represents the wavelet detail coefficient and cAj(k) represents the wavelet approximation coefficient of the jth level. F represents the feature matrix of the target signal. f = [W1, W2, …, Wm] represents the feature vector and fi is the feature vector of the ith segment of the target signal. Wm is the mth variables of the feature vector, f, and [w1m, w2m, …, wNm] is the observation sequence of the variable Wm.
If the sensors acquire acoustic and seismic signals of targets at a sampling rate of 4096 Hz and the signal generated by the target lasts 10 s, then the sensor node acquired 40,960 points sequence. With the K set to 512, then the acquired signal sequence is divided into 80 segments and every 512 points is taken as a segment for feature extracting using the WCER method. This means point 1 to point 512 is the first segment, point 513 to point 1024 is the second segment, and the entire data sequence is segmented in this way.
Subsequently, we utilize the feature extraction method, based on the WCER, to process the real-world acquired signal of vehicles in WSNs. The datasets were gathered from a real-world experiment, the third SensIT situational experiment [12]. In the experiment, 75 sensor nodes with acoustic, seismic, and infrared sensing capability were located at the Marine Corps Air Ground Combat Center. Testing runs were performed by driving different kinds of vehicles across the testing field. Four target vehicle classes were used: Assault amphibian vehicle (AAV), main battle tank, high-mobility multipurpose wheeled vehicle, and dragon wagon (DW). The acoustic and seismic data of each run were recorded and the sampling rate of the signals was 4096 Hz. There were three paths in the experimental area for the vehicles to pass and a testing run was accomplished by driving the vehicle on the path. A series of numbers was utilized to indicate these runs, e.g., DW3, DW4, DW5. The acoustic and seismic data of each run were recorded by the sensor nodes deployed at the side of the road. The details of the experiment are described in the paper mentioned above.
K, which is also the length of the segment of the target signal sequence, was set to 512 and the desired wavelet transform depth, J, was set to 8. In addition, a nearly orthogonal design of a bi-orthogonal filter bank, shown in Table 1, was utilized as the orthogonal filter bank to decompose the signal sequence.
Subsequently, we utilize the feature extraction method, based on the WCER, to extract the acoustic and seismic features of AAV3 runs.
In the AAV3 run, the acoustic and seismic signals of AAV were sensed and acquired by 18 sensor nodes in the WSN monitoring system. All the acquired signal sequences were divided into 4676 segments at intervals of 512 points for feature extraction. Figure 3b shows the extracted features of the segmented acoustic signal sequence using the WCER method and Figure 3d shows the extracted seismic features. The horizontal axis represents the number of signal segments, which also could be considered as a set of observations of a certain feature variable, while the vertical axis represents the WCER of each segmented signal.

3. Feature Vector Simplification Using Hierarchical Cluster Method

The acoustic and seismic signal features of AAV3, shown in Figure 3, indicate that differences between parts of the WCER feature variables are evident and can reflect the target characteristic, whereas the differences between the remaining feature variables are insignificant. Simplifying the feature vector will remove the redundant variables in the feature vector, which can decrease the computational complexity of the subsequent target classification and effectively reduce the time consumption of the process of target classification. The hierarchical cluster method [21] and principal component analysis (PCA) method [22] are two effective approaches to reduce the dimensions of the feature matrix.

3.1. Hierarchical Cluster Method

The hierarchical cluster method can aggregate variables into several clusters based on their similarity. After determining representative variables from aggregated clusters to reduce the number of variables, the dimension of the feature vector will be reduced. The feature vector of the signal obtained using the WCER method is f = [W1, W2, …, Wm], m = 2J, and J is the desired wavelet decomposition depth. The observation matrix of the feature vector of the target signal is denoted as F, whose column vector is the data sequence of the variable in the feature vector, f. By analyzing the sample observation matrix, the main variables in the feature vector are selected to realize its dimensionality reduction.
The hierarchical cluster method utilizes the distance between variables in the feature vector, f, and between clusters in the process of clustering to quantify the similarities between variables and the clusters. First, a pair of the nearest feature variables is selected to be merged into a new class according to the calculated feature variable distance matrix, D, expressed in Equation (7), then the two closest clusters are merged into a class. This process is repeated until a preset number of clusters is achieved:
D = [ 0 d i s t ( W 2 , W 1 ) 0 0 d i s t ( W 2 J , W 1 ) d i s t ( W 2 J 1 , W 1 ) 0 ]
where dist (Wu,Wv) represents the distance between the variable, Wu, and the variable, Wv. Wu is the uth variable in the feature vector and Wv is the vth variable.
Figure 4 shows the distance between variables in the feature vector and between clusters in the clustering process. d (r,s) represents the distance between the cluster, r, and the cluster, s.
The distance between the variables, Wu and Wv, is calculated using their correlation coefficient and the calculation method is expressed in Equation (8).
d i s t ( W u , W v ) = 1 R ( W u , W v ) = 1 C ( W u , W v ) C ( W u , W u ) C ( W v , W v )
where R (Wu,Wv) represents the correlation coefficient of Wu and Wv. C(Wu,Wv) is the covariance of Wu and Wv, and it is calculated as follows:
C ( W u , W v ) = E [ ( W u E ( W u ) ) · ( W v E ( W v ) ) ]
After obtaining the distance matrix, we must also calculate the distance between the clusters in the clustering process. In this study, the unweighted pair grouping method with arithmetic mean (UPGMA) [23] was utilized to calculate the distance between clusters. UPGMA is a widely used bottom-up hierarchical clustering method that defines cluster similarity [24] in terms of the average pairwise distance between all the objects in two different clusters. If the two clusters in the clustering process are the cluster, r, and the cluster, s, the formula for calculating the distance between them utilizing the UPGMA is expressed in Equation (10).
d ( r , s ) = 1 n r n s p = 1 n r q = 1 n s d i s t ( W r p , W s q )
where nr is the number of objects in cluster r and ns is the number of objects in cluster s. Wrp is the pth subject in cluster r and Wsq is the qth subject in cluster s.

3.2. PCA Method

The PCA method is another efficient approach to reduce the dimension of the feature vector. This approach reduces the dimensions of the feature by extracting several factors that contribute the most to signal differences in the observation matrix of the target feature vector.
The feature vector of the signal is f = [W1, W2, …, Wm] and the column vector, [wi1, wi2, …, wim], in the feature matrix, F, is the data sequence of the variable, Wi, in the feature vector, f. The variable should be normalized according to Equation (11) before performing principal component analysis.
W ˜ i = W i μ i s i
where μi is the mean and Si is the standard deviation of the sequence of the variable, Wi. After calculating the correlation coefficient of the variables in the feature vector, the correlation coefficient matrix can be achieved.
C = [ R ( W ˜ 1 , W ˜ 1 ) R ( W ˜ 1 , W ˜ m ) R ( W ˜ m , W ˜ 1 ) R ( W ˜ m , W ˜ m ) ]
Then, all the eigenvalues, λi (i = 1, 2, …, m), of the correlation coefficient matrix, C, and the corresponding eigenvectors, [ui1, ui2, …, uim], are calculated. Then the eigenvectors consists of m new indicator variables shown as follows:
{ y 1 = u 11 · W ˜ 1 + u 12 · W ˜ 2 + + u 1 m · W ˜ m y 2 = u 21 · W ˜ 1 + u 22 · W ˜ 2 + + u 2 m · W ˜ m y m = u m 1 · W ˜ 1 + u m 2 · W ˜ 2 + + u m m · W ˜ m
where yi is the ith principal component. The contribution rate of the principal component, which means how much the new indicator variable contributes to the difference of the target features, is calculated from its corresponding eigenvector:
  b k = λ k i = 1 m λ i k = 1 , 2 , , m
When the sum of the former P contribution rates of the of new indicator variables is approximately 1 (usually 0.9 or 0.95), the former P indicator variables, [y1, y2, …, yP], are selected as the new features vector instead of the original m features.

3.3. Comprasion of Two Feature Simplification Methods

To show the performance of these two methods, we calculated the time consumption of the signal feature extraction using the methods for feature dimensionality reduction. We also calculated the classification accuracy of the features after reducing the dimensions using two methods. The acoustic feature data of AAV3 and AAV4 after dimension reduction were used to train the SVM classifier. Subsequently, we used the trained classifier to classify the dataset of the other 15 runs (AAV5–AAV11, DW5–DW12) after their acoustic feature data were simplified using the two methods. The classification accuracy is denoted as Ac. We evaluated the accuracy as follows:
A c = # correctly   classified   data # total   testing   data × 100 %
The classification accuracy of the ith dataset is denoted as Ac(i) (i = 1, 2, …, M) and the average classification accuracy Aca is:
A c a = 1 M i = 1 M A c ( i )
where M is the number of datasets, acquired from different targets, that have been classified.
A comparison of the time needed and classification accuracy when using the hierarchical clustering and principal factor analysis methods to simplify features is shown in Table 2. fa_HC represents the acoustic features of the target after being simplified by hierarchical clustering and fa_PCA represents the acoustic features of the target after being simplified by principal factor analysis. It can be seen from Table 2 that after training the classifier, when using hierarchical clustering to reduce the dimensions of the feature, the feature extraction time is slightly less than when using principal factor analysis and the classification accuracy is higher. Therefore, the hierarchical clustering method was chosen for simplifying the feature vector.

3.4. Feature Vector Simplification Using Hierarchical Clustering

The acoustic and seismic signal features of AAV3 were considered as examples for simplifying the variables in the feature vectors using the hierarchical clustering method.
Figure 5 shows the clustering results of the feature vectors of AAV3. Eight-level wavelet decomposition of the acoustic and seismic signals of AAV3 yields two feature vectors with 16 variables. The acoustic feature vector is denoted as fa and the seismic feature vector is denoted as fs. After clustering variables with a distance less than 1, the variables with large differences in the feature vectors are retained and the variables with small differences are aggregated into a cluster.
The 16 variables in the acoustic feature vector, ( fa = [ W 1 , , W 16 ] ), were reduced to 9 after feature vector simplification. The variables, W1, W2, W3, and W4, were aggregated into a cluster and replaced with the variable, W4. The variable, W12, was utilized instead of the variables, W9, W10, W11, and W12, after they were aggregated into a cluster. Further, the variables, W15 and W16, were aggregated into a cluster and the variable, W15, was subsequently utilized instead.
The 16 variables in the acoustic feature vector, ( fs = [ W 1 , , W 16 ] ) ), were reduced to 7. The variables, W1, W2, W3, and W4, were aggregated into a cluster and replaced with the variable, W4. The variable, W14, was utilized instead of the variables, W9, W10, W11, W12, W13, and W14, after they were aggregated into a cluster. Further, the variables, W15 and W16, were aggregated into a cluster and the variable, W15, was subsequently utilized instead.
The simplified acoustic and seismic feature vectors are denoted as fanew and fsnew, where fanew = [W4, W5, W6, W7, W8, W12, W13, W14, W15], which corresponds to [WDcer(4), …, WDcer(8); WAcer(4), …, WAcer(7)]. fsnew = [W4, W5, W6, W7, W8, W14, W15], which corresponds to [WDcer(4), …, WDcer(8); WAcer(6), WAcer(7)].
The results of using simplified features for target classification will be discussed in Section 5.1.

4. Acoustic–Seismic Mixed Feature

After simplifying the features, we utilized the simplified acoustic and seismic features of the signal to classify the target.

4.1. Support Vector Machine Classifier

The commonly used classification methods are k-nearest neighbor (KNN) [25], decision tree (DT) [26], naive Bayes (NB) [27], and support vector machine (SVM). We used the acoustic feature data of AAV3 and AAV4 to train the classifiers and classify the acoustic feature data of the other 15 runs (AAV5–AAV11, DW5–DW12). Then, average classification accuracy and time needed were recorded to compare the performance of these methods in vehicle classification.
Table 3 shows the average classification accuracy and average time needed for different classification methods. The results indicate that KNN has the highest accuracy, but the classification takes too much time. The time needed for the DT and NB methods is very short, but their classification accuracy is lower than that of the KNN and SVM.
In summary, the SVM method has satisfactory classification accuracy and acceptable time needs. Therefore, the SVM algorithm was used as a classifier to validate the performance of the proposed WCER feature extraction method before and after feature simplification.
The SVM classification algorithm was performed by using the C support vector classifier (C-SVC) shared in LIBSVM [28], a commonly used SVM software library.
The principle of classification using SVM is to consider a multivariate feature as an independent point in a multidimensional space and, thereafter, utilize the training data to determine an optimal hyperplane to classify the independent points. Summarized mathematically, C-SVC is aimed at solving the following optimization problem:
m i n ( 1 2 z T z + R i = 1 N t ε i )
under the following constraints:
1 ε i y i ( z T ϕ ( f i ) + c ) 0 , ε i 0 ,   i = 1 , , N t
where Nt is the number of training vectors, fi is the sample feature vector for training, and yi is the class label corresponding to the feature vector. Function, ϕ ( f i ) , maps the feature vector, fi, to a high-dimensional space and R is the regularization parameter, which is set to 1 in this study. Usually, we solve the following dual problem:
m i n ( 1 2 α T Q α e T α )
under the following constraints:
y T α = 0 ,    0 α i R , i = 1 , , N t
where e = [1, …, 1]T is the unit vector. Q is an Nt by Nt positive semi-definite matrix. Q i j = y i y j K ( f i f j ) , and K ( f i f j ) = ϕ ( f i ) T ϕ ( f j ) is the kernel function. Moreover, the function for classifying a feature, f, is as follows:
sgn ( z T ϕ ( f ) + c ) = sgn ( i = 1 N t y i α i K ( f i f ) + c )
The label of the class is 1 or −1 for representing different kinds of targets, which indicates that y i { 1 , 1 } . Furthermore, we utilized a radial basis function kernel [29], with the following format:
K ( f i , f j ) = exp ( γ f i f j )
where the parameter, γ, in the kernel function is set to 0.5.

4.2. Vehicle Classification Using SVM Classifier

We extracted features from the acoustic and seismic signal data collected from 18 sensor nodes in runs AAV3, AAV4, DW3, and DW4. These extracted features were used to train the SVM classifier. Subsequently, we used the trained classifier to classify the dataset of the other 15 runs (AAV5–AAV11, DW5–DW12) before and after simplifying the acoustic and seismic WCER features.
The classification results, listed in Table 4, show that the classification accuracy of AAV is high, but that of DW was unsatisfactory, when classifying the vehicles using simplified acoustic features. By contrast, the classification accuracy of AAV was low and that of DW was satisfactory when the simplified seismic features were used. Therefore, we mixed acoustic and seismic features of the target signal to achieve satisfactory classification accuracy for both AAV and DW. The acoustic–seismic mixed features of the target signal can be described by the vectors, fmix and fmix = [fanew; fsnew].
Figure 6 shows the acoustic–seismic mixed features of signals acquired from AAV3 (Figure 6a) and DW3 (Figure 6b). The results of using acoustic–seismic mixed features for target classification are discussed in Section 5.2.

5. Performance of WCER Method

As there is no effective approach to directly evaluate the performance of the proposed WCER feature extraction method, we evaluated the method indirectly by comparing the performance of an SVM classifier using different features.

5.1. Effects of Feature Simplification and Mixed Features on Classification Accuracy

To study the performance when using simplified features for target classification, we compared the classification accuracy obtained using the features before and after simplification. The features extracted from runs AAV3, AAV4, DW3, and DW4 were utilized to train the SVM classifier before and after simplification. Subsequently, the trained classifier was used to classify the target using the features extracted from the dataset of the other 15 runs (AAV5–AAV11, DW5–DW12) before and after simplification. The set of parameters of the classifier is described in detail in Section 4. The classification accuracy obtained by using the acoustic–seismic mixed features was calculated similarly for studying the effect of feature mixing on vehicle classification.
Table 5 shows the classification accuracy obtained by using different feature vectors based on the WCER method. fa (fs) indicates that the targets were classified using the feature extracted from the acoustic (seismic) signal of the target without simplification and fanew (fsnew) indicates that the targets were classified using the feature extracted from the acoustic (seismic) signal of the target after feature simplification. fmix represents the classification of vehicles using acoustic–seismic mixed features of the target signal.
To further clarify the classification accuracy obtained by using different extracted features, Table 4 summarizes the average classification accuracy.
We also calculated the time consumption of target classification using these features to compare the efficiency of using different features. Two types of time use were calculated to compare the efficiency: Time consumption of the feature extraction procedure and the time needed for subsequent target classification using the extracted features. To compare the time consumption of the feature extraction step, we utilized the feature extraction methods to process the dataset of 15 runs (AAV5–AAV11, DW5–DW12) and calculated the average time consumed by feature extraction in each run. Subsequently, the average time consumption of target classification using these feature vectors was recorded to study the effect of using different feature vectors on the time consumption of the target classification.
The average classification accuracy and average time consumption when using different feature vectors are shown in Table 6. The results show that the WCER feature extraction method can effectively extract the target features from the target signals. When using the target acoustic and seismic features extracted using the WCER method to classify vehicles, the classification accuracy reached approximately 67%. Feature simplification can reduce the time consumption of feature extraction and classification, while having no effect on target classification accuracy. Compared with the use of feature vectors before simplification, the time consumption after feature simplification decreased slightly and the time consumption of target classification decreased by almost half. The average classification accuracy obtained by using simplified WCER features of the acoustic signal was 0.56% lower than that obtained by using unsimplified features, and the classification accuracy decreased by 0.38% after the features extracted from the seismic signal were simplified. This is mainly because the removed parts of the feature vectors in the simplification process were redundant and had no effect on the overall characteristics of the target signal.
Using mixed features can effectively improve the target classification accuracy. Compared with the use of acoustic or seismic features, classification accuracy was improved by approximately 12% when the acoustic–seismic mixed feature of the target signal was used.

5.2. Comparison with FFT-Based Feature Extraction Method

The experimental results in Section 5.1 show that the WCER feature extraction method can effectively extract the characteristics of the target signal and that the feature mixing approach can improve classification accuracy. However, whether the performance of this feature extraction method is better than that of the existing methods is yet to be determined.
The FFT-based feature extraction method mentioned in Section 1 is a mature and commonly used feature extraction method. To extract the features of the target signal, the FFT of the acoustic signals is computed every 512 points. Subsequently, the averages of every two points of the first 100 FFT values that have been normalized are used as the extracted features of 50 dimensions for target classification. We used this feature extraction method to compare with the proposed hybrid feature extraction method.
The data of runs AAV3, AAV4, DW3, and DW4 were used for training and the features extracted from the dataset of the other 15 runs (AAV5–AAV11, DW5–DW12) were classified by the trained SVM classifier.
The average time consumptions of the FFT-based and the WCER feature extraction methods were calculated to compare the efficiency of these methods.
As shown in Table 7 and Table 8, the average classification accuracy of the WCER method was 2.88% higher than that of the FFT-based method when the features extracted from the target acoustic signal were used. This is mainly because only the energy of the low-frequency band of the target signal is extracted by the FFT-based method and utilized for target classification. Compared with the FFT-based method, the proposed WCER feature extraction method utilizes all the signal spectrum energy characteristics to classify targets. We also observed that the time consumed by the proposed WCER method to extract features from the target acoustic signal is less than that of the FFT-based method because wavelet decomposition of the signal sequence takes less time than the FFT operation of the same signal sequence. Moreover, the number of variables in the signal feature extracted by the WCER method is less than the number of feature variables extracted by the FFT-based method, resulting in less time spent on subsequent target classification when using the proposed WCER method.
Compared with the FFT-based feature extraction method, average classification accuracy is improved by 15.5% when using acoustic–seismic mixed features. At the same time, the time consumption of target feature extraction is increased significantly. This is because the increased number of calculations leads to more time spent as the WCER method requires extraction of not only the features of the target acoustic signal, but also of the seismic features of the same target when using the acoustic–seismic mixed features.

6. Conclusions

In this paper, we proposed a novel feature extraction method named WCER based on WT for vehicle classification in WSNs. The acoustic-seismic mixed feature extraction method is presented after the acoustic and seismic WCER features of the target signal were simplified using the hierarchical clustering method. The flow of the proposed method for extracting acoustic–seismic mixed features of the target is shown in Figure 7. Finally, we studied the target classification accuracy and time consumption of the proposed feature extraction method. The experiment results show that classification accuracy of the targets detected in WSN is effectively improved by using the acoustic–seismic mixed feature of the target.
Another advantage of using the wavelet-based feature extraction method is that the wavelet coefficients can be processed using the thresholding method [30] for denoising before calculating the WCER of the target signal. Thus, the combination of the denoising and feature extraction methods can effectively reduce the time consumption of the entire signal processing of WSN from denoising to target classification.

Author Contributions

Conceptualization, H.Z. and Z.P.; software, H.Z.; supervision, Z.P.; validation, H.Z.; writing (original draft), H.Z.; writing, review and editing, W.Z.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript
WCERWavelet coefficient energy ratio
WSNsWireless sensor networks
WTWavelet transform
DWTDiscrete wavelet transform
FFTFast Fourier transform
AAVAssault amphibian vehicle
DWDragon wagon
UPGMAUnweighted pair grouping method with arithmetic mean
PCAPrincipal component analysis
KNNK-nearest neighbor
DTDecision tree
NBNaive Bayes
SVMSupport vector machine
C-SVCC support vector classifier

References

  1. Sohraby, K.; Minoli, D.; Znati, T. Wireless Sensor Networks: Technology, Protocols, and Applications, 1st ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2007. [Google Scholar]
  2. Mishra, D.P.; Dorale, S.S. An Application of Wireless Sensor Network in Intelligent Transportation System. In Proceedings of the 6th International Conference on Emerging Trends in Engineering and Technology, Nagpur, India, 16–18 December 2013; pp. 90–91. [Google Scholar]
  3. Balid, W.; Tafish, H.; Refai, H.H. Development of Portable Wireless Sensor Network System for Real-time Traffic Surveillance. In Proceedings of the IEEE 18th International Conference on Intelligent Transportation System, Las Palmas, Spain, 15–18 September 2015; pp. 1630–1637. [Google Scholar]
  4. Hortos, W.S. Multi-source feature extraction and target recognition in wireless sensor networks based on adaptive distributed wavelet compression algorithms. In Proceedings of the Conference on Automatic Target Recognition XVIII, Orlando, FL, USA, 19–20 March 2008. [Google Scholar]
  5. Li, F.; Lv, Z. Reliable vehicle type recognition based on information fusion in multiple sensor networks. Comput. Netw. 2017, 117, 76–84. [Google Scholar] [CrossRef]
  6. Klausner, A.; Tengg, A.; Rinner, B. Distributed Multilevel Data Fusion for Networked Embedded Systems. IEEE J. Sel. Top. Signal Process. 2008, 2, 538–555. [Google Scholar] [CrossRef]
  7. Wang, S.; Yang, B. Multi-sensor Data Fusion in Wireless Sensor Network for Target Detection. In Proceedings of the IEEE International Conference on Electronics Information and Emergency Communication, Beijing, China, 15–17 November 2014; pp. 29–32. [Google Scholar]
  8. Wang, X.; Wang, S.; Bi, D. Distributed Visual-Target-Surveillance System in Wireless Sensor Networks. IEEE Trans. Syst. Man Cybern. 2009, 39, 1134–1146. [Google Scholar] [CrossRef] [PubMed]
  9. Kim, Y.; Jeong, S.; Kim, D.; López, T.S. An efficient scheme of target classification and information fusion in wireless sensor networks. Pers. Ubiquitous Comput. 2009, 13, 499–508. [Google Scholar] [CrossRef]
  10. Wang, Y.; Zhou, J. Vehicle type recognition in WSN based on ITESP Algorithm. In Proceedings of the IEEE 10th International Conference on Ubiquitous Intelligence and Computing (UIC)/IEEE 10th International Conference on Autonomic and Trusted Computing (ATC), Sorrento Peninsula, Italy, 18–21 December 2013; pp. 668–671. [Google Scholar]
  11. Hsu, C.W; Lin, C.J. A comparison of methods for multiclass support vector machines. IEEE Trans. Neural Netw. 2002, 13, 415–425. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Duarte, M.F.; Hu, Y.H. Vehicle classification in distributed sensor networks. J. Parallel Distrib. Comput. 2004, 64, 826–838. [Google Scholar] [CrossRef] [Green Version]
  13. William, P.E.; Hoffman, M.W. Classification of Military Ground Vehicles Using Time Domain Harmonics' Amplitudes. IEEE Trans. Instrum. Meas. 2011, 60, 3720–3731. [Google Scholar] [CrossRef]
  14. Mamaghanian, H.; Khaled, N.; Atienza, D.; Vandergheynst, P. Compressed Sensing for Real-Time Energy-Efficient ECG Compression on Wireless Body Sensor Nodes. IEEE Trans. Biomed. Eng. 2011, 58, 2456–2466. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Shahbakhti, M. A Novel DWT Method for ECG Noise Elimination. IEEJ Trans. Electr. Electron. Eng. 2015, 10, 353–355. [Google Scholar] [CrossRef]
  16. Qin, Z.G.; Chen, L.; Bao, X.Y. Wavelet Denoising Method for Improving Detection Performance of Distributed Vibration Sensor. IEEE Photonics Technol. Lett. 2012, 24, 542–554. [Google Scholar] [CrossRef]
  17. Rioul, O.; Duhamel, P. Fast Algorithms for Discrete and Continuous Wavelet Transforms. IEEE Trans. Inf. Theory 1992, 38, 569–586. [Google Scholar] [CrossRef]
  18. Gargour, C.; Gabrea, M.; Ramachandran, V.; Lina, J.M. A Short Introduction to Wavelets and Their Applications. IEEE Circuits Syst. Mag. 2009, 9, 57–68. [Google Scholar] [CrossRef]
  19. Průša, Z.; Søndergaard, P.L.; Rajmic, P. Discrete Wavelet Transforms in the Large Time-Frequency Analysis Toolbox for MATLAB/GNU Octave. ACM Trans. Math. Softw. 2016, 42, 32. [Google Scholar] [CrossRef]
  20. Kozakevicius, A.; Schmidt, A.A. Wavelet transform with special boundary treatment for 1D data. Comput. Appl. Math. 2013, 32, 447–457. [Google Scholar] [CrossRef]
  21. Ackerman, M.; Ben-David, S. A Characterization of Linkage-Based Hierarchical Clustering. ‎J. Mach. Learn. Res. 2016, 17, 8182–8198. [Google Scholar]
  22. Cao, L.J.; Chua, K.S.; Chong, W.K.; Lee, H.P.; Gu, Q.M. A comparison of PCA, KPCA and ICA for dimensionality reduction in support vector machine. Neurocomputing 2003, 55, 321–336. [Google Scholar] [CrossRef]
  23. Gronau, I.; Moran, S. Optimal implementations of UPGMA and other common clustering algorithms. Inf. Process. Lett. 2007, 104, 205–210. [Google Scholar] [CrossRef] [Green Version]
  24. Murtagh, F.; Contreras, P. Algorithms for hierarchical clustering: An overview. Wiley Interdiscip. Data Min. Knowl. Discov. 2012, 2, 86–97. [Google Scholar] [CrossRef]
  25. Guo, G.D.; Wang, H.; Bell, D.; Bi, Y.X.; Greer, K. KNN model-based approach in classification. In Proceedings of the OTM Confederated International Conference CoopIS, DOA and ODBASE, Catania, Italy, 3–7 November 2003; Volume 2888, pp. 986–996. [Google Scholar]
  26. Myles, A.J.; Feudale, R.N.; Liu, Y.; Woody, N.A.; Brown, S.D. An introduction to decision tree modeling. J. Chemom. 2004, 18, 275–285. [Google Scholar] [CrossRef]
  27. Langseth, H.; Nielsen, T.D. Classification using Hierarchical Naive Bayes models. Mach. Learn. 2006, 63, 135–159. [Google Scholar] [CrossRef]
  28. Chang, C.C.; Lin, C.J. LIBSVM: A Library for Support Vector Machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27. [Google Scholar] [CrossRef]
  29. Liu, Q.Z.; Chen, C.H.; Zhang, Y.; Hu, Z.G. Feature selection for support vector machines with RBF kernel. Artif. Intell. Rev. 2011, 36, 99–115. [Google Scholar] [CrossRef]
  30. Jagtiani, A.V.; Sawant, R.; Carletta, J.; Zhe, J. Wavelet transform-based methods for denoising of Coulter counter signals. Meas. Sci. Technol. 2008, 19, 065102. [Google Scholar] [CrossRef]
Figure 1. Three-level wavelet decomposition algorithm based on the à trous algorithm, hj represents h0 up-sampled by 2j and gj represents g0 up-sampled by 2j.
Figure 1. Three-level wavelet decomposition algorithm based on the à trous algorithm, hj represents h0 up-sampled by 2j and gj represents g0 up-sampled by 2j.
Sensors 18 01862 g001
Figure 2. Process of feature extraction using wavelet coefficient energy ratio (WCER).
Figure 2. Process of feature extraction using wavelet coefficient energy ratio (WCER).
Sensors 18 01862 g002
Figure 3. WCER features of AAV3 signals: (a) Acoustic signal of AAV3 acquired from 18 sensor nodes; (b) features extracted from the acoustic signal of AAV3; (c) seismic signal of AAV3 acquired from 18 sensor nodes; and (d) features extracted from the seismic signal of AAV3.
Figure 3. WCER features of AAV3 signals: (a) Acoustic signal of AAV3 acquired from 18 sensor nodes; (b) features extracted from the acoustic signal of AAV3; (c) seismic signal of AAV3 acquired from 18 sensor nodes; and (d) features extracted from the seismic signal of AAV3.
Sensors 18 01862 g003aSensors 18 01862 g003b
Figure 4. Distance between variables and between clusters.
Figure 4. Distance between variables and between clusters.
Sensors 18 01862 g004
Figure 5. Hierarchical clustering results of feature vectors of AAV3: (a) Result of acoustic feature vector; (b) result of seismic feature vector.
Figure 5. Hierarchical clustering results of feature vectors of AAV3: (a) Result of acoustic feature vector; (b) result of seismic feature vector.
Sensors 18 01862 g005
Figure 6. Acoustic–seismic mixed features of different vehicles: (a) AAV3; (b) DW3.
Figure 6. Acoustic–seismic mixed features of different vehicles: (a) AAV3; (b) DW3.
Sensors 18 01862 g006
Figure 7. Flowchart of the acoustic–seismic mixed feature extraction method based on WCER.
Figure 7. Flowchart of the acoustic–seismic mixed feature extraction method based on WCER.
Sensors 18 01862 g007
Table 1. Coefficients of quasi-orthogonal bi-orthogonal filters.
Table 1. Coefficients of quasi-orthogonal bi-orthogonal filters.
lh0(l)g0(l)lh0(l)g0(l)
00.56130.5601
−10.2865−0.296110.3030−0.2961
−2−0.0432−0.04702−0.0508−0.0470
−3−0.04650.05523−0.05820.0552
−40.01660.022040.02440.0220
−50.0055−0.010550.0112−0.0105
−6−0.0027−0.00586−0.0064−0.0057
−700.00187−0.00180.0018
−800.000780.00080.0007
−90−0.000390.0004−0.0003
−1000100.00010
−11001100
Table 2. Average classification accuracy and time consumption of feature extraction using different simplification methods. fa_HC, acoustic features after hierarchical clustering; fa_PCA, acoustic features after principal factor analysis.
Table 2. Average classification accuracy and time consumption of feature extraction using different simplification methods. fa_HC, acoustic features after hierarchical clustering; fa_PCA, acoustic features after principal factor analysis.
Using fa_HCUsing fa_PCA
Average classification accuracy66.98%63.52%
Time consumption of feature extraction4.1543 s4.1809 s
Table 3. Average classification accuracy and time needed for different classification methods using acoustic features without simplification. KNN, k-nearest neighbor; DT, decision tree; NB, naïve Bayes; SVM, support vector machine.
Table 3. Average classification accuracy and time needed for different classification methods using acoustic features without simplification. KNN, k-nearest neighbor; DT, decision tree; NB, naïve Bayes; SVM, support vector machine.
KNNDTNBSVM
Average classification accuracy71.55%65.77%64.29%67.54%
Average time needed for classification6.6225 s0.0029 s0.0051 s0.4282 s
Table 4. Classification accuracy with different feature vectors.
Table 4. Classification accuracy with different feature vectors.
RunnameUsingfanewUsingfsnewRunnameUsingfanewUsingfsnew
AAV591.3609%43.9645%DW585.4323%80.968%
AAV690.6977%51.292%DW684.5255%88.4185%
AAV790.0334%48.8294%DW725.4846%86.7161%
AAV891.7002%49.4762%DW829.8495%69.5652%
AAV999.9269%51.462%DW96.6007%72.7723%
AAV1089.5207%55.9371%DW1035.2547%91.1081%
AAV1188.8577%42.8371%DW1151.8025%94.4691%
DW1243.695%85.9726%
Table 5. Comparison of classification accuracy using different feature vectors.
Table 5. Comparison of classification accuracy using different feature vectors.
RunnameUsing faUsingfanewUsing fsUsingfsnewUsingfmix
AAV591.6568%91.3609%44.9704%43.9645%90.5917%
AAV691.7313%90.6977%51.2489%51.292%89.4488%
AAV790.6355%90.0334%50.903%48.8294%89.4314%
AAV892.3449%91.7002%51.0073%49.4762%90.2498%
AAV999.8538%99.9269%53.655%51.462%97.2953%
AAV1090.3076%89.5207%56.0086%55.9371%84.2275%
AAV1189.5599%88.8577%43.4457%42.8371%86.0955%
DW585.1034%85.4323%80.9211%80.968%79.3703%
DW685.1582%84.5255%89.2457%88.4185%87.6399%
DW726.1688%25.4846%85.8609%86.7161%71.2657%
DW831.1037%29.8495%68.7291%69.5652%60.9532%
DW97.75578%6.6007%72.5248%72.7723%49.9175%
DW1036.6399%35.2547%90.4379%91.1081%69.3923%
DW1152.0988%51.8025%94.6173%94.4691%72.2469%
DW1242.913%43.695% 85.9726%85.9726%75.8065%
Table 6. Average classification accuracy and time consumption using different feature vectors.
Table 6. Average classification accuracy and time consumption using different feature vectors.
Using faUsingfanewUsing fsUsingfsnewUsingfmix
Average classification accuracy67.54%66.98%67.97%67.59%79.60%
Average time consumption of feature extraction)4.1696 s4.1543 s7.3511 s7.2377 s15.3392 s
Average time consumption of classification0.4282 s0.2761 s0.6403 s0.3734 s0.3869 s
Table 7. Comparison of classification accuracy of different feature extraction methods.
Table 7. Comparison of classification accuracy of different feature extraction methods.
RunnameUsingfmixUsing FFT-Based FeaturesRunnameUsingfmixUsing FFT-Based Features
AAV590.5917%93.8462%DW579.3703%90.7895%
AAV689.4488%97.7606%DW687.6399%94.7445%
AAV789.4314%52.5753%DW771.2657%25.0855%
AAV890.2498%74.0532%DW860.9532%40.1338%
AAV997.2953%62.7924%DW949.9175%21.6172%
AAV1084.2275%88.7339%DW1069.3923%42.8061%
AAV1186.0955%90.2154%DW1172.2469%46.1728%
DW1275.8065%40.176%
Table 8. Average classification accuracy and time needed for different feature extraction methods.
Table 8. Average classification accuracy and time needed for different feature extraction methods.
UsingfmixUsingfanewUsing FFT-Based Features
Average classification accuracy79.60%66.98%64.10%
Efficiency (time consumption of feature extraction)15.3392 s4.1543 s5.5515 s
Time consumption of classification0.3869 s0.2761 s0.3749 s

Share and Cite

MDPI and ACS Style

Zhang, H.; Pan, Z.; Zhang, W. Acoustic–Seismic Mixed Feature Extraction Based on Wavelet Transform for Vehicle Classification in Wireless Sensor Networks. Sensors 2018, 18, 1862. https://doi.org/10.3390/s18061862

AMA Style

Zhang H, Pan Z, Zhang W. Acoustic–Seismic Mixed Feature Extraction Based on Wavelet Transform for Vehicle Classification in Wireless Sensor Networks. Sensors. 2018; 18(6):1862. https://doi.org/10.3390/s18061862

Chicago/Turabian Style

Zhang, Heng, Zhongming Pan, and Wenna Zhang. 2018. "Acoustic–Seismic Mixed Feature Extraction Based on Wavelet Transform for Vehicle Classification in Wireless Sensor Networks" Sensors 18, no. 6: 1862. https://doi.org/10.3390/s18061862

APA Style

Zhang, H., Pan, Z., & Zhang, W. (2018). Acoustic–Seismic Mixed Feature Extraction Based on Wavelet Transform for Vehicle Classification in Wireless Sensor Networks. Sensors, 18(6), 1862. https://doi.org/10.3390/s18061862

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop