Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
A Platinized Carbon Fiber Microelectrode-Based Oxidase Biosensor for Amperometric Monitoring of Lactate in Brain Slices
Previous Article in Journal
The Use of Computational Geometry Techniques to Resolve the Issues of Coverage and Connectivity in Wireless Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

WCNN3D: Wavelet Convolutional Neural Network-Based 3D Object Detection for Autonomous Driving

by
Simegnew Yihunie Alaba
* and
John E. Ball
Department of Electrical and Computer Engineering, James Worth Bagley College of Engineering, Mississippi State University, Starkville, MS 39762, USA
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(18), 7010; https://doi.org/10.3390/s22187010
Submission received: 24 August 2022 / Revised: 5 September 2022 / Accepted: 14 September 2022 / Published: 16 September 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Three-dimensional object detection is crucial for autonomous driving to understand the driving environment. Since the pooling operation causes information loss in the standard CNN, we designed a wavelet-multiresolution-analysis-based 3D object detection network without a pooling operation. Additionally, instead of using a single filter like the standard convolution, we used the lower-frequency and higher-frequency coefficients as a filter. These filters capture more relevant parts than a single filter, enlarging the receptive field. The model comprises a discrete wavelet transform (DWT) and an inverse wavelet transform (IWT) with skip connections to encourage feature reuse for contrasting and expanding layers. The IWT enriches the feature representation by fully recovering the lost details during the downsampling operation. Element-wise summation was used for the skip connections to decrease the computational burden. We trained the model for the Haar and Daubechies (Db4) wavelets. The two-level wavelet decomposition result shows that we can build a lightweight model without losing significant performance. The experimental results on KITTI’s BEV and 3D evaluation benchmark show that our model outperforms the PointPillars-based model by up to 14% while reducing the number of trainable parameters.

1. Introduction

LiDAR 3D object detection in robotics applications, such as autonomous vehicles (AVs), is essential for detecting other vehicles, pedestrians, and cyclists. Technology development companies such as Waymo [1] and car manufacturing companies such as Bosch [2] have tested vehicles equipped with different sensors, including LiDAR. Because of the rapid growth of deep learning (DL) methods in computer vision, research on 3D object detection has increased. Due to this growth, the performance of DL methods has improved. Among these, CNNs have shown a significant performance improvement and state-of-the-art performance in several computer vision tasks, such as image classification and 2D object detection. Although the pooling operation reduces the size of the features at every layer for strides of more than one, it causes information loss, which can affect the detection performance.
Another limitation of CNN is the lack of interpretability. A CNN is a universal function approximator. However, it is considered a black-box model because of nonlinear mapping and unclear working mechanisms [3]. This characteristic makes it difficult to understand how CNN learns and interacts with other models. Thus, it is hard to understand the soundness and weaknesses of the model and diagnose, as well as correct, potential problems [3]. Because of these problems, model development often relies on trial and error [4].
The discrete wavelet transform (DWT) can also enlarge the receptive field by considering more spatial context and employing multilevel decomposition filters (using low-frequency and high-frequency filters at each level). We can also enlarge the receptive field by using filters of larger sizes, but this increases the computational cost. Dilated convolution is implemented to enlarge the receptive field [5]. However, dilated convolution suffers from a gridding effect [6] because of the combination of pixels from different neighbors.
In the signal-processing community, it is common to transform the spatial signal into a frequency domain representation, such as using a Fourier transform to use the dual property of convolution in the spatial domain and element-wise multiplication in the frequency domain. Rippel et al. [7] proposed a Fourier-transform-based convolutional neural network to use the dual property and tested it on the ImageNet [8] dataset. However, this is analogous to the conventional CNN in the frequency domain. The wavelet transform helps to keep both the spatial and frequency information. In this paper, we built a multiresolution analysis in the neural network with skip connections, WCNN3D: Wavelet Convolutional-Neural-Network-Based 3D Object Detection, to mitigate the information loss due to the pooling operation. The standard 2D convolution operates in the spatial domain, which drops the high-frequency component. Designing wavelet-based CNN helps use high-frequency components essential for learning and classifying images or other data. In a DWT, only the low-frequency component is decomposed into low-frequency and high-frequency components at each level of decomposition. However, we kept the high-frequency component in each multilevel DWT decomposition by concatenating the lower-frequency and higher-frequency components to enrich the features at each level of decomposition.
The point cloud data were encoded into point pillars to make the dataset suitable for 2D convolution and were input for the backbone and detection modules. PointPillars uses pointNets to learn a representation of point clouds in a vertical column (pillars) [9] instead of using fixed encoders. We can apply 2D convolution to the encoded pointpillar data, which reduces the computation burden of 3D convolutions. The rest of the paper is structured as follows. Section 2 presents related work. The background information and intuition of wavelets are summarized in Section 3. Section 4 summarizes the proposed model architecture and subnetworks, such as feature extraction and detection head, loss functions, and data augmentation techniques. Results and analyses of the experiment are presented in Section 5. The last Section 6 concludes the work.

2. Related Work

This section summarizes 3D object detection using camera images and LiDAR point clouds.

2.1. 3D Object Detection Using Camera Images

Camera sensors are rich in texture and color information but lack depth information. Thus, 3D object detection using the camera is challenging. Different methods have been proposed to solve the lack of depth and minimize the performance drop. Some methods use pseudo-LiDAR techniques, whereas others use stereo images and geometric constraints to solve the lack of depth in images [10]. The pseudo-LiDAR methods predict the depth of each image pixel and generate pseudo-LiDAR representation, such as [11,12]. The stereo-image-based methods use stereo images to predict depth information and further generate 3D bounding box information, such as [13,14]. On the other hand, geometric-based methods use different geometric constraints, such as ground planes and the object shape [15,16], to estimate depth information.
Although different techniques have been proposed to solve the camera’s lack of depth information, the performance of the proposed models is lower than 3D sensors such as LiDAR. For example, the pseudo-LiDAR data do not perform similarly to LiDAR data because of the image-to-LiDAR generation error [10].

2.2. 3D Object Detection Using LiDAR Point Clouds

Over the last few years, many LiDAR-based 3D object detection methods have been developed. The unstructured and sparse point cloud data are encoded into different representations to learn features: voxels, projections, and raw point cloud methods. VoxelNet [17] transforms point clouds into volumetric (voxel) representations and stacks the volumetric representations to form LiDAR point clouds. SECOND [18] improved voxelNet by introducing sparse convolution [19]. Voxel-FPN [20] also used voxel representation to conduct 3D object detection on point clouds. PointPillars [9] encoded the point cloud data into pillar representation to solve the computational burden of voxel representation due to 3D convolution. Some works, such as PointNet [21] and PointNet++ [22], directly process the raw point cloud for 3D object detection. The raw point cloud implementation has a high computational burden because of the 3D convolutions, besides the sparse and unstructured nature of the point cloud. On the one hand, works such as BirdNet+ [23] projected the LiDAR point cloud into bird’s-eye view (BEV) and applied 2D standard convolution on the projected data. Some methods use more than one sensor to solve the limitation of individual sensors and use the best out of each sensor. MV3D [24] and AVOD [25] fuse the projected BEV data with image data to enhance the 3D object detection performance. Another technique is integrating more than one LiDAR data representation, such as voxel and raw point cloud. PV-RCNN [26] integrated the voxel representation and the raw point cloud to obtain the best of the two representations. Although the model showed a large margin performance improvement over the PointPillars [9] network, the model’s parameter is larger due to the 3D convolutions.
Many works adopt the pillar representation, such as [27,28,29], because of its ability to learn feature representation instead of extracting features using fixed encoders like VoxelNet [17]. Additionally, pillar representation is fast because we can apply 2D convolutions to the extracted features instead of 3D convolutions. Caine et al. [27] proposed a pseudo-labeling domain adaptation method for a 3D object detection student–teacher network. Zhou et al. [29] presented a multiview fusion 3D object detection network from vertical column pillars and a perspective view. In this work, besides using the pillar encoding of point pillars to reduce the computational burden, we designed a new wavelet-based convolutional neural network for 3D object detection to minimize the information loss because of the pooling operation in the standard 2D convolution and to enlarge the receptive field.
Wavelet decomposition is well known and widely used in signal processing, but its usage in the computer vision community is limited. Works such as [30,31,32] embed wavelets into the neural network to carry out 2D classification, image segmentation, and image restoration. Fujieda and Takayama [30] presented a multi-resolution analysis and CNNs combination for texture classification and image annotation. The authors achieved a significant performance improvement over the CNN models with fewer parameters. Shen proposed WaveSNet [31], a wavelet-integrated deep network for image segmentation applications. The author integrated discrete wavelet transform into U-Net [33], SegNet [34], and DeepLabv3+ [35] models for effective image segmentation. Liu et al. [32] proposed a multi-level wavelet convolutional neural network (MWCNN) for image denoising, single image super-resolution, JPEG image artifacts removal, and classification. Even though the wavelet has shown significant performance in the above works and other signal processing tasks, it is not common in object detection, especially in 3D object detection. Three-dimensional object detection networks need to fulfill criteria such as being lightweight to run on real-time speed and accurate to give error-free information [36]. We introduce the first wavelet-based lightweight 3D object detection model for autonomous driving. This paper uses the term wavelet and discrete wavelet transforms interchangeably. We focus only on discrete wavelet transforms in this work, so both terms refer to the discrete wavelet transform.

2.3. Contributions

We designed a neural network by adding a high-frequency component to the standard spatial domain of CNN to enrich the feature representation. Additionally, we used the DWT as a downsampling operator to avoid information loss because of the pooling operation and to enlarge the receptive field using the multiresolution filters. The main contributions of the paper can be summarized as follows.
1.
To the best of our knowledge, our wavelet-based convolutional neural network model that incorporates the high-frequency component into the spatial domain to enhance the performance is the first work in 3D object detection.
2.
The model has high-frequency and low-frequency filters covering a large spatial domain to enlarge the receptive field, which improves the performance of the models.
3.
We removed the standard pooling operation of CNN to avoid information loss and to design wavelet-based CNN without pooling using the subsampling operation. The biorthogonal property of wavelets helps us to perform subsampling without information loss. We applied IWT to enrich the feature maps for the detection head and fully recover the lost detail information.
We show the performance of the model on the KITTI dataset. The model significantly improves the performance over the standard convolutional neural network models.

3. Wavelet-Based Convolutional Neural Network

Spectral approaches have been studied well in the signal and image processing community and have achieved significant results, such as [31,37,38]. To use the advantage of wavelets, we designed a multiresolution [39] wavelet-analysis-based CNN for 3D object detection. The multiresolution analysis property of wavelets is advantageous in detecting objects of different scales.
The 2D wavelet decomposition on images has approximate (lower-frequency) and detailed (higher-frequency) coefficients. Then, at every stage of decomposition, the lower-frequency component convolves with the image and decomposes further into lower-frequency and higher-frequency coefficients. In a single-stage decomposition, there are four convolutional filters: f L L , f L H , f H L , and f H H as shown in Figure 1. The high frequencies f L H , f H L , and f H H represent the horizontal, vertical, and diagonal components of the decomposition, respectively. The subscript L stands for low-pass filtered and H for high-pass filtered. The approximation information is represented by f L L , which contains the maximum information of the decomposition. For example, convolving an image x with the four convolutional filters decomposes the image into sub-band images of x L L , x L H , x H L , and x H H . The convolutional filters differ based on the wavelet, such as a Haar or a Daubechies wavelet. For example, for the Haar wavelet, the filter coefficients are defined as:
f L L = 1 / 2 1 1 1 1 , f L H = 1 / 2 1 1 1 1 , f H L = 1 / 2 1 1 1 1 , f H H = 1 / 2 1 1 1 1 .
The operation of DWT for sub-band images can be expressed as x L L =( f L L x)↓2, X L H = ( f L H x)↓2, x H L = ( f H L x)↓2, and x H H = ( f H H x) ↓2, where ∗ is the convolution operator and ↓2 is the downsampling/subsampling operator with a factor of two. Although there is a downsampling operation in the DWT, the original image/data x are fully recovered using the IWT without information loss due to the biorthogonal property of wavelets. In the standard convolution, deconvolution and max-unpooling operations are commonly used to increase the resolution of the feature map, but these operations cannot recover the lost data details. The IWT increases the resolution of the feature maps and helps to recover the lost data details. Generally, we used the DWT as a downsampling operator and the IWT as an upsampling operator.
The multiresolution analysis of convolution followed by the downsampling operation is equivalent to the standard convolution followed by the pooling operation. For a convolution layer of a standard CNN, for a training set x = ( x 0 , x 1 , . . . , x n 1 ) R n , we can obtain predicted labels y = ( y 0 , y 1 , . . . , y n 1 ) R n .
y n = i N n w i x i + b i a s ,
where n is the number of samples, N n is the set of indices of neighbors of x i , and w i is the weight. The label y can be written using a cross-correlation (i.e., convolution operator in most machine learning usage) ∗ as:
y = x w + b i a s ,
where x and w are vectors.
The convolution operation of the input x with weight w is analogous to the DWT sub-band image x convolving with the filters. The subsampling layer (pooling layer) is applied immediately to the convolution layer to simplify the information in the standard CNN. For example, average pooling with a stride of two reduces the number of outputs by half compared to the corresponding number of inputs. We omitted biases for the next operations to simplify the notation. We can express the generalized form of convolution followed by the pooling operation as:
y = ( x w ) p ,
where p is downsampling by a factor of p and p > 1. When p = 2, this operation is analogous to the multiresolution analysis of wavelets with lower-frequency and higher-frequency kernels. This operation can be written as:
x l = ( x f l ) 2 x h = ( x f h ) 2 ,
where f l is the lower-frequency kernel ( f L L ) and f h is the high-frequency kernel, which consists of f L H , f H L , and f H H . Equation (4) can be expressed as:
x L L = ( x f L L ) 2 x L H = ( x f L H ) 2 x H L = ( x f H L ) 2 x H H = ( x f H H ) 2 ,
The lower-frequency component is decomposed into hierarchical decomposition repeatedly for several decomposition levels. However, we used high-frequency and low-frequency kernels at each decomposition level to obtain richer information. At a decomposition level m, we can generalize the multiresolution analysis as:
x l , m + 1 = ( x f l , m ) 2 x h , m + 1 = ( x f h , m ) 2 ,
The standard CNN discards x h , m and uses only the lower-frequency kernel.
x l , m + 1 = ( x f l , m ) 2 ,
A single kernel is used in the standard CNN, but we used both low and high-frequency filters as a kernel. Using low and high-frequency filters increases the chance of taking into account more input pixels during the operation, which increases the field of view (receptive field). The receptive field is the size of a region in the input space that a particular feature looks at, which can be described by its central location and size. When the receptive field size increases, more relevant input parts are considered for detection [5]. The receptive field size of single-path networks for each layer can be calculated using the following arithmetic [40]:
r i = r i 1 + ( f 1 ) j i 1 ,
where r i is the receptive field size for layer i, f is a filter size of layer i, and j i 1 is the cumulative stride. The cumulative strides can be updated using j o u t = j i n s , where s is the stride. To understand the intuition of how the DWT increases the receptive field, let us compare the PointPillars [9] backbone and our backbone. All of the filers are 3 × 3 in the PointPillars network; however, we used a combination of low and high-frequency filters. In the DWT, only a low-frequency filter is decomposed in each decomposition layer, but we used a combination of low and high-frequency filters to increase the receptive field size. Each low and high-frequency filter has a size of 2 × 2. When four of the filters, f L L , f L H , f H L , a n d f H H , are combined, we can obtain 8 × 8 filters, which is larger than the PointPillars filter size.
Equation (8) works only for single-path networks (i.e., a single input to each layer). When multiple inputs to layer/layers exist, the receptive field size should be calculated for each input to the layer. Our network has multiple inputs from previous layers: both the downsampling and upsampling layers. The receptive field size increases when there are multiple inputs [41]. However, the detection /classification accuracy has a logarithmic relationship with the receptive field size [41]. Let us consider only the filter size by keeping all other parameters, such as padding, the same for both PointPillars [9] and our network. From (8), we can see that the receptive field size of our network is greater than the PointPillars network due to the larger filter size (i.e., 8 × 8 is greater than 3 × 3). However, due to the feature reuse design in the downsampling and upsampling layers, our proposed network has multiple inputs, potentially increasing the receptive field size further.
Note that the receptive field is not the only contributing factor to the performance improvement of DL networks. Other factors, such as the number of layers, number of filters per layer, and batch normalization, contribute to the performance of a DL model. Therefore, using a set of kernels with lower-frequency and higher-frequency filters increases the receptive field. Integrating the higher frequency into the standard CNN outperforms the existing 3D object detection models, as shown in the experiment in Section 5.

4. WCNN3D Network Architecture and Implementation Details

The WCNN3D network comprises three major components: the feature network, the base network, and the detection head, as shown in Figure 2. We used the pillar FeatureNet from the PointPillars [9] network to convert the raw point cloud into vertical columns’ pseudo image representation. The base network, our main network, transforms the pillar features into high-level representation. Finally, the detection head predicts classes of objects and regresses 3D bounding boxes for objects.

4.1. Pillar FeatureNet

The point cloud data were converted into a pillar representation to avoid the computational burden of 3D convolutions. We used the PointPillar representation from the PointPillars [9] network because it reduces the computational complexity by using 2D convolutions instead of 3D convolutions like VoxelNet [17] while maintaining a high detection performance. The raw point cloud with x, y, z and reflectance is discretized into a set of pillars on an evenly spaced x-y grid. A pillar is a voxel that has unlimited spatial extent in the z-direction [9]. Then, the points in each pillar were decorated with reflectance (r), x c , y c , z c , x p , and y p . The subscript c denotes the distance to the mean of all points in the pillar, whereas the subscript p denotes the offset from the pillar x and y centers. The LiDAR point at this stage is nine-dimensional (i.e., x, y, z, r, x c , y c , z c , x p , and y p ). Most of the pillars are empty due to the sparsity of LiDAR points, so a dense tensor of size (D, P, N) was created to reduce the sparsity of the decoded LiDAR point, where D denotes the dimension, P the number of non-empty pillars per sample, and N is the number of points per pillar. This operation reduces memory complexity by focusing on the non-empty pillars only. Applying a linear layer followed by BatchNorm [42] and ReLU [43] generates a tensor of size (C, P, N), where C denotes the number of channels. The encoded feature was scattered back to a pseudo-image of size (C, H, W), where H and W denote the height and width, respectively. This representation avoids the computational burden because of 3D convolutions.

4.2. Base Network

We designed a simple backbone network based on the DWT and IWT as shown in Figure 2. The network comprises the downsampling and upsampling layers. The downsampling layer performs a 2D DWT, convolution, and ReLU activation. The layer also comprises skip-connections of features from previous layers. The upsampling network was designed using the IWT to increase the feature map resolution and fully recover lost details. The wavelet’s lossless property helps to fully recover the lost information during downsampling, which is essential for the detection head network. We also designed skip-connections between consecutive layers of the downsampling and upsampling layers to enrich feature representation by encouraging feature reuse (knowledge preservation) and mitigating gradient problems for deep networks. An element-wise summation was applied for skip-connections instead of concatenation to decrease the computation burden. We also applied 1×1 convolution to reduce the channel dimensions and maintain the number of trainable parameters to be as small as possible while maintaining a high detection performance. The feature map size and the number of kernels used are given in Figure 2. The input to the first convolution layer is 248 × 296 with 64 kernels. The feature size decreases at each DWT operation and increases at each IWT operation. The batch normalization did not improve the performance, so we opted not to use it for the experiment. This condition might be due to the small number of batch sizes (two in this case because our machine cannot handle more than two batches). All weights were initialized randomly using the uniform distribution as in [44] and PointPillars [9].

4.3. Detection Head

We used the single shot detector (SSD) [45] as a detection head followed by regression for 3D bounding box generation. The SSD detection head comprises fully convolutional layers to predict four object predictions for each location (cell). Multiple predictions contain boundary box and confidence scores, which are essential in detecting objects of different scales and aspect ratios. Due to these properties of the SSD, we chose to use the SSD detection head as our object detection head. Then, regression was performed to obtain the 3D bounding box of objects. During the prediction level, non-maximum suppression (NMS) was also applied to suppress duplicate predictions.

4.4. Loss

The same loss function as that of VoxelNet [17], SECOND [18], and PointPillars [9], was used. The seven-point bounding box encoding technique (x, y, z, w, h, l, θ) was adopted. x, y, and z are the center of coordinates; w, l, h are the width, length, and height, respectively, and θ is the yaw rotation around the z-axis. The regression operation between the ground truth and anchors can be defined as:
Δ x = x g t x a d a , Δ y = y g t y a d a , Δ z = z g t z a d a
Δ w = log w g t w a , Δ h = log h g t h a , Δ l = log l g t l a
Δ θ = sin ( w g t w a ) ,
where the superscripts gt and a represent the ground truth and the anchor boxes, respectively. d a = ( w a ) 2 + ( l a ) 2 is the diagonal of the anchor box. The SmoothL1 loss and focal loss [46] were used as the localization loss and classification loss, respectively. We used a direction classifier loss, which uses the Softmax loss function, used for angle localization as in SECOND [18] and PointPillars [9]. The SmoothL1 loss can be expressed as:
L l o c = b ( x , y , z , w , l , h , θ ) S m o o t h L 1 ( Δ b ) ,
where Δ b can be Δ x when b is x, Δ y when b is y, Δ ϑ when b is θ, etc.
Similarly, the focal loss can be expressed as follows:
L c l s = α a ( 1 p a ) γ log ( p a ) ,
where γ is an anchor class probability and p a is an anchor class prediction. For fair comparison, we used the same values as [46] for α = 0.25 and γ = 2. The total loss, which includes the localization loss, classification loss, and directional loss, can be expressed as:
L = 1 N p o s ( β l o c L l o c + β c l s L c l s + β d i r L d i r ) ,
where N p o s is the number of positive anchors. We used β l o c = 2, β c l s = 1, and β d i r = 0.2, which is the same as [9,46]. Adam optimization with a learning rate of 3 ∗ 10−4 with a decay factor of 0.8 for every 15 epochs was used. We trained the model for 160 and 320 epochs with a batch size of two. The model converges faster with fewer epochs too.

4.5. Dataset and Data Augmentation

4.5.1. Dataset

We used the KITTI dataset [47], which comprises 7481 training and 7518 testing samples. The model was trained with the LiDAR point clouds that comprise car, pedestrian, and cyclist categories. Each class was evaluated on three difficulty levels: easy, moderate, and hard, based on object size, occlusion, and truncation levels (truncation level refers to what portion of an object is outside the camera view). We divided the original training set into 3712 training samples and 3769 validation samples like [13] and PointPillars [9]. Training the cars with one network and pedestrians and cyclists with another network has become common for 3D object detection [9,17,18,25]. Therefore, we followed the same practice to train cars for one network and pedestrians and cyclists for another network.
We used a maximum number of pillars of 12,000, a maximum number of points per pillar of 100, and a xy resolution of 0.16 m. We also trained the model for a maximum of 15,000 pillars and a maximum number of points per pillar of 120. However, increasing the number of points adds a computational burden, so the above values are optimal for performance and computation. At the inference time, NMS was applied to suppress the detection of less than a given IOU value. For the KITTI dataset, cars were evaluated with an IOU of 0.7, whereas pedestrians and cyclists were evaluated with an IOU of 0.5. The x, y, and z ranges of [(0, 70.4), (−40, 40), (−3, 1)] meters, respectively, were used for cars. A width of 1.6 m, length of 3.9 m, height of 1.5 m, and a z-center of −1 m were used for the car’s anchor box. Similarly, the x, y, z ranges for pedestrians and cyclists used were [(0, 48), (−20, 20), (−2.5, 0.5)] meters, respectively. A width of 0.6 m, length of 0.8 m, height of 1.73 m, and a z-center of −0.6 m were used for the pedestrian anchor box, whereas a width, length, and height of (0.6, 1.76, 1.73) meters, respectively, with a z-center of −0.6 m, were used for cyclist anchors. These values are commonly used on the KITTI dataset [47] for different works, such as [9,17,18].

4.5.2. 3D Data Augmentation

Data augmentation has been well-studied in image classification problems compared to object detection models. An appropriate data augmentation, such as color transformation and geometric transformation, improves the generalization of a model. We followed SECOND [18] and PointPillars [9] augmentation techniques. PointPillars created a lookup table for all ground truth 3D bounding box classes. Then, it randomly selected fifteen cars, zero pedestrians, and eight cyclists’ ground truth samples for each sample. However, SECOND [18] used eight, eight, and fifteen ground truth samples for pedestrians, cyclists, and cars, respectively. We used the PointPillars [9] setting for this experiment because of its better performance compared to SECOND [18]. Then, each bounding box was rotated using a uniformly generated rotation angle in the range of [ π / 2 , π / 2 ].
The local translation was also carried out with samples generated from a normal distribution of zero mean and a standard deviation of [0.25, 0.25, 0.25] for x, y, and z. Finally, global augmentation was performed to enrich the dataset further. A random mirror flip along the x-axis [9,48] was applied, followed by global rotation and scaling [9,17,18]. A global translation was applied that was drawn from a normal distribution with a mean of zero and standard deviation of [0.2, 0.2, 0.2] for x, y, and z. All experiments were conducted using NVIDIA GeForce RTX 2080 SUPER GPU running on Ubuntu 20.04. Code will be released https://github.com/Simeon340703/WCNN3D accessed on 28 August 2022.

5. Results and Discussions

We used the BEV and 3D KITTI datasets as official evaluation metrics for all experiments. Each class’ detection was evaluated based on the three difficulty levels: easy, moderate, and hard. The KITTI dataset’s mean average precision (mAP) was calculated based on the moderate difficulty level evaluation. Table 1 and Table 2 show the experimental results of our model compared with the base model PointPillars [9] and other related state-of-the-art models. Our model outperforms other networks in mAP for both BEV and 3D KITTI dataset evaluation benchmarks. We trained the car’s network for two, three, and four levels of decomposition, but we did not train the pedestrians and cyclists’ network for four levels of decomposition because of the small size of the objects to be decomposed after three levels of decomposition. The three difficult performance levels for cars, pedestrians, and cyclists show that our model’s performance is much higher than other models for both Haar wavelets and Daubechies (Db4) wavelets, except for PV-RCNN for car networks. Although the cars’ performance in the PV-RCNN model, which integrates the voxel form of LiDAR and the raw point cloud, is higher than our model, the pedestrians and cyclists’ network performance is much lower than our model. Additionally, the mAP of our model is higher than PV-RCNN for both BEV and 3D KITTI evaluation metrics. For both the Haar wavelet and Db4 wavelet, the four-level wavelet decomposition gives a higher performance than the three-level and two-level wavelets decomposition, but the drop in the two levels of the wavelet decomposition performance is insignificant compared to the decrease in training parameters. The Db4 performance is higher than the Haar wavelet performance, especially for pedestrians and cyclists.
Embedding the higher frequency into the spatial domain, which contains only the low frequency, enriches the features and helps the detection head for classification and regression tasks. Additionally, the network’s IWT part helps to recover lost details during downsampling. This recovery of lost details helps to increase the performance of the model. The performance improvement ratio of the model on smaller objects, such as pedestrians and cyclists, is greater than the cars’ performance improvement, as shown in Table 1 and Table 2. Therefore, recovering the lost details using IWT is helpful, especially for small objects. Note that most DL models struggle to detect small objects. The moderate and hard difficulty levels of the KITTI dataset decrease the performance of all of the models. However, the detection result of our model even increases at a higher rate for moderate and hard difficulty levels, which proves that adding the higher frequency component helps the model to generalize detection for such conditions.
Even though PV-RCNN improves the performance of the PointPillars by a large margin, the 3D convolution is still the bottleneck and is not a lightweight model for real-time processing. The major challenge of autonomous driving is increasing the performance while keeping the model lightweight. Our car model has 2.64 million, 3.33 million, and 5.6 million trainable parameters for two, three, and four decomposition levels. The pedestrians and cyclists model has 2 million and 2.97 million trainable parameters for two and three decomposition levels, respectively. The only lightweight model is the PointPillars network, with 5.9 million trainable parameters. The PV-RCNN model is not lightweight because of the 3D convolutions used for voxel processing. Therefore, our model is a lightweight model for even real-time processing. Our work will open the door to other researchers in this area to see the object detection network from the other direction to make 3D object detection interpretable, lightweight, and robust for real-time processing.
Our model produces an excellent result for cars, pedestrians, and cyclists, as shown in Figure 3. The network can detect strongly overlapped and occluded objects but struggles with a few overlapped, occluded, and faraway objects. Generally, the performance of our network, especially for small objects, such as pedestrians and cyclists, is promising. Figure 4 shows a qualitative comparison of the PointPillars [9] network and our proposed network. It is challenging to differentiate the output of different networks with visualization unless there is a miss detection. However, the images have a confidence score difference for each class. For example, the bottom car in the left and right images has different confidence scores. In the left image (PointPillars network), the bottom car has a 0.93 confidence score and an IOU of 0.83. On the other hand, the same car in the right image has a confidence score of 0.96 with an IOU of 0.87. Other cars in the left and right images also have a different confidence score, as shown in Figure 4.

6. Conclusions

In this work, we introduce WCNN3D, a new wavelet-based 3D object detection model for autonomous driving. The experimental results show that removing the pooling operation of CNN and replacing it with wavelet-based CNN increases the detection performance. Designing wavelet-based CNN helps use the high-frequency component of the data thrown away by the standard convolution to enrich features and increase the detection performance of the model. We also used the lower-frequency and high-frequency coefficients of the wavelet as a kernel for the convolution operation, which enlarges the receptive field. The detection result for two levels of wavelet decomposition shows that we can obtain lightweight models without losing much performance while decreasing the number of parameters. The model’s overall performance on the BEV and 3D KITTI evaluation metrics shows that our model outperforms the pillar-based LiDAR 3D object detection models and gives competitive results to sensor fusion-based state-of-the-art models. Therefore, the advantage of wavelet-based CNN is multi-fold. The results of Table 1 and Table 2 show that our model is more suitable for small object detection than other models.

Author Contributions

Conceptualization, S.Y.A.; methodology, S.Y.A.; software, S.Y.A.; validation, J.E.B.; formal analysis, S.Y.A.; investigation, S.Y.A.; data curation, S.Y.A.; writing—original draft preparation, S.Y.A.; writing—review and editing, J.E.B.; visualization, S.Y.A.; supervision, J.E.B.; project administration, J.E.B.; funding acquisition, J.E.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The KITTI public dataset is available here http://www.cvlibs.net/datasets/kitti/eval_object.php?obj_benchmark=3d (Accessed on 16 February 2022). Code will also be released. https://github.com/Simeon340703/WCNN3D accessed on 28 August 2022.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNNConvolutional Neural Network
DLDeep Learning
DWTDiscrete Wavelet Transform
IOUIntersection Over Union
IWTInverse Wavelet Transform
mAPMean Average Precision
MWCNNMulti-level Wavelet Convolutional Neural Network
NMSNon-maximal Suppression
ReLURectified Linear Unit
RCNNRegion-based Convolutional Neural Network
SSDSingle Shot MultiBox Detector

References

  1. Waymo. Waymo Driver-Waymo; Technical Report; WAYMO: Mountain View, CA, USA, 2022. [Google Scholar]
  2. Dechant, M. Self-Driving Car Technology—Between Man and Machine; Technical Report; Bosch: Gerlingen, Germany, 2022. [Google Scholar]
  3. Dong, Y.; Su, H.; Zhu, J.; Zhang, B. Improving interpretability of deep neural networks with semantic information. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4306–4314. [Google Scholar]
  4. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 818–833. [Google Scholar]
  5. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  6. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding convolution for semantic segmentation. In Proceedings of the 2018 IEEE winter conference on applications of computer vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar]
  7. Rippel, O.; Snoek, J.; Adams, R.P. Spectral Representations for Convolutional Neural Networks. arXiv 2015, arXiv:1506.03767. Available online: http://xxx.lanl.gov/abs/1506.03767 (accessed on 28 August 2022).
  8. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  9. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12697–12705. [Google Scholar]
  10. Alaba, S.; Ball, J. Deep Learning-based Image 3D Object Detection for Autonomous Driving: Review. TechRxiv 2022, Preprint. [Google Scholar] [CrossRef]
  11. Wang, Y.; Chao, W.L.; Garg, D.; Hariharan, B.; Campbell, M.; Weinberger, K.Q. Pseudo-lidar from visual depth estimation: Bridging the gap in 3D object detection for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8445–8453. [Google Scholar]
  12. Ma, X.; Wang, Z.; Li, H.; Zhang, P.; Ouyang, W.; Fan, X. Accurate monocular 3d object detection via color-embedded 3D reconstruction for autonomous driving. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 6851–6860. [Google Scholar]
  13. Chen, X.; Kundu, K.; Zhu, Y.; Berneshawi, A.G.; Ma, H.; Fidler, S.; Urtasun, R. 3D object proposals for accurate object class detection. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 424–432. [Google Scholar]
  14. Shi, Y.; Mi, Z.; Guo, Y. Stereo CenterNet based 3D Object Detection for Autonomous Driving. arXiv 2021, arXiv:2103.11071. [Google Scholar] [CrossRef]
  15. Mousavian, A.; Anguelov, D.; Flynn, J.; Kosecka, J. 3D bounding box estimation using deep learning and geometry. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7074–7082. [Google Scholar]
  16. Liu, Z.; Wu, Z.; Tóth, R. Smoke: Single-stage monocular 3D object detection via keypoint estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 996–997. [Google Scholar]
  17. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3D object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4490–4499. [Google Scholar]
  18. Yan, Y.; Mao, Y.; Li, B. Second: Sparsely embedded convolutional detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [PubMed]
  19. Graham, B.; van der Maaten, L. Submanifold sparse convolutional networks. arXiv 2017, arXiv:1706.01307. [Google Scholar]
  20. Kuang, H.; Wang, B.; An, J.; Zhang, M.; Zhang, Z. Voxel-FPN: Multi-scale voxel feature aggregation for 3D object detection from LIDAR point clouds. Sensors 2020, 20, 704. [Google Scholar] [CrossRef] [PubMed]
  21. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  22. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. arXiv 2017, arXiv:1706.02413. [Google Scholar]
  23. Barrera, A.; Guindel, C.; Beltrán, J.; García, F. Birdnet+: End-to-end 3D object detection in lidar bird’s eye view. In Proceedings of the 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC), Rhodes, Greece, 20–23 September 2020; pp. 1–6. [Google Scholar]
  24. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3D object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  25. Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. Joint 3D proposal generation and object detection from view aggregation. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–8. [Google Scholar]
  26. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. Pv-rcnn: Point-voxel feature set abstraction for 3D object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10529–10538. [Google Scholar]
  27. Caine, B.; Roelofs, R.; Vasudevan, V.; Ngiam, J.; Chai, Y.; Chen, Z.; Shlens, J. Pseudo-labeling for Scalable 3D Object Detection. arXiv 2021, arXiv:2103.02093. [Google Scholar]
  28. Xu, Q.; Zhou, Y.; Wang, W.; Qi, C.R.; Anguelov, D. Spg: Unsupervised domain adaptation for 3D object detection via semantic point generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 15446–15456. [Google Scholar]
  29. Zhou, Y.; Sun, P.; Zhang, Y.; Anguelov, D.; Gao, J.; Ouyang, T.; Guo, J.; Ngiam, J.; Vasudevan, V. End-to-end multi-view fusion for 3D object detection in lidar point clouds. In Proceedings of the Conference on Robot Learning, Virtual, 16–18 November 2020; pp. 923–932. [Google Scholar]
  30. Fujieda, S.; Takayama, K.; Hachisuka, T. Wavelet convolutional neural networks. arXiv 2018, arXiv:1805.08620. [Google Scholar]
  31. Li, Q.; Shen, L. Wavesnet: Wavelet integrated deep networks for image segmentation. arXiv 2020, arXiv:2005.14461. [Google Scholar]
  32. Liu, P.; Zhang, H.; Lian, W.; Zuo, W. Multi-level wavelet convolutional neural networks. IEEE Access 2019, 7, 74973–74985. [Google Scholar] [CrossRef]
  33. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  34. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  35. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  36. Alaba, S.; Gurbuz, A.; Ball, J. A Comprehensive Survey of Deep Learning Multisensor Fusion-based 3D Object Detection for Autonomous Driving: Methods, Challenges, Open Issues, and Future Directions. TechRxiv 2022, Preprint. [Google Scholar] [CrossRef]
  37. Fujieda, S.; Takayama, K.; Hachisuka, T. Wavelet convolutional neural networks for texture classification. arXiv 2017, arXiv:1707.07394. [Google Scholar]
  38. Kanchana, M.; Varalakshmi, P. Texture classification using discrete shearlet transform. Int. J. Sci. Res. 2013, 5, 3. [Google Scholar] [CrossRef]
  39. Mallat, S.G. Multifrequency channel decompositions of images and wavelet models. IEEE Trans. Acoust. Speech Signal Process. 1989, 37, 2091–2110. [Google Scholar] [CrossRef] [Green Version]
  40. Hien, D. A Guide to Receptive Field Arithmetic for Convolutional Neural Networks. 2017. Available online: https://medium.com/syncedreview/a-guide-to-receptive-field-arithmetic-for-convolutional-neural-networks-42f33d4378e0 (accessed on 28 August 2022).
  41. Araujo, A.; Norris, W.; Sim, J. Computing Receptive Fields of Convolutional Neural Networks. Distill 2019. Available online: https://distill.pub/2019/computing-receptive-fields (accessed on 28 August 2022). [CrossRef]
  42. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  43. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the ICML, Haifa, Israel, 21–24 June 2010. [Google Scholar]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  45. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  46. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  47. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  48. Yang, B.; Luo, W.; Urtasun, R. Pixor: Real-time 3D object detection from point clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June2018; pp. 7652–7660. [Google Scholar]
  49. Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum pointnets for 3D object detection from rgb-d data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June2018; pp. 918–927. [Google Scholar]
  50. Yang, B.; Liang, M.; Urtasun, R. Hdnet: Exploiting hd maps for 3D object detection. In Proceedings of the Conference on Robot Learning, Zurich, Switzerland, 29–31 October 2018; pp. 146–155. [Google Scholar]
Figure 1. Two−stage 2D wavelet decomposition. The lower-frequency filter coefficient (approximation coefficient) decomposes into lower and higher filter coefficients at each stage of the wavelet decomposition. The decomposition starts with rows downsampling by a factor of two, followed by the same operation along with the columns (the operation can be carried out vice versa, starting with a column and then a row), where c is the low-frequency coefficient (coarse coefficient), d is the high-frequency coefficient (detailed coefficient), g is the high-frequency filter, and h is the low-frequency filter. The ↓2 operator is the downsampling/subsampling operator with a factor of two. At every stage of wavelet decomposition, the low-frequency coefficient decomposed into high-frequency and low-frequency coefficients.
Figure 1. Two−stage 2D wavelet decomposition. The lower-frequency filter coefficient (approximation coefficient) decomposes into lower and higher filter coefficients at each stage of the wavelet decomposition. The decomposition starts with rows downsampling by a factor of two, followed by the same operation along with the columns (the operation can be carried out vice versa, starting with a column and then a row), where c is the low-frequency coefficient (coarse coefficient), d is the high-frequency coefficient (detailed coefficient), g is the high-frequency filter, and h is the low-frequency filter. The ↓2 operator is the downsampling/subsampling operator with a factor of two. At every stage of wavelet decomposition, the low-frequency coefficient decomposed into high-frequency and low-frequency coefficients.
Sensors 22 07010 g001
Figure 2. Architecture of the proposed WCNN3D object detection pedestrians and cyclists model for two levels of wavelet decomposition. Each convolutional block comprises three convolutional layers at each level of decomposition. We also used 1×1 convolutions to reduce the channel dimension of feature maps. The raw LiDAR data are encoded through the pillar feature network into a pillar representation. Then, the base network learns features using an expanding and contrasting wavelet convolutional network. The upsampling subnetwork recovers the lost details and increases the feature map of the downsampled features before detection. We did not apply the activation function for the last convolution layer. The upsampled features feed the detection head to predict 3D bounding boxes and classifications of objects. The convolution of down and up layers represent the convolutions of the contrasting and expanding layers, respectively. We also included the feature size (height and width ) and the number of kernels used in each operation. We used 3×3 kernel size, except for 1×1 convolutions, which used a 1×1 kernel size.
Figure 2. Architecture of the proposed WCNN3D object detection pedestrians and cyclists model for two levels of wavelet decomposition. Each convolutional block comprises three convolutional layers at each level of decomposition. We also used 1×1 convolutions to reduce the channel dimension of feature maps. The raw LiDAR data are encoded through the pillar feature network into a pillar representation. Then, the base network learns features using an expanding and contrasting wavelet convolutional network. The upsampling subnetwork recovers the lost details and increases the feature map of the downsampled features before detection. We did not apply the activation function for the last convolution layer. The upsampled features feed the detection head to predict 3D bounding boxes and classifications of objects. The convolution of down and up layers represent the convolutions of the contrasting and expanding layers, respectively. We also included the feature size (height and width ) and the number of kernels used in each operation. We used 3×3 kernel size, except for 1×1 convolutions, which used a 1×1 kernel size.
Sensors 22 07010 g002
Figure 3. (a) Cars, (b) pedestrians, (c) cyclists. Qualitative analysis of our LiDAR-based 3D object detection model on the KITTI dataset (zoomed view). For better visualization, we used the bird’s eye view (top) and image perspective (bottom). The ground truth boxes are shown in blue, whereas the predicted boxes are in green. Even though the cars (a) are overlapped, and some are occluded, the network can detect them.
Figure 3. (a) Cars, (b) pedestrians, (c) cyclists. Qualitative analysis of our LiDAR-based 3D object detection model on the KITTI dataset (zoomed view). For better visualization, we used the bird’s eye view (top) and image perspective (bottom). The ground truth boxes are shown in blue, whereas the predicted boxes are in green. Even though the cars (a) are overlapped, and some are occluded, the network can detect them.
Sensors 22 07010 g003
Figure 4. Sample qualitative car class output comparison of the (a) PointPillars [9] network and (b) our proposed network (zoomed view). For better visualization, we used the bird’s eye view (top) and image perspective (bottom). The output shows a class score for each image with the corresponding IOU value. The higher the IOU value, the more confident the prediction.
Figure 4. Sample qualitative car class output comparison of the (a) PointPillars [9] network and (b) our proposed network (zoomed view). For better visualization, we used the bird’s eye view (top) and image perspective (bottom). The output shows a class score for each image with the corresponding IOU value. The higher the IOU value, the more confident the prediction.
Sensors 22 07010 g004
Table 1. 3D object detection result for Haar wavelet and Daubechies (Db4) on the KITTI test BEV detection benchmark. Note that ‘2L’ refers to 2-level wavelet decomposition, ‘3L’ refers to 3-level wavelet decomposition, and ‘4L’ refers to 4-level wavelet decomposition. The best results are in bold.
Table 1. 3D object detection result for Haar wavelet and Daubechies (Db4) on the KITTI test BEV detection benchmark. Note that ‘2L’ refers to 2-level wavelet decomposition, ‘3L’ refers to 3-level wavelet decomposition, and ‘4L’ refers to 4-level wavelet decomposition. The best results are in bold.
MethodsmAPCarPedestrianCyclist
EasyMod.HardEasyMod.HardEasyMod.Hard
MV3D [24]-86.6278.9369.80------
F-pointNet [49]65.2091.1784.6774.7757.1349.5745.4877.2661.3753.78
PIXOR++ [50]-89.3883.7077.97------
VoxelNet [17]58.2589.3579.2677.3946.1340.7438.1166.7054.7650.55
SECOND [18]60.5688.0779.3777.9555.1046.2744.7673.6756.0448.78
PointPillars [9]66.1988.3586.1079.8358.6650.2347.1979.1962.2556.00
PV-RCNN [26]70.0494.9890.6586.1459.8650.5746.7482.4968.8962.41
WCNN3D (Haar-2L) ( ours)70.9989.7787.7686.3267.1262.7159.0480.7462.4959.28
WCNN3D (Haar-3L) (ours)71.0290.0587.5186.1169.2063.2959.2681.5462.2658.32
WCNN3D (Haar-4L) (ours)-90.0287.9086.35------
WCNN3D (Db4-2L) (ours)71.6690.1187.8386.2769.4864.5260.0783.6962.6359.45
WCNN3D (Db4-3L) (ours)71.8490.1287.9786.4668.4063.2059.3682.7864.3460.29
WCNN3D (Db4-4L) (ours)-90.2088.0486.31-----
Table 2. Three-dimensional object detection result for Haar wavelet and Daubechies (Db4) wavelet on the KITTI test 3D detection benchmark. Note that ‘2L’ refers to 2-level wavelet decomposition, ‘3L’ refers to 3-level wavelet decomposition, and ‘4L’ refers to 4-level wavelet decomposition. The best results are in bold.
Table 2. Three-dimensional object detection result for Haar wavelet and Daubechies (Db4) wavelet on the KITTI test 3D detection benchmark. Note that ‘2L’ refers to 2-level wavelet decomposition, ‘3L’ refers to 3-level wavelet decomposition, and ‘4L’ refers to 4-level wavelet decomposition. The best results are in bold.
MethodsmAPCarPedestrianCyclist
EasyMod.HardEasyMod.HardEasyMod.Hard
MV3D [24]-74.9763.6354.0------
F-pointNet [49]56.0482.1969.7960.5950.5342.1538.0872.2756.1749.01
VoxelNet [17]49.0577.4765.1157.7339.4833.6931.561.2248.3644.37
SECOND [18]56.6983.1373.6666.2041.0742.5637.2970.5153.8546.90
PointPillars [9]59.2079.0574.9968.3052.0843.5341.4975.7859.0752.92
PV-RCNN [26]62.8190.2581.4376.8252.1743.2940.2978.6063.7157.65
WCNN3D (Haar-2L) ( ours)64.3687.0977.3975.4958.6254.5750.0279.5661.1357.37
WCNN3D (Haar-3L) (ours)63.1687.8077.6175.7158.8653.4148.9479.7458.4754.71
WCNN3D (Haar-4L) (ours)-87.8477.6776.00------
WCNN3D (Db4-2L) (ours)65.4187.7577.5675.4061.9357.6752.0682.7461.0157.66
WCNN3D (Db4-3L) (ours)64.1688.5778.0476.1657.8853.7549.8280.1460.7056.70
WCNN3D (Db4-4L) (ours)-87.8377.7575.74------
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alaba, S.Y.; Ball, J.E. WCNN3D: Wavelet Convolutional Neural Network-Based 3D Object Detection for Autonomous Driving. Sensors 2022, 22, 7010. https://doi.org/10.3390/s22187010

AMA Style

Alaba SY, Ball JE. WCNN3D: Wavelet Convolutional Neural Network-Based 3D Object Detection for Autonomous Driving. Sensors. 2022; 22(18):7010. https://doi.org/10.3390/s22187010

Chicago/Turabian Style

Alaba, Simegnew Yihunie, and John E. Ball. 2022. "WCNN3D: Wavelet Convolutional Neural Network-Based 3D Object Detection for Autonomous Driving" Sensors 22, no. 18: 7010. https://doi.org/10.3390/s22187010

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop