Figure 1.
Block scheme of the simple VO system used in this research.
Figure 1.
Block scheme of the simple VO system used in this research.
Figure 2.
PUTKK dataset examples: Visualization of RGB-colored point clouds registered with the Kinect v1 and ground-truth camera poses for one of the PUTKK dataset sequences (a), sample Kinect v1 depth frame (b), and sample Kinect v2 depth frame (c) from this dataset.
Figure 2.
PUTKK dataset examples: Visualization of RGB-colored point clouds registered with the Kinect v1 and ground-truth camera poses for one of the PUTKK dataset sequences (a), sample Kinect v1 depth frame (b), and sample Kinect v2 depth frame (c) from this dataset.
Figure 3.
Block scheme of the Particle Swarm Optimization algorithm.
Figure 3.
Block scheme of the Particle Swarm Optimization algorithm.
Figure 4.
Block scheme of the adaptive Evolutionary Algorithm.
Figure 4.
Block scheme of the adaptive Evolutionary Algorithm.
Figure 5.
The U-Net architecture of a CNN is based on the Monodepth network, which is used for depth completion with RGB or RGB-D input from Kinect v1.
Figure 5.
The U-Net architecture of a CNN is based on the Monodepth network, which is used for depth completion with RGB or RGB-D input from Kinect v1.
Figure 6.
Block scheme of the fine-tuning procedure for the Monodepth model.
Figure 6.
Block scheme of the fine-tuning procedure for the Monodepth model.
Figure 7.
Search method for the best learning rate with FastAi (a) and learning results (b).
Figure 7.
Search method for the best learning rate with FastAi (a) and learning results (b).
Figure 8.
Depth maps: (a) Original, (b) Navier–Stokes (NS), (c) Telea, and (d) learned with RGB-D frames.
Figure 8.
Depth maps: (a) Original, (b) Navier–Stokes (NS), (c) Telea, and (d) learned with RGB-D frames.
Figure 9.
Kinect v1 depth maps: (a) Original, (b) estimated by the original Monodepth model, and (c) original depth image completed by learned depth with RGB inference.
Figure 9.
Kinect v1 depth maps: (a) Original, (b) estimated by the original Monodepth model, and (c) original depth image completed by learned depth with RGB inference.
Figure 10.
Colormap visualization of the difference between the estimated scene depth and the Kinect v2 ground-truth for the improved Monodepth model inference with Kinect v1 RGB frames only (a) and with both RGB and depth frames from Kinect v1 (b). See text for further explanation.
Figure 10.
Colormap visualization of the difference between the estimated scene depth and the Kinect v2 ground-truth for the improved Monodepth model inference with Kinect v1 RGB frames only (a) and with both RGB and depth frames from Kinect v1 (b). See text for further explanation.
Figure 11.
Trajectory estimation results for the putkk_Dataset_1_Kin_1 sequence for our VO system working with: (a,e) no inpainting, (b,f) NS inpainting, (c,g) Telea inpainting, or (d,h) learned depth with RGB-D inference. First row: ATE error plots; second: translational RPE plots.
Figure 11.
Trajectory estimation results for the putkk_Dataset_1_Kin_1 sequence for our VO system working with: (a,e) no inpainting, (b,f) NS inpainting, (c,g) Telea inpainting, or (d,h) learned depth with RGB-D inference. First row: ATE error plots; second: translational RPE plots.
Figure 12.
Trajectory estimation results for the putkk_Dataset_2_Kin_1 sequence for our VO system working with: (a,e) no inpainting, (b,f) NS inpainting, (c,g) Telea inpainting, or (d,h) learned depth with RGB-D inference. First row: ATE error plots; second: translational RPE plots.
Figure 12.
Trajectory estimation results for the putkk_Dataset_2_Kin_1 sequence for our VO system working with: (a,e) no inpainting, (b,f) NS inpainting, (c,g) Telea inpainting, or (d,h) learned depth with RGB-D inference. First row: ATE error plots; second: translational RPE plots.
Figure 13.
Trajectory estimation results for the putkk_Dataset_3_Kin_1 sequence for our VO system working with: (a,e) no inpainting, (b,f) NS inpainting, (c,g) Telea inpainting, or (d,h) learned depth with RGB-D inference. First row: ATE error plots; second: translational RPE plots.
Figure 13.
Trajectory estimation results for the putkk_Dataset_3_Kin_1 sequence for our VO system working with: (a,e) no inpainting, (b,f) NS inpainting, (c,g) Telea inpainting, or (d,h) learned depth with RGB-D inference. First row: ATE error plots; second: translational RPE plots.
Figure 14.
Trajectory estimation results for Kinect v1 frames on the putkk_Dataset_2_Kin_1 sequence for VO system working with (a,d) no inpainting, (b,e) learned depth completion with RGB inference, and (c,f) learned depth completion with RGB-D inference. The first row presents ATE error plots. The second includes translational RPE plots.
Figure 14.
Trajectory estimation results for Kinect v1 frames on the putkk_Dataset_2_Kin_1 sequence for VO system working with (a,d) no inpainting, (b,e) learned depth completion with RGB inference, and (c,f) learned depth completion with RGB-D inference. The first row presents ATE error plots. The second includes translational RPE plots.
Figure 15.
Trajectory estimation results for Kinect v1 frames on the putkk_Dataset_3_Kin_1 sequence for VO system working with (a,d) no inpainting, (b,e) learned depth completion with RGB inference, and (c,f) learned depth completion with RGB-D inference. The first row presents ATE error plots. The second includes translational RPE plots.
Figure 15.
Trajectory estimation results for Kinect v1 frames on the putkk_Dataset_3_Kin_1 sequence for VO system working with (a,d) no inpainting, (b,e) learned depth completion with RGB inference, and (c,f) learned depth completion with RGB-D inference. The first row presents ATE error plots. The second includes translational RPE plots.
Table 1.
Visual odometry parameters that were optimized jointly using PSO.
Table 1.
Visual odometry parameters that were optimized jointly using PSO.
Method | | | | | |
---|
PSO ATE | 0.060 | 0.890 | 0.068 | 0.892 | 0.00001 |
PSO RPE | 0.048 | 0.805 | 0.043 | 0.800 | 0.00001 |
Table 2.
ATE and RPE values on the putkk_Dataset_5_Kin_1 sequence for different parameter sets.
Table 2.
ATE and RPE values on the putkk_Dataset_5_Kin_1 sequence for different parameter sets.
Method/Params | ATE RMSE [m] | Trans. RPE RMSE [m] | Rot. RPE RMSE [°] |
---|
PSO ATE | 0.048 | 0.030 | 0.675 |
PSO RPE | 0.056 | 0.029 | 0.643 |
Table 3.
Comparison of ATE and RPE results on the putkk_Dataset_5_Kin_1 sequence for parameter sets obtained with different optimization procedure variants of the AKAZE detector threshold.
Table 3.
Comparison of ATE and RPE results on the putkk_Dataset_5_Kin_1 sequence for parameter sets obtained with different optimization procedure variants of the AKAZE detector threshold.
| putkk_Dataset_5_Kin_1 |
---|
Error Metric
| |
PSO ATE
|
PSO RPE
|
EA ATE
|
EA RPE
|
---|
ATE RMSE | [m] | 0.049 | 0.056 | 0.070 | 0.072 |
Trans. RPE RMSE | [m] | 0.030 | 0.029 | 0.035 | 0.033 |
Rot. RPE RMSE | [°] | 0.665 | 0.643 | 0.794 | 0.745 |
Table 4.
Trajectory estimation results for Kinect v1 frames with no depth inpainting, Telea and Navier–Stokes inpainting (radius 3), and learned depth with optimized parameters for the putkk_Dataset_5_Kin_1 sequence.
Table 4.
Trajectory estimation results for Kinect v1 frames with no depth inpainting, Telea and Navier–Stokes inpainting (radius 3), and learned depth with optimized parameters for the putkk_Dataset_5_Kin_1 sequence.
| | putkk_Dataset_5_Kin_1 |
---|
Metric
| |
No Inpainting
|
NS
|
Telea
|
Learned RGB-D Inference
|
---|
ATE RMSE | [m] | 0.198 | 0.201 | 0.204 | 0.049 |
Trans. RPE RMSE | [m] | 0.012 | 0.011 | 0.012 | 0.010 |
Rot. RPE RMSE | [°] | 0.174 | 0.187 | 0.183 | 0.665 |
Table 5.
Trajectory estimation results with conventional and deep learning methods for three PUTKK sequences (not used for training or optimization of parameters).
Table 5.
Trajectory estimation results with conventional and deep learning methods for three PUTKK sequences (not used for training or optimization of parameters).
| | putkk_Dataset_1_Kin_1 |
---|
Metric
| |
No Inpainting
|
NS
|
Telea
|
Learned RGB-D Inference
|
---|
ATE RMSE | [m] | 0.596 | 1.112 | 0.443 | 0.359 |
Trans. RPE RMSE | [m] | 0.009 | 0.021 | 0.016 | 0.015 |
Rot. RPE RMSE | [°] | 0.168 | 0.507 | 0.369 | 0.375 |
| | putkk_Dataset_2_Kin_1 |
ATE RMSE | [m] | 0.677 | 1.148 | 0.616 | 0.502 |
Trans. RPE RMSE | [m] | 0.010 | 0.030 | 0.033 | 0.020 |
Rot. RPE RMSE | [°] | 0.243 | 0.713 | 0.694 | 0.446 |
| | putkk_Dataset_3_Kin_1 |
ATE RMSE | [m] | 1.145 | 0.934 | 1.092 | 0.773 |
Trans. RPE RMSE | [m] | 0.012 | 0.033 | 0.030 | 0.012 |
Rot. RPE RMSE | [°] | 0.251 | 0.607 | 0.532 | 0.212 |
Table 6.
Trajectory optimization results for no inpainting and two variants of learned depth with optimized parameters.
Table 6.
Trajectory optimization results for no inpainting and two variants of learned depth with optimized parameters.
| | putkk_Dataset_5_Kin_1 |
---|
Metric
| |
No Inpainting
|
Learned RGB Inference
|
Learned RGB-D Inference
|
---|
ATE RMSE | [m] | 0.198 | 0.176 | 0.049 |
Trans. RPE RMSE | [m] | 0.012 | 0.015 | 0.012 |
Rot. RPE RMSE | [°] | 0.174 | 0.289 | 0.212 |
Table 7.
Trajectory estimation results with different deep learning methods for three PUTKK sequences (not used for training or optimization of parameters).
Table 7.
Trajectory estimation results with different deep learning methods for three PUTKK sequences (not used for training or optimization of parameters).
| | putkk_Dataset_1_Kin_1 |
---|
Metric | | No Inpainting | Learned RGB Inference | Learned RGB-D Inference |
ATE RMSE | [m] | 0.596 | 0.701 | 0.359 |
Trans. RPE RMSE | [m] | 0.009 | 0.022 | 0.015 |
Rot. RPE RMSE | [°] | 0.168 | 0.535 | 0.375 |
| | putkk_Dataset_2_Kin_1 |
ATE RMSE | [m] | 0.677 | 1.122 | 0.502 |
Trans. RPE RMSE | [m] | 0.010 | 0.027 | 0.020 |
Rot. RPE RMSE | [°] | 0.243 | 0.593 | 0.446 |
| | putkk_Dataset_3_Kin_1 |
ATE RMSE | [m] | 1.145 | 0.886 | 0.773 |
Trans. RPE RMSE | [m] | 0.012 | 0.027 | 0.019 |
Rot. RPE RMSE | [°] | 0.251 | 0.591 | 0.407 |
Table 8.
Execution time results for different depth completion methods on the putkk_Dataset_5_Kin_1 sequence.
Table 8.
Execution time results for different depth completion methods on the putkk_Dataset_5_Kin_1 sequence.
Depth Completion Time | NS | Telea | Learned RGB-D Inference |
---|
Mean value [ms] | 221.0 | 220.9 | 6.5 |
Standard deviation [ms] | 33.3 | 33.3 | 1.9 |