Author Contributions
Conceptualization, X.Z. and M.H.; methodology, X.Z.; software, X.Z.; validation, X.Z., F.X. and M.H.; writing—original draft preparation, X.Z.; writing—review and editing, M.H., D.Z. and S.J. All authors have read and agreed to the published version of the manuscript.
Figure 1.
The structure of the ConvNeXt block.
Figure 1.
The structure of the ConvNeXt block.
Figure 2.
The overall processes of our framework. Two independent branches are designed to extract the features from different modalities. In each branch, RPCM is used to extract features with larger receptive fields, and dense connections are applied between RPCMs to reuse intermediate features. Then, TCCM is used to compensate for texture details and contrast of features.
Figure 2.
The overall processes of our framework. Two independent branches are designed to extract the features from different modalities. In each branch, RPCM is used to extract features with larger receptive fields, and dense connections are applied between RPCMs to reuse intermediate features. Then, TCCM is used to compensate for texture details and contrast of features.
Figure 3.
The structure of RPCM. It consists of a convolutional layer and residual connections of three PCBs. Each PCB is constructed by integrating a 7 × 7 regular convolution into the first layer of the ConvNeXt block.
Figure 3.
The structure of RPCM. It consists of a convolutional layer and residual connections of three PCBs. Each PCB is constructed by integrating a 7 × 7 regular convolution into the first layer of the ConvNeXt block.
Figure 4.
The architecture of TCCM. TCCM consists of two gradient residuals with SimAM attention and a CSAM, where the specific implementation of the CSAM is indicated in the orange rectangle.
Figure 4.
The architecture of TCCM. TCCM consists of two gradient residuals with SimAM attention and a CSAM, where the specific implementation of the CSAM is indicated in the orange rectangle.
Figure 5.
Ablation analysis of loss function on the TNO dataset.
Figure 5.
Ablation analysis of loss function on the TNO dataset.
Figure 6.
Ablation analysis of the network structure on the TNO dataset.
Figure 6.
Ablation analysis of the network structure on the TNO dataset.
Figure 7.
Subjective comparison of different fusion methods on 00123D from the MSRS dataset.
Figure 7.
Subjective comparison of different fusion methods on 00123D from the MSRS dataset.
Figure 8.
Subjective comparison of different fusion methods on 00537D from the MSRS dataset.
Figure 8.
Subjective comparison of different fusion methods on 00537D from the MSRS dataset.
Figure 9.
Subjective comparison of different fusion methods on 00858N from the MSRS dataset.
Figure 9.
Subjective comparison of different fusion methods on 00858N from the MSRS dataset.
Figure 10.
Subjective comparison of different fusion methods on 01024N from the MSRS dataset.
Figure 10.
Subjective comparison of different fusion methods on 01024N from the MSRS dataset.
Figure 11.
Subjective comparison of different fusion methods on Kaptain_1654 from the TNO dataset.
Figure 11.
Subjective comparison of different fusion methods on Kaptain_1654 from the TNO dataset.
Figure 12.
Subjective comparison of different fusion methods on soldier_in_trench_2 from the TNO dataset.
Figure 12.
Subjective comparison of different fusion methods on soldier_in_trench_2 from the TNO dataset.
Figure 13.
Subjective comparison of different fusion methods on a representative example from the MFD dataset.
Figure 13.
Subjective comparison of different fusion methods on a representative example from the MFD dataset.
Table 1.
Ablation study of loss function on the TNO dataset (the best results are in bold, and the second-best results are underlined).
Table 1.
Ablation study of loss function on the TNO dataset (the best results are in bold, and the second-best results are underlined).
Loss Function | EN | SF | AG | SD | SCD | VIF | Q | SSIM | MS-SSIM |
---|
Without-ssim | 7.0570 | 12.7583 | 4.9627 | 44.8685 | 1.6374 | 0.7672 | 0.5999 | 1.0168 | 1.0257 |
Without-int | 6.9578 | 11.0307 | 4.3695 | 39.4983 | 1.7283 | 0.6511 | 0.5380 | 0.9954 | 1.0785 |
Without-text | 6.9374 | 8.9991 | 3.4665 | 43.0512 | 1.6836 | 0.6661 | 0.4540 | 1.0434 | 1.0762 |
DTFusion | 7.0729 | 12.6585 | 4.9350 | 45.8861 | 1.7614 | 0.7199 | 0.5660 | 1.0445 | 1.0788 |
Table 2.
Ablation study of the network structure on the TNO dataset (the best results are in bold, and the second-best results are underlined).
Table 2.
Ablation study of the network structure on the TNO dataset (the best results are in bold, and the second-best results are underlined).
Models | EN | SF | AG | SD | SCD | VIF | Q | SSIM | MS-SSIM |
---|
No-RPCM | 7.0417 | 12.6129 | 4.8514 | 44.4450 | 1.6988 | 0.7598 | 0.5863 | 1.0539 | 1.0628 |
No-PConv | 7.0563 | 12.6217 | 4.9650 | 44.6881 | 1.7133 | 0.7457 | 0.5947 | 1.0371 | 1.0683 |
No-TCCM | 7.0063 | 12.0825 | 4.6448 | 43.1418 | 1.6482 | 0.7363 | 0.5645 | 1.0697 | 1.0648 |
DTFusion | 7.0729 | 12.6585 | 4.9350 | 45.8861 | 1.7614 | 0.7199 | 0.5660 | 1.0445 | 1.0788 |
Table 3.
Objective comparison of different fusion methods on the MSRS dataset (the best results are in bold, and the second-best results are underlined).
Table 3.
Objective comparison of different fusion methods on the MSRS dataset (the best results are in bold, and the second-best results are underlined).
Models | EN | SF | AG | SD | SCD | VIF | Q | SSIM | MS-SSIM |
---|
DenseFuse | 5.9455 | 6.0676 | 2.0973 | 23.5937 | 1.2519 | 0.6845 | 0.3651 | 0.9019 | 1.0070 |
CSF | 5.8701 | 7.1226 | 2.3864 | 26.6909 | 1.3490 | 0.6119 | 0.3697 | 0.7672 | 0.9813 |
RFN-Nest | 6.2102 | 6.1924 | 2.1460 | 29.1218 | 1.4717 | 0.6495 | 0.3873 | 0.7604 | 1.0090 |
SDNet | 5.2526 | 8.6887 | 2.7218 | 17.3375 | 0.9880 | 0.4948 | 0.3726 | 0.7322 | 0.8932 |
GANMcC | 6.2296 | 5.9199 | 2.2062 | 28.2696 | 1.4816 | 0.6273 | 0.3298 | 0.8082 | 0.9734 |
U2Fusion | 5.3743 | 9.0844 | 2.8687 | 25.5003 | 1.2428 | 0.5297 | 0.4076 | 0.7635 | 0.9099 |
SMoA | 6.1164 | 7.5615 | 2.5667 | 28.8636 | 1.3884 | 0.6566 | 0.4339 | 0.8599 | 1.0094 |
SeAFusion | 6.7031 | 11.5627 | 3.8788 | 43.0600 | 1.7014 | 0.9749 | 0.6653 | 0.9759 | 1.0363 |
UNFusion | 6.5652 | 10.4136 | 3.3311 | 40.9665 | 1.6590 | 0.9456 | 0.6513 | 0.9984 | 1.0476 |
DATFuse | 6.5043 | 10.9776 | 3.6513 | 36.5945 | 1.4092 | 0.9023 | 0.6394 | 0.9060 | 1.0046 |
DTFusion | 6.6331 | 12.6396 | 4.0827 | 44.1737 | 1.7358 | 0.9200 | 0.6550 | 1.0199 | 1.0685 |
Table 4.
Objective comparison of different fusion methods on Kaptain_1654 from the TNO dataset (the best results are in bold, and the second-best results are underlined).
Table 4.
Objective comparison of different fusion methods on Kaptain_1654 from the TNO dataset (the best results are in bold, and the second-best results are underlined).
Models | EN | SF | AG | SD | SCD | VIF | Q | SSIM | MS-SSIM |
---|
DenseFuse | 6.3276 | 6.3882 | 2.5563 | 27.3346 | 1.5923 | 0.5736 | 0.3268 | 1.0078 | 0.8233 |
CSF | 6.5561 | 7.2109 | 3.0127 | 32.7923 | 1.7185 | 0.5393 | 0.3621 | 0.9734 | 0.8851 |
RFN-Nest | 6.5620 | 5.0514 | 2.1920 | 31.7220 | 1.6656 | 0.5214 | 0.2983 | 0.7364 | 0.8402 |
SDNet | 6.2167 | 10.9474 | 4.3723 | 24.4714 | 1.5514 | 0.4706 | 0.3880 | 1.0114 | 0.8498 |
GANMcC | 6.6511 | 5.4166 | 2.3573 | 35.3765 | 1.7362 | 0.5031 | 0.2565 | 0.8921 | 0.8195 |
U2Fusion | 6.8309 | 10.2560 | 4.4312 | 34.0316 | 1.7381 | 0.5588 | 0.4257 | 1.0221 | 0.9091 |
SMoA | 6.4447 | 6.3792 | 2.6192 | 32.7428 | 1.6986 | 0.5007 | 0.3122 | 0.7795 | 0.8306 |
SeAFusion | 6.7105 | 11.2986 | 4.6320 | 41.4455 | 1.7177 | 0.5796 | 0.4512 | 0.8727 | 0.8911 |
UNFusion | 6.4740 | 8.7765 | 3.3405 | 36.8532 | 1.6765 | 0.7865 | 0.5305 | 1.0426 | 0.8576 |
DATFuse | 6.2341 | 9.5534 | 3.6206 | 29.7492 | 1.4967 | 0.7020 | 0.5107 | 0.9582 | 0.7725 |
DTFusion | 6.7212 | 12.4824 | 4.8506 | 42.7354 | 1.7761 | 0.6860 | 0.5579 | 0.9889 | 0.9248 |
Table 5.
Objective comparison of different fusion methods on soldier_in_trench_2 from the TNO dataset (the best results are in bold, and the second-best results are underlined).
Table 5.
Objective comparison of different fusion methods on soldier_in_trench_2 from the TNO dataset (the best results are in bold, and the second-best results are underlined).
Models | EN | SF | AG | SD | SCD | VIF | Q | SSIM | MS-SSIM |
---|
DenseFuse | 6.3070 | 7.5459 | 2.7402 | 20.0567 | 1.5948 | 0.6927 | 0.3953 | 0.8886 | 0.9066 |
CSF | 6.9345 | 8.8690 | 3.8940 | 31.7183 | 1.7707 | 0.6516 | 0.5417 | 0.9249 | 0.9967 |
RFN-Nest | 7.0609 | 7.4689 | 3.2704 | 33.2321 | 1.8660 | 0.6336 | 0.5001 | 0.8497 | 0.9975 |
SDNet | 6.0010 | 14.7995 | 5.4140 | 19.1660 | 1.5612 | 0.6632 | 0.6035 | 0.9788 | 0.8821 |
GANMcC | 7.0623 | 7.0105 | 2.7162 | 34.8298 | 1.8269 | 0.6139 | 0.3044 | 0.8123 | 0.9479 |
U2Fusion | 6.5952 | 13.0062 | 5.2435 | 24.9478 | 1.7591 | 0.7047 | 0.5988 | 0.9762 | 0.9933 |
SMoA | 6.4538 | 7.0206 | 2.3414 | 22.7674 | 1.6671 | 0.5060 | 0.2235 | 0.6436 | 0.9036 |
SeAFusion | 6.7607 | 13.2970 | 4.9708 | 30.0178 | 1.7232 | 0.5967 | 0.4465 | 0.8141 | 0.9498 |
UNFusion | 6.6805 | 11.3169 | 3.7477 | 29.6441 | 1.6904 | 0.9285 | 0.5403 | 0.9761 | 0.9639 |
DATFuse | 6.0989 | 9.2577 | 2.9807 | 20.1861 | 1.4767 | 0.6958 | 0.4169 | 0.8643 | 0.8581 |
DTFusion | 6.7794 | 13.7397 | 5.4402 | 31.0218 | 1.7768 | 0.7701 | 0.6868 | 0.9435 | 0.9543 |
Table 6.
Objective comparison of different fusion methods on 20 image pairs from the TNO dataset (the best results are in bold, and the second-best results are underlined).
Table 6.
Objective comparison of different fusion methods on 20 image pairs from the TNO dataset (the best results are in bold, and the second-best results are underlined).
Models | EN | SF | AG | SD | SCD | VIF | Q | SSIM | MS-SSIM |
---|
DenseFuse | 6.5312 | 7.1421 | 2.7540 | 28.2952 | 1.5355 | 0.5997 | 0.3581 | 1.0187 | 1.0348 |
CSF | 6.9622 | 9.0222 | 3.7094 | 39.1585 | 1.7211 | 0.5920 | 0.4140 | 0.9579 | 1.0824 |
RFN-Nest | 6.9988 | 6.1282 | 2.7297 | 38.4950 | 1.7179 | 0.5640 | 0.3373 | 0.8064 | 1.0476 |
SDNet | 6.6880 | 12.0056 | 4.6001 | 34.1692 | 1.4700 | 0.5747 | 0.4447 | 1.0092 | 1.0253 |
GANMcC | 6.9580 | 6.7637 | 2.7919 | 38.8544 | 1.6641 | 0.5450 | 0.3079 | 0.8900 | 1.0260 |
U2Fusion | 6.9716 | 11.8527 | 4.8584 | 38.6913 | 1.7087 | 0.6076 | 0.4494 | 0.9835 | 1.0790 |
SMoA | 6.7095 | 7.3701 | 2.8099 | 33.3535 | 1.6411 | 0.5230 | 0.3291 | 0.8108 | 1.0424 |
SeAFusion | 7.1809 | 12.4793 | 4.9169 | 47.1625 | 1.7248 | 0.6138 | 0.4488 | 0.8744 | 1.0467 |
UNFusion | 7.0402 | 10.5738 | 3.9037 | 44.0714 | 1.6746 | 0.8613 | 0.5580 | 1.0430 | 1.0609 |
DATFuse | 6.6465 | 10.7138 | 3.9266 | 31.7927 | 1.4900 | 0.7049 | 0.5083 | 0.9507 | 0.9679 |
DTFusion | 7.0729 | 12.6585 | 4.9350 | 45.8861 | 1.7614 | 0.7199 | 0.5660 | 1.0445 | 1.0788 |
Table 7.
Objective comparison of different fusion methods on 300 image pairs from the MFD dataset (the best results are in bold, and the second-best results are underlined).
Table 7.
Objective comparison of different fusion methods on 300 image pairs from the MFD dataset (the best results are in bold, and the second-best results are underlined).
Models | EN | SF | AG | SD | SCD | VIF | Q | SSIM | MS-SSIM |
---|
DenseFuse | 6.4268 | 8.057 | 2.8312 | 24.9986 | 1.5358 | 0.5955 | 0.3793 | 0.9416 | 1.0463 |
CSF | 6.8299 | 10.2153 | 3.7065 | 33.8474 | 1.7298 | 0.6197 | 0.4788 | 0.9340 | 1.0825 |
RFN-Nest | 6.8271 | 8.2239 | 3.0637 | 32.8573 | 1.7279 | 0.5745 | 0.4155 | 0.8007 | 1.0444 |
SDNet | 6.8481 | 14.3417 | 5.0301 | 35.6101 | 1.5508 | 0.5613 | 0.5372 | 0.9690 | 0.9391 |
GANMcC | 6.8482 | 7.7419 | 2.7907 | 33.3603 | 1.6540 | 0.5359 | 0.3195 | 0.8452 | 0.9873 |
U2Fusion | 6.9532 | 13.8198 | 5.1340 | 35.2004 | 1.7635 | 0.6612 | 0.5668 | 0.9696 | 1.0835 |
SMoA | 6.6054 | 8.0088 | 2.7677 | 28.6175 | 1.6261 | 0.4939 | 0.3015 | 0.6419 | 0.9655 |
SeAFusion | 7.0037 | 15.3494 | 5.4669 | 37.1395 | 1.6843 | 0.5821 | 0.5152 | 0.7518 | 0.9897 |
UNFusion | 6.8598 | 12.1175 | 4.0451 | 34.4041 | 1.5702 | 0.7594 | 0.5477 | 1.0049 | 1.0344 |
DATFuse | 6.4019 | 10.4622 | 3.4408 | 26.3103 | 1.2867 | 0.6444 | 0.4935 | 0.9193 | 0.9343 |
DTFusion | 6.9707 | 17.1730 | 5.8161 | 37.5127 | 1.6856 | 0.7615 | 0.6548 | 0.9506 | 1.0432 |
Table 8.
Comparison of running time of different methods on the MSRS, TNO, and MFD datasets (the best results are in bold, and the second-best results are underlined).
Table 8.
Comparison of running time of different methods on the MSRS, TNO, and MFD datasets (the best results are in bold, and the second-best results are underlined).
Models | MSRS (s) | TNO (s) | MFD (s) |
---|
DenseFuse | 0.0778 | 0.0864 | 0.1896 |
CSF | 32.2675 | 31.9109 | 77.0992 |
RFN-Nest | 0.1361 | 0.2166 | 0.2613 |
SDNet | 0.0239 | 0.0153 | 0.0569 |
GANMcC | 5.4610 | 5.4283 | 13.2267 |
U2Fusion | 0.1897 | 0.1828 | 0.5035 |
SMoA | 8.2952 | 8.6730 | 9.1976 |
SeAFusion | 0.1307 | 0.5514 | 0.2711 |
UNFusion | 0.1102 | 0.2673 | 0.2491 |
DATFuse | 0.0203 | 0.0269 | 0.0429 |
DTFusion | 0.3765 | 0.5447 | 0.8927 |