Better and Faster Deep Image Fusion with Spatial Frequency
Citations Over Time
Abstract
Recent years have witnessed wide application of infrared and visible image fusion. However, most existing deep fusion methods focused primarily on improving the accuracy without taking much consideration of efficiency. In this paper, our goal is to build a better, faster and stronger image fusion method, which can reduce the computation complexity significantly while keep the fusion quality unchanged. To this end, we systematically analyzed the image fusion accuracy for different depth of image features and designed a lightweight backbone network with spatial frequency for infrared and visible image fusion. Unlikely previous methods based on traditional convolutional neural networks, our method can greatly preserve the detail information during image fusion. We analyze the spatial frequency strategy of our prototype and show that it can maintain more edges and textures information during fusion. Furthermore, our method has fewer parameters and lower computation in comparison of state-of-the-art fusion methods. Experiments conducted on benchmarks demonstrate that our method can achieve compelling fusion results over 97.0% decline of parameter size, running 5 times faster than state-of-the-art fusion methods.
Related Papers
- → Visible and infrared image fusion based on visual saliency detection(2020)3 cited
- → Impact of Multispectral Image Fusion Methods on Visual Search Behavior(2006)3 cited
- On the Image Fusion Rules of Strip Surface Defects Based on Dissimilarity Between Local Variances(2011)
- Multiscale Image Fusion Methods based on Improved Region Energy Rules(2010)
- Image fusion algorithm based on high frequency channel(2003)