Image deblurring has the goal of restoring a sharp image from a degraded one. Currently, most deblurring methods are designed for natural images; these methods may not perform well when applied to remote sensing images. There are many differences between remote sensing and natural images, e.g., shooting distance, content complexity, and clarity. Therefore, a blind motion deblurring method specifically designed for remote sensing images called dual scale parallel spatial fusion network (DSPF-Net) is proposed. It has three innovative aspects: a dual-scale connection module is added between the two scales of the bottleneck layer and the decoder to realize the fusion of spatial detail and semantic features. Second, an adaptive spatial selection module is designed, which adds the function of selecting global and local spatial features. Finally, the cross-scale fusion (CSF) module is designed to restore the edge details and main structures by fusing the multi-scale features between the encoder and decoder. Extensive experiments are established on the synthetic dataset Blur-RS, the averaged peak signal-to-noise ratio and structural similarity are improved by 0.7916% and 0.0265%, respectively, compared to the best-performing comparison method. It shows that DSPF-Net has advantages in the task of blind motion deblurring of remote sensing images.
In recent years, the applications of hyperspectral imaging in the protection and analysis of cultural relics have received widespread attention. However, due to the limitation of imaging sensors, the spatial resolution of existing hyperspectral images is low, which hinders the development of hyperspectral digitization of cultural relics. Hyperspectral (HS) and RGB image fusion technology can generate hyperspectral images with high spatial resolution, which has gradually become a research hotspot. Inspired by the astounding performance of deep learning in various hyperspectral image processing tasks, this paper proposes a hyperspectral image fusion method based on dual-resolution fusion feature mutual guidance network (DRFFMG). Firstly, two feature extraction networks for HS and RGB images with different resolution pairs are designed to increase the richness of extracted features and reduce the loss of original hyperspectral information. Then, the spatial and spectral features extracted from the above feature extraction networks are fused, and a fusion feature mutual guidance module is designed to promote the mutual learning of different spatial features through information transmission, effectively reducing spatial distortion. Finally, the desired high spatial resolution HS image is restored from the fused features through an image reconstruction network. Experiments demonstrate that the proposed DRFFMG network can produce fusion images competitive with even better to state of the arts, and retain spectral information while improving spatial resolution.
KEYWORDS: Data hiding, Mining, Convolution, Machine learning, Hyperspectral imaging, Feature extraction, Data modeling, Principal component analysis, Data processing, Image fusion
Implicit information exploration techniques are of great importance for the restoration and conservation of cultural relics. At present, the hyperspectral image analysis technique is one of the main methods to extract hidden information, which mainly contains two analysis methods such as principal component analysis (PCA) and minimum noise fraction rotation (MNF), both of which have achieved certain information extraction effects. In recent years, with the development of artificial intelligence, deep learning, and other technologies, nonlinear methods such as neural networks are expected to further improve the effect of implicit information mining. Therefore, this paper is oriented to the problem of extracting hidden information from pottery artifacts and tries to study and explore the hidden information mining method based on deep neural networks, expecting to obtain more stable and richer hidden information. In this paper, an auto-encoder-based implied information mining method is proposed first, and the auto-encoder (AE) framework achieves good performance in feature learning by automatically learning low-dimensional embedding and reconstructing data. However, during the experiments, it is found that some important detailed information (e.g., implicit information) is often lost in the reconstruction process because the traditional autoencoder network only focuses more on the pixel-level reconstruction loss and ignores the overall distribution. Therefore, this paper further proposes a multi-scale convolutional autoencoder network (MSCAE). It constructs a multi-scale convolutional module based on the traditional AE and designs a cyclic consistency loss in addition to the reconstruction loss, to reduce the loss of detailed information in the reconstruction process and improve the implicit information mining effect. In the experiments, we find that the proposed method can achieve effective implied information mining by extracting implied information from cocoon-shaped pots, and its visual effect has been improved compared with the traditional AE network.
Video super-resolution (VSR) aims to generate high-resolution (HR) video by exploiting temporal consistency and contextual similarity of low-resolution (LR) video sequences. The key to improving the quality of VSR lies in accurate frame alignment and the feature fusion of adjacent frames. We propose a dual channel attention deep and shallow super-resolution network, which combines with HR optical flow compensation to construct an end-to-end VSR framework HOFADS-VSR (attention deep and shallow VSR network union HR optical flow compensation). HR optical flow calculated by spatiotemporal dependency of consecutive LR frames is used to compensate adjacent frames to implement accurate frame alignment. Deep and shallow channels with attention residual block restore small-scale detail features and large-scale contour features, respectively, and strengthen the rich features of global and local regions through weight adjustment. Extensive experiments have been performed to demonstrate the effectiveness and robustness of HOFADS-VSR. Comparative results on the Vid4, SPMC-12, and Harmonic-8 datasets show that our network not only achieves good performance on peak signal-to-noise ratio and structural similarity index but also the restored structure and texture have excellent fidelity.
Generative adversarial network (GAN) for super-resolution (SR) has attracted enormous interest in recent years. It has been widely used to solve the single-image super-resolution (SISR) task and made superior performance. However, GAN is rarely used for video super-resolution (VSR). VSR aims to improve video resolution by exploiting the temporal continuity and spatial similarity of video sequence frames. We design a GAN with multi-feature discriminators and combine it with optical flow estimation compensation to construct an end-to-end VSR framework OFC-MFGAN. Optical flow estimation compensation makes use of temporal continuity and spatial similarity features of adjacent frames to provide rich detailed information for GAN. Multi-feature discriminators based on visual attention mechanism include the pixel discriminator, edge discriminator, gray discriminator, and color discriminator. GAN with multi-feature discriminators makes the data distribution and visually sensitive features (edge, texture, and color) of SR frames similar to high-resolution frames. OFC-MFGAN effectively integrates the time, space, and visually sensitive features of videos. Extensive experiments on public video datasets and surveillance videos show the effectiveness and robustness of the proposed method. Compared with several state-of-the-art VSR methods and SISR methods, the proposed method can not only recover prominent edges, clear textures, and realistic colors but also make a pleasant visual feeling and competitive perceptual index.
We focus on the range migration (RM) and Doppler frequency migration (DFM) corrections in the long-time coherent integration, and a fast detection method based on two-dimensional trilinear autocorrelation function is proposed for the maneuvering target with jerk motion. This proposed method can integrate the echoes’ energy into peaks in a three-dimensional parameter space coherently and estimate the target’s radial range, acceleration, and jerk simultaneously by the peak detection technique. Then through the estimations of radial range, acceleration, and jerk, the radial velocity can be obtained through one-dimensional parameter searching. Finally, RM and DFM can be compensated simultaneously, and the target can be detected through the constant false alarm technique. This proposed method can strike a good balance between the computational complexity and detection performance. Experiments with the simulation and real measured radar data are conducted to verify the proposed method.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.