Deep convolutional neural networks(CNNs)have demonstrated remarkable performance in video super-resolution(VSR).However,the ability of most existing methods to recover fine details in complex scenes is often hindered ...Deep convolutional neural networks(CNNs)have demonstrated remarkable performance in video super-resolution(VSR).However,the ability of most existing methods to recover fine details in complex scenes is often hindered by the loss of shallow texture information during feature extraction.To address this limitation,we propose a 3D Convolutional Enhanced Residual Video Super-Resolution Network(3D-ERVSNet).This network employs a forward and backward bidirectional propagation module(FBBPM)that aligns features across frames using explicit optical flow through lightweight SPyNet.By incorporating an enhanced residual structure(ERS)with skip connections,shallow and deep features are effectively integrated,enhancing texture restoration capabilities.Furthermore,3D convolution module(3DCM)is applied after the backward propagation module to implicitly capture spatio-temporal dependencies.The architecture synergizes these components where FBBPM extracts aligned features,ERS fuses hierarchical representations,and 3DCM refines temporal coherence.Finally,a deep feature aggregation module(DFAM)fuses the processed features,and a pixel-upsampling module(PUM)reconstructs the high-resolution(HR)video frames.Comprehensive evaluations on REDS,Vid4,UDM10,and Vim4 benchmarks demonstrate well performance including 30.95 dB PSNR/0.8822 SSIM on REDS and 32.78 dB/0.8987 on Vim4.3D-ERVSNet achieves significant gains over baselines while maintaining high efficiency with only 6.3M parameters and 77ms/frame runtime(i.e.,20×faster than RBPN).The network’s effectiveness stems from its task-specific asymmetric design that balances explicit alignment and implicit fusion.展开更多
Background Recurrent recovery is a common method for video super-resolution(VSR)that models the correlation between frames via hidden states.However,the application of this structure in real-world scenarios can lead t...Background Recurrent recovery is a common method for video super-resolution(VSR)that models the correlation between frames via hidden states.However,the application of this structure in real-world scenarios can lead to unsatisfactory artifacts.We found that in real-world VSR training,the use of unknown and complex degradation can better simulate the degradation process in the real world.Methods Based on this,we propose the RealFuVSR model,which simulates real-world degradation and mitigates artifacts caused by the VSR.Specifically,we propose a multiscale feature extraction module(MSF)module that extracts and fuses features from multiple scales,thereby facilitating the elimination of hidden state artifacts.To improve the accuracy of the hidden state alignment information,RealFuVSR uses an advanced optical flow-guided deformable convolution.Moreover,a cascaded residual upsampling module was used to eliminate noise caused by the upsampling process.Results The experiment demonstrates that RealFuVSR model can not only recover high-quality videos but also outperforms the state-of-the-art RealBasicVSR and RealESRGAN models.展开更多
This letter proposes a novel method of compressed video super-resolution reconstruction based on MAP-POCS (Maximum Posterior Probability-Projection Onto Convex Set). At first assuming the high-resolution model subject...This letter proposes a novel method of compressed video super-resolution reconstruction based on MAP-POCS (Maximum Posterior Probability-Projection Onto Convex Set). At first assuming the high-resolution model subject to Poisson-Markov distribution, then constructing the projecting convex based on MAP. According to the characteristics of compressed video, two different convexes are constructed based on integrating the inter-frame and intra-frame information in the wavelet-domain. The results of the experiment demonstrate that the new method not only outperforms the traditional algorithms on the aspects of PSNR (Peak Signal-to-Noise Ratio), MSE (Mean Square Error) and reconstruction vision effect, but also has the advantages of rapid convergence and easy extension.展开更多
Unmanned aerial vehicle(UAV)-borne gamma-ray spectrum survey plays a crucial role in geological mapping,radioactive mineral exploration,and environmental monitoring.However,raw data are often compromised by flight and...Unmanned aerial vehicle(UAV)-borne gamma-ray spectrum survey plays a crucial role in geological mapping,radioactive mineral exploration,and environmental monitoring.However,raw data are often compromised by flight and instrument background noise,as well as detector resolution limitations,which affect the accuracy of geological interpretations.This study aims to explore the application of the Real-ESRGAN algorithm in the super-resolution reconstruction of UAV-borne gamma-ray spectrum images to enhance spatial resolution and the quality of geological feature visualization.We conducted super-resolution reconstruction experiments with 2×,4×and 6×magnification using the Real-ESRGAN algorithm,comparing the results with three other mainstream algorithms(SRCNN,SRGAN,FSRCNN)to verify the superiority in image quality.The experimental results indicate that Real-ESRGAN achieved a structural similarity index(SSIM)value of 0.950 at 2×magnification,significantly higher than the other algorithms,demonstrating its advantage in detail preservation.Furthermore,Real-ESRGAN effectively reduced ringing and overshoot artifacts,enhancing the clarity of geological structures and mineral deposit sites,thus providing high-quality visual information for geological exploration.展开更多
Single Image Super-Resolution(SISR)seeks to reconstruct high-resolution(HR)images from lowresolution(LR)inputs,thereby enhancing visual fidelity and the perception of fine details.While Transformer-based models—such ...Single Image Super-Resolution(SISR)seeks to reconstruct high-resolution(HR)images from lowresolution(LR)inputs,thereby enhancing visual fidelity and the perception of fine details.While Transformer-based models—such as SwinIR,Restormer,and HAT—have recently achieved impressive results in super-resolution tasks by capturing global contextual information,these methods often suffer from substantial computational and memory overhead,which limits their deployment on resource-constrained edge devices.To address these challenges,we propose a novel lightweight super-resolution network,termed Binary Attention-Guided Information Distillation(BAID),which integrates frequency-aware modeling with a binary attention mechanism to significantly reduce computational complexity and parameter count whilemaintaining strong reconstruction performance.The network combines a high–low frequency decoupling strategy with a local–global attention sharing mechanism,enabling efficient compression of redundant computations through binary attention guidance.At the core of the architecture lies the Attention-Guided Distillation Block(AGDB),which retains the strengths of the information distillation framework while introducing a sparse binary attention module to enhance both inference efficiency and feature representation.Extensive×4 superresolution experiments on four standard benchmarks—Set5,Set14,BSD100,and Urban100—demonstrate that BAID achieves Peak Signal-to-Noise Ratio(PSNR)values of 32.13,28.51,27.47,and 26.15,respectively,with only 1.22 million parameters and 26.1 G Floating-Point Operations(FLOPs),outperforming other state-of-the-art lightweight methods such as Information Multi-Distillation Network(IMDN)and Residual Feature Distillation Network(RFDN).These results highlight the proposed model’s ability to deliver high-quality image reconstruction while offering strong deployment efficiency,making it well-suited for image restoration tasks in resource-limited environments.展开更多
This study develops a surrogate super-resolution(SR)framework that accelerates finite element method(FEM)-based computational fluid dynamics(CFD)using deep learning.High-resolution(HR)FEM-based CFDremains computationa...This study develops a surrogate super-resolution(SR)framework that accelerates finite element method(FEM)-based computational fluid dynamics(CFD)using deep learning.High-resolution(HR)FEM-based CFDremains computationally prohibitive for time-sensitive applications,including patient-specific aneurysm hemodynamics where rapid turnaround is valuable.The proposed pipeline learns to reconstruct HR velocity-magnitude fields fromlow-resolution(LR)FEM solutions generated under the same governing equations and boundary conditions.It consistsof three modules:(i)offline pre-training of a residual network on representative vascular geometries;(ii)lightweightfine-tuning to adapt the pretrained model to geometric variability,including patient-specific aneurysm morphologies;and(iii)an unstructured-to-structured sampling strategy with region-of-interest upsampling that concentrates resolution in flow-critical zones(e.g.,the aneurysm sac)rather than the full domain.This targeted reconstruction substantiallyreduces inference and post-processing cost while preserving key HR flow features.Experiments on cerebral aneurysmmodels show that HR velocity-magnitude fields can be recovered with accuracy comparable to direct HR simulationsat less than 1%of the direct HR simulation cost per analysis(LR simulation and SR inference),while adaptation to newgeometries requires only lightweight fine-tuning with limited target-specific HR data.While clinical endpoints andadditional variables(e.g.,pressure or wall-based metrics)are left for future work,the results indicate that the proposedsurrogate SR approach can streamline FEM-based CFD workflows toward near real-time hemodynamic analysis acrossmorphologically similar vascular models.展开更多
High-resolution remote sensing imagery is essential for critical applications such as precision agriculture,urban management planning,and military reconnaissance.Although significant progress has been made in singleim...High-resolution remote sensing imagery is essential for critical applications such as precision agriculture,urban management planning,and military reconnaissance.Although significant progress has been made in singleimage super-resolution(SISR)using generative adversarial networks(GANs),existing approaches still face challenges in recovering high-frequency details,effectively utilizing features,maintaining structural integrity,and ensuring training stability—particularly when dealing with the complex textures characteristic of remote sensing imagery.To address these limitations,this paper proposes the Improved ResidualModule and AttentionMechanism Network(IRMANet),a novel architecture specifically designed for remote sensing image reconstruction.IRMANet builds upon the Super-Resolution Generative Adversarial Network(SRGAN)framework and introduces several key innovations.First,the Enhanced Residual Unit(ERU)enhances feature reuse and stabilizes training through deep residual connections.Second,the Self-Attention Residual Block(SARB)incorporates a self-attentionmechanism into the Improved Residual Module(IRM)to effectivelymodel long-range dependencies and automatically emphasize salient features.Additionally,the IRM adopts amulti-scale feature fusion strategy to facilitate synergistic interactions between local detail and global semantic information.The effectiveness of each component is validated through ablation studies,while comprehensive comparative experiments on standard remote sensing datasets demonstrate that IRMANet significantly outperforms both the baseline and state-of-the-art methods in terms of perceptual quality and quantitative metrics.Specifically,compared to the baseline model,at a magnification factor of 2,IRMANet achieves an improvement of 0.24 dB in peak signal-to-noise ratio(PSNR)and 0.54 in structural similarity index(SSIM);at a magnification factor of 4,it achieves gains of 0.22 dB in PSNR and 0.51 in SSIM.These results confirm that the proposedmethod effectively enhances detail representation and structural reconstruction accuracy in complex remote sensing scenarios,offering robust technical support for high-precision detection and identification of both military and civilian aircraft.展开更多
Remote sensing image super-resolution technology is pivotal for enhancing image quality in critical applications including environmental monitoring,urban planning,and disaster assessment.However,traditional methods ex...Remote sensing image super-resolution technology is pivotal for enhancing image quality in critical applications including environmental monitoring,urban planning,and disaster assessment.However,traditional methods exhibit deficiencies in detail recovery and noise suppression,particularly when processing complex landscapes(e.g.,forests,farmlands),leading to artifacts and spectral distortions that limit practical utility.To address this,we propose an enhanced Super-Resolution Generative Adversarial Network(SRGAN)framework featuring three key innovations:(1)Replacement of L1/L2 loss with a robust Charbonnier loss to suppress noise while preserving edge details via adaptive gradient balancing;(2)A multi-loss joint optimization strategy dynamically weighting Charbonnier loss(β=0.5),Visual Geometry Group(VGG)perceptual loss(α=1),and adversarial loss(γ=0.1)to synergize pixel-level accuracy and perceptual quality;(3)A multi-scale residual network(MSRN)capturing cross-scale texture features(e.g.,forest canopies,mountain contours).Validated on Sentinel-2(10 m)and SPOT-6/7(2.5 m)datasets covering 904 km2 in Motuo County,Xizang,our method outperforms the SRGAN baseline(SR4RS)with Peak Signal-to-Noise Ratio(PSNR)gains of 0.29 dB and Structural Similarity Index(SSIM)improvements of 3.08%on forest imagery.Visual comparisons confirm enhanced texture continuity despite marginal Learned Perceptual Image Patch Similarity(LPIPS)increases.The method significantly improves noise robustness and edge retention in complex geomorphology,demonstrating 18%faster response in forest fire early warning and providing high-resolution support for agricultural/urban monitoring.Future work will integrate spectral constraints and lightweight architectures.展开更多
Video Super-Resolution (SR) reconstruction produces video sequences with High Resolution (HR) via the fusion of several Low-Resolution (LR) video frames. Traditional methods rely on the accurate estimation of su...Video Super-Resolution (SR) reconstruction produces video sequences with High Resolution (HR) via the fusion of several Low-Resolution (LR) video frames. Traditional methods rely on the accurate estimation of subpixel motion, which constrains their applicability to video sequences with relatively simple motions such as global translation. We propose an efficient iterative spatio-temporal adaptive SR reconstruction model based on Zemike Moment (ZM), which is effective for spatial video sequences with arbitrary motion. The model uses region correlation judgment and self-adaptive threshold strategies to improve the effect and time efficiency of the ZM-based SR method. This leads to better mining of non-local self-similarity and local structural regularity, and is robust to noise and rotation. An efficient iterative curvature-based interpolation scheme is introduced to obtain the initial HR estimation of each LR video frame. Experimental results both on spatial and standard video sequences demonstrate that the proposed method outperforms existing methods in terms of both subjective visual and objective quantitative evaluations, and greatly improves the time efficiency.展开更多
Space-time video super-resolution(STVSR)serves the purpose to reconstruct high-resolution high-frame-rate videos from their low-resolution low-frame-rate counterparts.Recent approaches utilize end-to-end deep learning...Space-time video super-resolution(STVSR)serves the purpose to reconstruct high-resolution high-frame-rate videos from their low-resolution low-frame-rate counterparts.Recent approaches utilize end-to-end deep learning models to achieve STVSR.They first interpolate intermediate frame features between given frames,then perform local and global refinement among the feature sequence,and finally increase the spatial resolutions of these features.However,in the most important feature interpolation phase,they only capture spatial-temporal information from the most adjacent frame features,ignoring modelling long-term spatial-temporal correlations between multiple neighbouring frames to restore variable-speed object movements and maintain long-term motion continuity.In this paper,we propose a novel long-term temporal feature aggregation network(LTFA-Net)for STVSR.Specifically,we design a long-term mixture of experts(LTMoE)module for feature interpolation.LTMoE contains multiple experts to extract mutual and complementary spatial-temporal information from multiple consecutive adjacent frame features,which are then combined with different weights to obtain interpolation results using several gating nets.Next,we perform local and global feature refinement using the Locally-temporal Feature Comparison(LFC)module and bidirectional deformable ConvLSTM layer,respectively.Experimental results on two standard benchmarks,Adobe240 and GoPro,indicate the effectiveness and superiority of our approach over state of the art.展开更多
The rapid development of super-resolution microscopy has made it possible to observe subcellular structures and dynamic behaviors in living cells with nanoscale spatial resolution, greatly advancing progress in life s...The rapid development of super-resolution microscopy has made it possible to observe subcellular structures and dynamic behaviors in living cells with nanoscale spatial resolution, greatly advancing progress in life sciences. As hardware technology continues to evolve, the availability of new fluorescent probes with superior performance is becoming increasingly important. In recent years, fluorescent nanoprobes (FNPs) have emerged as highly promising fluorescent probes for bioimaging due to their high brightness and excellent photostability. This paper focuses on the development and applications of FNPs as probes for live-cell super-resolution imaging. It provides an overview of different super-resolution methods, discusses the performance requirements for FNPs in these methods, and reviews the latest applications of FNPs in the super-resolution imaging of living cells. Finally, it addresses the challenges and future outlook in this field.展开更多
Hyper Spectral Image Super-Resolution(HSI-SR) has gained significant attention in recent years due to its potential applications.However,the challenge of obtaining high-resolution hyperspectral images is compounded by...Hyper Spectral Image Super-Resolution(HSI-SR) has gained significant attention in recent years due to its potential applications.However,the challenge of obtaining high-resolution hyperspectral images is compounded by limitations in sensor resolution and the high dimensionality of spectral data.Traditional approaches,including interpolation-based methods and sparse representation techniques,often struggle to capture the intricate spectral-spatial dependencies in hyperspectral images.To address these limitations,this study proposes a Hadamard Self-Attention Network(HSAN) for fusing a High-resolution Multispectral Image(Hr-MSI) and a Low-resolution Hyper Spectral Image(Lr-HSI),achieving HSI-SR for obtaining a High-resolution Hyper Spectral Image(Hr-HSI).The core of HSAN is a new Hadamard self-attention mechanism that can be more efficient than traditional dot-product attention because it avoids matrix multiplications and softmax operations.Considering that deep learning-based data fusion typically entails a significant computational and storage burden,this new approach can be integrated with convolutional layers to form an unsupervised lightweight network,which significantly reduces dependence on computational resources.Experimental results across four datasets validate the effectiveness and advantages of HSAN,compared with state-of-the-art approaches.The source code will be available at https://github.com/zxnhkdm/HSAN.展开更多
Existing imaging techniques cannot simultaneously achieve high resolution and a wide field of view,and manual multi-mineral segmentation in shale lacks precision.To address these limitations,we propose a comprehensive...Existing imaging techniques cannot simultaneously achieve high resolution and a wide field of view,and manual multi-mineral segmentation in shale lacks precision.To address these limitations,we propose a comprehensive framework based on generative adversarial network(GAN)for characterizing pore structure properties of shale,which incorporates image augmentation,super-resolution reconstruction,and multi-mineral auto-segmentation.Using real 2D and 3D shale images,the framework was assessed through correlation function,entropy,porosity,pore size distribution,and permeability.The application results show that this framework enables the enhancement of 3D low-resolution digital cores by a scale factor of 8,without paired shale images,effectively reconstructing the unresolved fine-scale pores under a low resolution,rather than merely denoising,deblurring,and edge clarification.The trained GAN-based segmentation model effectively improves manual multi-mineral segmentation results,resulting in a strong resemblance to real samples in terms of pore size distribution and permeability.This framework significantly improves the characterization of complex shale microstructures and can be expanded to other heterogeneous porous media,such as carbonate,coal,and tight sandstone reservoirs.展开更多
Video inpainting is a technique that fills in the missing regions or gaps in a video by using its known pixels.The existing video inpainting algorithms are computationally expensive and introduce seam in the target re...Video inpainting is a technique that fills in the missing regions or gaps in a video by using its known pixels.The existing video inpainting algorithms are computationally expensive and introduce seam in the target region that arises due to variation in brightness or contrast of the patches.To overcome these drawbacks,the authors propose a novel two-stage framework.In the first step,sub-bands of wavelets of a low-resolution image are obtained using the dualtree complex wavelet transform.Criminisi algorithm and auto-regression technique are then applied to these subbands to inpaint the missing regions.The fuzzy logic-based histogram equalisation is used to further enhance the image by preserving the image brightness and improve the local contrast.In the second step,the image is enhanced using super-resolution technique.The process of down-sampling,inpainting and subsequently enhancing the video using the super-resolution technique reduces the video inpainting time.The framework is tested on video sequences by comparing and analysing the structural similarity index matrix,peak-signal-to-noise ratio,visual information fidelity in pixel domain and execution time with the state-of-the-art algorithms.The experimental analysis gives visually pleasing results for object removal and error concealment.展开更多
Super-Resolution (SR) technique means to reconstruct High-Resolution (HR) images from a sequence of Low-Resolution (LR) observations,which has been a great focus for compressed video. Based on the theory of Projection...Super-Resolution (SR) technique means to reconstruct High-Resolution (HR) images from a sequence of Low-Resolution (LR) observations,which has been a great focus for compressed video. Based on the theory of Projection Onto Convex Set (POCS),this paper constructs Quantization Constraint Set (QCS) using the quantization information extracted from the video bit stream. By combining the statistical properties of image and the Human Visual System (HVS),a novel Adaptive Quantization Constraint Set (AQCS) is proposed. Simulation results show that AQCS-based SR al-gorithm converges at a fast rate and obtains better performance in both objective and subjective quality,which is applicable for compressed video.展开更多
Existing learning-based super-resolution (SR) reconstruction algorithms are mainly designed for single image, which ignore the spatio-temporal relationship between video frames. Aiming at applying the advantages of ...Existing learning-based super-resolution (SR) reconstruction algorithms are mainly designed for single image, which ignore the spatio-temporal relationship between video frames. Aiming at applying the advantages of learning-based algorithms to video SR field, a novel video SR reconstruction algorithm based on deep convolutional neural network (CNN) and spatio-temporal similarity (STCNN-SR) was proposed in this paper. It is a deep learning method for video SR reconstruction, which considers not onlv the mapping relationship among associated low-resolution (LR) and high-resolution (HR) image blocks, but also the spatio-temporal non-local complementary and redundant information between adjacent low-resolution video frames. The reconstruction speed can be improved obviously with the pre-trained end-to-end reconstructed coefficients. Moreover, the performance of video SR will be further improved by the optimization process with spatio-temporal similarity. Experimental results demonstrated that the proposed algorithm achieves a competitive SR quality on both subjective and objective evaluations, when compared to other state-of-the-art algorithms.展开更多
Image super-resolution reconstruction technology is currently widely used in medical imaging,video surveillance,and industrial quality inspection.It not only enhances image quality but also improves details and visual...Image super-resolution reconstruction technology is currently widely used in medical imaging,video surveillance,and industrial quality inspection.It not only enhances image quality but also improves details and visual perception,significantly increasing the utility of low-resolution images.In this study,an improved image superresolution reconstruction model based on Generative Adversarial Networks(SRGAN)was proposed.This model introduced a channel and spatial attention mechanism(CSAB)in the generator,allowing it to effectively leverage the information from the input image to enhance feature representations and capture important details.The discriminator was designed with an improved PatchGAN architecture,which more accurately captured local details and texture information of the image.With these enhanced generator and discriminator architectures and an optimized loss function design,this method demonstrated superior performance in image quality assessment metrics.Experimental results showed that this model outperforms traditional methods,presenting more detailed and realistic image details in the visual effects.展开更多
Significant advancements have been achieved in the field of Single Image Super-Resolution(SISR)through the utilization of Convolutional Neural Networks(CNNs)to attain state-of-the-art performance.Recent efforts have e...Significant advancements have been achieved in the field of Single Image Super-Resolution(SISR)through the utilization of Convolutional Neural Networks(CNNs)to attain state-of-the-art performance.Recent efforts have explored the incorporation of Transformers to augment network performance in SISR.However,the high computational cost of Transformers makes them less suitable for deployment on lightweight devices.Moreover,the majority of enhancements for CNNs rely predominantly on small spatial convolutions,thereby neglecting the potential advantages of large kernel convolution.In this paper,the authors propose a Multi-Perception Large Kernel convNet(MPLKN)which delves into the exploration of large kernel convolution.Specifically,the authors have architected a Multi-Perception Large Kernel(MPLK)module aimed at extracting multi-scale features and employ a stepwise feature fusion strategy to seamlessly integrate these features.In addition,to enhance the network's capacity for nonlinear spatial information processing,the authors have designed a Spatial-Channel Gated Feed-forward Network(SCGFN)that is capable of adapting to feature interactions across both spatial and channel dimensions.Experimental results demonstrate that MPLKN outperforms other lightweight image super-resolution models while maintaining a minimal number of parameters and FLOPs.展开更多
Recent Super-Resolution(SR)algorithms often suffer from excessive model complexity,high computational costs,and limited flexibility across varying image scales.To address these challenges,we propose DDNet,a dynamic an...Recent Super-Resolution(SR)algorithms often suffer from excessive model complexity,high computational costs,and limited flexibility across varying image scales.To address these challenges,we propose DDNet,a dynamic and lightweight SR framework designed for arbitrary scaling factors.DDNet integrates a residual learning structure with an Adaptively fusion Feature Block(AFB)and a scale-aware upsampling module,effectively reducing parameter overhead while preserving reconstruction quality.Additionally,we introduce DDNetGAN,an enhanced variant that leverages a relativistic Generative Adversarial Network(GAN)to further improve texture realism.To validate the proposed models,we conduct extensive training using the DIV2K and Flickr2K datasets and evaluate performance across standard benchmarks including Set5,Set14,Urban100,Manga109,and BSD100.Our experiments cover both symmetric and asymmetric upscaling factors and incorporate ablation studies to assess key components.Results show that DDNet and DDNetGAN achieve competitive performance compared with mainstream SR algorithms,demonstrating a strong balance between accuracy,efficiency,and flexibility.These findings highlight the potential of our approach for practical real-world super-resolution applications.展开更多
Blood cells are the most integral part of the body,which are made up of erythrocytes,platelets and white blood cells.The examination of subcellular structures and proteins within blood cells at the nanoscale can provi...Blood cells are the most integral part of the body,which are made up of erythrocytes,platelets and white blood cells.The examination of subcellular structures and proteins within blood cells at the nanoscale can provide valuable insights into the health status of an individual,accurate diagnosis,and efficient treatment strategies for diseases.Super-resolution microscopy(SRM)has recently emerged as a cutting-edge tool for the study of blood cells,providing numerous advantages over traditional methods for examining subcellular structures and proteins.In this paper,we focus on outlining the fundamental principles of various SRM techniques and their applications in both normal and diseased states of blood cells.Furthermore,future prospects of SRM techniques in the analysis of blood cells are also discussed.展开更多
基金supported in part by the Basic and Applied Basic Research Foundation of Guangdong Province[2025A1515011566]in part by the State Key Laboratory for Novel Software Technology,Nanjing University[KFKT2024B08]+1 种基金in part by Leading Talents in Gusu Innovation and Entrepreneurship[ZXL2023170]in part by the Basic Research Programs of Taicang 2024,[TC2024JC32].
文摘Deep convolutional neural networks(CNNs)have demonstrated remarkable performance in video super-resolution(VSR).However,the ability of most existing methods to recover fine details in complex scenes is often hindered by the loss of shallow texture information during feature extraction.To address this limitation,we propose a 3D Convolutional Enhanced Residual Video Super-Resolution Network(3D-ERVSNet).This network employs a forward and backward bidirectional propagation module(FBBPM)that aligns features across frames using explicit optical flow through lightweight SPyNet.By incorporating an enhanced residual structure(ERS)with skip connections,shallow and deep features are effectively integrated,enhancing texture restoration capabilities.Furthermore,3D convolution module(3DCM)is applied after the backward propagation module to implicitly capture spatio-temporal dependencies.The architecture synergizes these components where FBBPM extracts aligned features,ERS fuses hierarchical representations,and 3DCM refines temporal coherence.Finally,a deep feature aggregation module(DFAM)fuses the processed features,and a pixel-upsampling module(PUM)reconstructs the high-resolution(HR)video frames.Comprehensive evaluations on REDS,Vid4,UDM10,and Vim4 benchmarks demonstrate well performance including 30.95 dB PSNR/0.8822 SSIM on REDS and 32.78 dB/0.8987 on Vim4.3D-ERVSNet achieves significant gains over baselines while maintaining high efficiency with only 6.3M parameters and 77ms/frame runtime(i.e.,20×faster than RBPN).The network’s effectiveness stems from its task-specific asymmetric design that balances explicit alignment and implicit fusion.
基金Supported by Open Project of the Ministry of Industry and Information Technology Key Laboratory of Performance and Reliability Testing and Evaluation for Basic Software and Hardware。
文摘Background Recurrent recovery is a common method for video super-resolution(VSR)that models the correlation between frames via hidden states.However,the application of this structure in real-world scenarios can lead to unsatisfactory artifacts.We found that in real-world VSR training,the use of unknown and complex degradation can better simulate the degradation process in the real world.Methods Based on this,we propose the RealFuVSR model,which simulates real-world degradation and mitigates artifacts caused by the VSR.Specifically,we propose a multiscale feature extraction module(MSF)module that extracts and fuses features from multiple scales,thereby facilitating the elimination of hidden state artifacts.To improve the accuracy of the hidden state alignment information,RealFuVSR uses an advanced optical flow-guided deformable convolution.Moreover,a cascaded residual upsampling module was used to eliminate noise caused by the upsampling process.Results The experiment demonstrates that RealFuVSR model can not only recover high-quality videos but also outperforms the state-of-the-art RealBasicVSR and RealESRGAN models.
基金Supported by the Natural Science Foundation of Jiangsu Province (No. BK2004151).
文摘This letter proposes a novel method of compressed video super-resolution reconstruction based on MAP-POCS (Maximum Posterior Probability-Projection Onto Convex Set). At first assuming the high-resolution model subject to Poisson-Markov distribution, then constructing the projecting convex based on MAP. According to the characteristics of compressed video, two different convexes are constructed based on integrating the inter-frame and intra-frame information in the wavelet-domain. The results of the experiment demonstrate that the new method not only outperforms the traditional algorithms on the aspects of PSNR (Peak Signal-to-Noise Ratio), MSE (Mean Square Error) and reconstruction vision effect, but also has the advantages of rapid convergence and easy extension.
基金supported by the National Natural Science Foundation of China(Nos.12205044 and 12265003)2024 Jiangxi Province Civil-Military Integration Research Institute‘BeiDou+’Project Subtopic(No.2024JXRH0Y06).
文摘Unmanned aerial vehicle(UAV)-borne gamma-ray spectrum survey plays a crucial role in geological mapping,radioactive mineral exploration,and environmental monitoring.However,raw data are often compromised by flight and instrument background noise,as well as detector resolution limitations,which affect the accuracy of geological interpretations.This study aims to explore the application of the Real-ESRGAN algorithm in the super-resolution reconstruction of UAV-borne gamma-ray spectrum images to enhance spatial resolution and the quality of geological feature visualization.We conducted super-resolution reconstruction experiments with 2×,4×and 6×magnification using the Real-ESRGAN algorithm,comparing the results with three other mainstream algorithms(SRCNN,SRGAN,FSRCNN)to verify the superiority in image quality.The experimental results indicate that Real-ESRGAN achieved a structural similarity index(SSIM)value of 0.950 at 2×magnification,significantly higher than the other algorithms,demonstrating its advantage in detail preservation.Furthermore,Real-ESRGAN effectively reduced ringing and overshoot artifacts,enhancing the clarity of geological structures and mineral deposit sites,thus providing high-quality visual information for geological exploration.
基金funded by Project of Sichuan Provincial Department of Science and Technology under 2025JDKP0150the Fundamental Research Funds for the Central Universities under 25CAFUC03093.
文摘Single Image Super-Resolution(SISR)seeks to reconstruct high-resolution(HR)images from lowresolution(LR)inputs,thereby enhancing visual fidelity and the perception of fine details.While Transformer-based models—such as SwinIR,Restormer,and HAT—have recently achieved impressive results in super-resolution tasks by capturing global contextual information,these methods often suffer from substantial computational and memory overhead,which limits their deployment on resource-constrained edge devices.To address these challenges,we propose a novel lightweight super-resolution network,termed Binary Attention-Guided Information Distillation(BAID),which integrates frequency-aware modeling with a binary attention mechanism to significantly reduce computational complexity and parameter count whilemaintaining strong reconstruction performance.The network combines a high–low frequency decoupling strategy with a local–global attention sharing mechanism,enabling efficient compression of redundant computations through binary attention guidance.At the core of the architecture lies the Attention-Guided Distillation Block(AGDB),which retains the strengths of the information distillation framework while introducing a sparse binary attention module to enhance both inference efficiency and feature representation.Extensive×4 superresolution experiments on four standard benchmarks—Set5,Set14,BSD100,and Urban100—demonstrate that BAID achieves Peak Signal-to-Noise Ratio(PSNR)values of 32.13,28.51,27.47,and 26.15,respectively,with only 1.22 million parameters and 26.1 G Floating-Point Operations(FLOPs),outperforming other state-of-the-art lightweight methods such as Information Multi-Distillation Network(IMDN)and Residual Feature Distillation Network(RFDN).These results highlight the proposed model’s ability to deliver high-quality image reconstruction while offering strong deployment efficiency,making it well-suited for image restoration tasks in resource-limited environments.
文摘This study develops a surrogate super-resolution(SR)framework that accelerates finite element method(FEM)-based computational fluid dynamics(CFD)using deep learning.High-resolution(HR)FEM-based CFDremains computationally prohibitive for time-sensitive applications,including patient-specific aneurysm hemodynamics where rapid turnaround is valuable.The proposed pipeline learns to reconstruct HR velocity-magnitude fields fromlow-resolution(LR)FEM solutions generated under the same governing equations and boundary conditions.It consistsof three modules:(i)offline pre-training of a residual network on representative vascular geometries;(ii)lightweightfine-tuning to adapt the pretrained model to geometric variability,including patient-specific aneurysm morphologies;and(iii)an unstructured-to-structured sampling strategy with region-of-interest upsampling that concentrates resolution in flow-critical zones(e.g.,the aneurysm sac)rather than the full domain.This targeted reconstruction substantiallyreduces inference and post-processing cost while preserving key HR flow features.Experiments on cerebral aneurysmmodels show that HR velocity-magnitude fields can be recovered with accuracy comparable to direct HR simulationsat less than 1%of the direct HR simulation cost per analysis(LR simulation and SR inference),while adaptation to newgeometries requires only lightweight fine-tuning with limited target-specific HR data.While clinical endpoints andadditional variables(e.g.,pressure or wall-based metrics)are left for future work,the results indicate that the proposedsurrogate SR approach can streamline FEM-based CFD workflows toward near real-time hemodynamic analysis acrossmorphologically similar vascular models.
基金funded by the Henan Province Key R&D Program Project,“Research and Application Demonstration of Class Ⅱ Superlattice Medium Wave High Temperature Infrared Detector Technology”,grant number 231111210400.
文摘High-resolution remote sensing imagery is essential for critical applications such as precision agriculture,urban management planning,and military reconnaissance.Although significant progress has been made in singleimage super-resolution(SISR)using generative adversarial networks(GANs),existing approaches still face challenges in recovering high-frequency details,effectively utilizing features,maintaining structural integrity,and ensuring training stability—particularly when dealing with the complex textures characteristic of remote sensing imagery.To address these limitations,this paper proposes the Improved ResidualModule and AttentionMechanism Network(IRMANet),a novel architecture specifically designed for remote sensing image reconstruction.IRMANet builds upon the Super-Resolution Generative Adversarial Network(SRGAN)framework and introduces several key innovations.First,the Enhanced Residual Unit(ERU)enhances feature reuse and stabilizes training through deep residual connections.Second,the Self-Attention Residual Block(SARB)incorporates a self-attentionmechanism into the Improved Residual Module(IRM)to effectivelymodel long-range dependencies and automatically emphasize salient features.Additionally,the IRM adopts amulti-scale feature fusion strategy to facilitate synergistic interactions between local detail and global semantic information.The effectiveness of each component is validated through ablation studies,while comprehensive comparative experiments on standard remote sensing datasets demonstrate that IRMANet significantly outperforms both the baseline and state-of-the-art methods in terms of perceptual quality and quantitative metrics.Specifically,compared to the baseline model,at a magnification factor of 2,IRMANet achieves an improvement of 0.24 dB in peak signal-to-noise ratio(PSNR)and 0.54 in structural similarity index(SSIM);at a magnification factor of 4,it achieves gains of 0.22 dB in PSNR and 0.51 in SSIM.These results confirm that the proposedmethod effectively enhances detail representation and structural reconstruction accuracy in complex remote sensing scenarios,offering robust technical support for high-precision detection and identification of both military and civilian aircraft.
基金This study was supported by:Inner Mongolia Academy of Forestry Sciences Open Research Project(Grant No.KF2024MS03)The Project to Improve the Scientific Research Capacity of the Inner Mongolia Academy of Forestry Sciences(Grant No.2024NLTS04)The Innovation and Entrepreneurship Training Program for Undergraduates of Beijing Forestry University(Grant No.X202410022268).
文摘Remote sensing image super-resolution technology is pivotal for enhancing image quality in critical applications including environmental monitoring,urban planning,and disaster assessment.However,traditional methods exhibit deficiencies in detail recovery and noise suppression,particularly when processing complex landscapes(e.g.,forests,farmlands),leading to artifacts and spectral distortions that limit practical utility.To address this,we propose an enhanced Super-Resolution Generative Adversarial Network(SRGAN)framework featuring three key innovations:(1)Replacement of L1/L2 loss with a robust Charbonnier loss to suppress noise while preserving edge details via adaptive gradient balancing;(2)A multi-loss joint optimization strategy dynamically weighting Charbonnier loss(β=0.5),Visual Geometry Group(VGG)perceptual loss(α=1),and adversarial loss(γ=0.1)to synergize pixel-level accuracy and perceptual quality;(3)A multi-scale residual network(MSRN)capturing cross-scale texture features(e.g.,forest canopies,mountain contours).Validated on Sentinel-2(10 m)and SPOT-6/7(2.5 m)datasets covering 904 km2 in Motuo County,Xizang,our method outperforms the SRGAN baseline(SR4RS)with Peak Signal-to-Noise Ratio(PSNR)gains of 0.29 dB and Structural Similarity Index(SSIM)improvements of 3.08%on forest imagery.Visual comparisons confirm enhanced texture continuity despite marginal Learned Perceptual Image Patch Similarity(LPIPS)increases.The method significantly improves noise robustness and edge retention in complex geomorphology,demonstrating 18%faster response in forest fire early warning and providing high-resolution support for agricultural/urban monitoring.Future work will integrate spectral constraints and lightweight architectures.
基金the National Basic Research Program of China (973 Program) under Grant No.2012CB821200,the National Natural Science Foundation of China under Grants No.91024001,No.61070142,the Beijing Natural Science Foundation under Grant No.4111002
文摘Video Super-Resolution (SR) reconstruction produces video sequences with High Resolution (HR) via the fusion of several Low-Resolution (LR) video frames. Traditional methods rely on the accurate estimation of subpixel motion, which constrains their applicability to video sequences with relatively simple motions such as global translation. We propose an efficient iterative spatio-temporal adaptive SR reconstruction model based on Zemike Moment (ZM), which is effective for spatial video sequences with arbitrary motion. The model uses region correlation judgment and self-adaptive threshold strategies to improve the effect and time efficiency of the ZM-based SR method. This leads to better mining of non-local self-similarity and local structural regularity, and is robust to noise and rotation. An efficient iterative curvature-based interpolation scheme is introduced to obtain the initial HR estimation of each LR video frame. Experimental results both on spatial and standard video sequences demonstrate that the proposed method outperforms existing methods in terms of both subjective visual and objective quantitative evaluations, and greatly improves the time efficiency.
文摘Space-time video super-resolution(STVSR)serves the purpose to reconstruct high-resolution high-frame-rate videos from their low-resolution low-frame-rate counterparts.Recent approaches utilize end-to-end deep learning models to achieve STVSR.They first interpolate intermediate frame features between given frames,then perform local and global refinement among the feature sequence,and finally increase the spatial resolutions of these features.However,in the most important feature interpolation phase,they only capture spatial-temporal information from the most adjacent frame features,ignoring modelling long-term spatial-temporal correlations between multiple neighbouring frames to restore variable-speed object movements and maintain long-term motion continuity.In this paper,we propose a novel long-term temporal feature aggregation network(LTFA-Net)for STVSR.Specifically,we design a long-term mixture of experts(LTMoE)module for feature interpolation.LTMoE contains multiple experts to extract mutual and complementary spatial-temporal information from multiple consecutive adjacent frame features,which are then combined with different weights to obtain interpolation results using several gating nets.Next,we perform local and global feature refinement using the Locally-temporal Feature Comparison(LFC)module and bidirectional deformable ConvLSTM layer,respectively.Experimental results on two standard benchmarks,Adobe240 and GoPro,indicate the effectiveness and superiority of our approach over state of the art.
基金supported by the following grants:National Natural Science Foundation of China(grant nos.92354305,32271428,and 32201132)National Key R&D Program of China(grant no.2022YFC3401100)+1 种基金Fund for Knowledge Innovation of Wuhan Science and Technology Bureau(grant no.2022020801010558)Director Fund of WNLO.
文摘The rapid development of super-resolution microscopy has made it possible to observe subcellular structures and dynamic behaviors in living cells with nanoscale spatial resolution, greatly advancing progress in life sciences. As hardware technology continues to evolve, the availability of new fluorescent probes with superior performance is becoming increasingly important. In recent years, fluorescent nanoprobes (FNPs) have emerged as highly promising fluorescent probes for bioimaging due to their high brightness and excellent photostability. This paper focuses on the development and applications of FNPs as probes for live-cell super-resolution imaging. It provides an overview of different super-resolution methods, discusses the performance requirements for FNPs in these methods, and reviews the latest applications of FNPs in the super-resolution imaging of living cells. Finally, it addresses the challenges and future outlook in this field.
基金National Natural Science Foundation of China(No. 42571463)Macao Young Scholars Program(No. AM2023033)+1 种基金Shaanxi Province Youth Science and Technology Star Program(No. 2024ZCKJXX-115)Natural Science Foundation of Shaanxi Province(No.2025JC-YBMS-257)。
文摘Hyper Spectral Image Super-Resolution(HSI-SR) has gained significant attention in recent years due to its potential applications.However,the challenge of obtaining high-resolution hyperspectral images is compounded by limitations in sensor resolution and the high dimensionality of spectral data.Traditional approaches,including interpolation-based methods and sparse representation techniques,often struggle to capture the intricate spectral-spatial dependencies in hyperspectral images.To address these limitations,this study proposes a Hadamard Self-Attention Network(HSAN) for fusing a High-resolution Multispectral Image(Hr-MSI) and a Low-resolution Hyper Spectral Image(Lr-HSI),achieving HSI-SR for obtaining a High-resolution Hyper Spectral Image(Hr-HSI).The core of HSAN is a new Hadamard self-attention mechanism that can be more efficient than traditional dot-product attention because it avoids matrix multiplications and softmax operations.Considering that deep learning-based data fusion typically entails a significant computational and storage burden,this new approach can be integrated with convolutional layers to form an unsupervised lightweight network,which significantly reduces dependence on computational resources.Experimental results across four datasets validate the effectiveness and advantages of HSAN,compared with state-of-the-art approaches.The source code will be available at https://github.com/zxnhkdm/HSAN.
基金Supported by the National Natural Science Foundation of China(U23A20595,52034010,52288101)National Key Research and Development Program of China(2022YFE0203400)+1 种基金Shandong Provincial Natural Science Foundation(ZR2024ZD17)Fundamental Research Funds for the Central Universities(23CX10004A).
文摘Existing imaging techniques cannot simultaneously achieve high resolution and a wide field of view,and manual multi-mineral segmentation in shale lacks precision.To address these limitations,we propose a comprehensive framework based on generative adversarial network(GAN)for characterizing pore structure properties of shale,which incorporates image augmentation,super-resolution reconstruction,and multi-mineral auto-segmentation.Using real 2D and 3D shale images,the framework was assessed through correlation function,entropy,porosity,pore size distribution,and permeability.The application results show that this framework enables the enhancement of 3D low-resolution digital cores by a scale factor of 8,without paired shale images,effectively reconstructing the unresolved fine-scale pores under a low resolution,rather than merely denoising,deblurring,and edge clarification.The trained GAN-based segmentation model effectively improves manual multi-mineral segmentation results,resulting in a strong resemblance to real samples in terms of pore size distribution and permeability.This framework significantly improves the characterization of complex shale microstructures and can be expanded to other heterogeneous porous media,such as carbonate,coal,and tight sandstone reservoirs.
文摘Video inpainting is a technique that fills in the missing regions or gaps in a video by using its known pixels.The existing video inpainting algorithms are computationally expensive and introduce seam in the target region that arises due to variation in brightness or contrast of the patches.To overcome these drawbacks,the authors propose a novel two-stage framework.In the first step,sub-bands of wavelets of a low-resolution image are obtained using the dualtree complex wavelet transform.Criminisi algorithm and auto-regression technique are then applied to these subbands to inpaint the missing regions.The fuzzy logic-based histogram equalisation is used to further enhance the image by preserving the image brightness and improve the local contrast.In the second step,the image is enhanced using super-resolution technique.The process of down-sampling,inpainting and subsequently enhancing the video using the super-resolution technique reduces the video inpainting time.The framework is tested on video sequences by comparing and analysing the structural similarity index matrix,peak-signal-to-noise ratio,visual information fidelity in pixel domain and execution time with the state-of-the-art algorithms.The experimental analysis gives visually pleasing results for object removal and error concealment.
基金the Natural Science Foundation of Jiangsu Province (No.BK2004151).
文摘Super-Resolution (SR) technique means to reconstruct High-Resolution (HR) images from a sequence of Low-Resolution (LR) observations,which has been a great focus for compressed video. Based on the theory of Projection Onto Convex Set (POCS),this paper constructs Quantization Constraint Set (QCS) using the quantization information extracted from the video bit stream. By combining the statistical properties of image and the Human Visual System (HVS),a novel Adaptive Quantization Constraint Set (AQCS) is proposed. Simulation results show that AQCS-based SR al-gorithm converges at a fast rate and obtains better performance in both objective and subjective quality,which is applicable for compressed video.
基金supported by the National Natural Science Foundation of China (61320106006, 61532006, 61502042)
文摘Existing learning-based super-resolution (SR) reconstruction algorithms are mainly designed for single image, which ignore the spatio-temporal relationship between video frames. Aiming at applying the advantages of learning-based algorithms to video SR field, a novel video SR reconstruction algorithm based on deep convolutional neural network (CNN) and spatio-temporal similarity (STCNN-SR) was proposed in this paper. It is a deep learning method for video SR reconstruction, which considers not onlv the mapping relationship among associated low-resolution (LR) and high-resolution (HR) image blocks, but also the spatio-temporal non-local complementary and redundant information between adjacent low-resolution video frames. The reconstruction speed can be improved obviously with the pre-trained end-to-end reconstructed coefficients. Moreover, the performance of video SR will be further improved by the optimization process with spatio-temporal similarity. Experimental results demonstrated that the proposed algorithm achieves a competitive SR quality on both subjective and objective evaluations, when compared to other state-of-the-art algorithms.
文摘Image super-resolution reconstruction technology is currently widely used in medical imaging,video surveillance,and industrial quality inspection.It not only enhances image quality but also improves details and visual perception,significantly increasing the utility of low-resolution images.In this study,an improved image superresolution reconstruction model based on Generative Adversarial Networks(SRGAN)was proposed.This model introduced a channel and spatial attention mechanism(CSAB)in the generator,allowing it to effectively leverage the information from the input image to enhance feature representations and capture important details.The discriminator was designed with an improved PatchGAN architecture,which more accurately captured local details and texture information of the image.With these enhanced generator and discriminator architectures and an optimized loss function design,this method demonstrated superior performance in image quality assessment metrics.Experimental results showed that this model outperforms traditional methods,presenting more detailed and realistic image details in the visual effects.
文摘Significant advancements have been achieved in the field of Single Image Super-Resolution(SISR)through the utilization of Convolutional Neural Networks(CNNs)to attain state-of-the-art performance.Recent efforts have explored the incorporation of Transformers to augment network performance in SISR.However,the high computational cost of Transformers makes them less suitable for deployment on lightweight devices.Moreover,the majority of enhancements for CNNs rely predominantly on small spatial convolutions,thereby neglecting the potential advantages of large kernel convolution.In this paper,the authors propose a Multi-Perception Large Kernel convNet(MPLKN)which delves into the exploration of large kernel convolution.Specifically,the authors have architected a Multi-Perception Large Kernel(MPLK)module aimed at extracting multi-scale features and employ a stepwise feature fusion strategy to seamlessly integrate these features.In addition,to enhance the network's capacity for nonlinear spatial information processing,the authors have designed a Spatial-Channel Gated Feed-forward Network(SCGFN)that is capable of adapting to feature interactions across both spatial and channel dimensions.Experimental results demonstrate that MPLKN outperforms other lightweight image super-resolution models while maintaining a minimal number of parameters and FLOPs.
基金supported by Sichuan Science and Technology Program[2023YFSY0026,2023YFH0004].
文摘Recent Super-Resolution(SR)algorithms often suffer from excessive model complexity,high computational costs,and limited flexibility across varying image scales.To address these challenges,we propose DDNet,a dynamic and lightweight SR framework designed for arbitrary scaling factors.DDNet integrates a residual learning structure with an Adaptively fusion Feature Block(AFB)and a scale-aware upsampling module,effectively reducing parameter overhead while preserving reconstruction quality.Additionally,we introduce DDNetGAN,an enhanced variant that leverages a relativistic Generative Adversarial Network(GAN)to further improve texture realism.To validate the proposed models,we conduct extensive training using the DIV2K and Flickr2K datasets and evaluate performance across standard benchmarks including Set5,Set14,Urban100,Manga109,and BSD100.Our experiments cover both symmetric and asymmetric upscaling factors and incorporate ablation studies to assess key components.Results show that DDNet and DDNetGAN achieve competitive performance compared with mainstream SR algorithms,demonstrating a strong balance between accuracy,efficiency,and flexibility.These findings highlight the potential of our approach for practical real-world super-resolution applications.
基金supported by the following grants:National Key R&D Program of China(Grant no.2022YFC3401100)National Natural Science Foundation of China(Grant nos.32271428,92054110,32201132 and 31600692).
文摘Blood cells are the most integral part of the body,which are made up of erythrocytes,platelets and white blood cells.The examination of subcellular structures and proteins within blood cells at the nanoscale can provide valuable insights into the health status of an individual,accurate diagnosis,and efficient treatment strategies for diseases.Super-resolution microscopy(SRM)has recently emerged as a cutting-edge tool for the study of blood cells,providing numerous advantages over traditional methods for examining subcellular structures and proteins.In this paper,we focus on outlining the fundamental principles of various SRM techniques and their applications in both normal and diseased states of blood cells.Furthermore,future prospects of SRM techniques in the analysis of blood cells are also discussed.