Gamma-ray imaging systems are powerful tools in radiographic diagnosis.However,the recorded images suffer from degradations such as noise,blurring,and downsampling,consequently failing to meet high-precision diagnosti...Gamma-ray imaging systems are powerful tools in radiographic diagnosis.However,the recorded images suffer from degradations such as noise,blurring,and downsampling,consequently failing to meet high-precision diagnostic requirements.In this paper,we propose a novel single-image super-resolution algorithm to enhance the spatial resolution of gamma-ray imaging systems.A mathematical model of the gamma-ray imaging system is established based on maximum a posteriori estimation.Within the plug-and-play framework,the half-quadratic splitting method is employed to decouple the data fidelit term and the regularization term.An image denoiser using convolutional neural networks is adopted as an implicit image prior,referred to as a deep denoiser prior,eliminating the need to explicitly design a regularization term.Furthermore,the impact of the image boundary condition on reconstruction results is considered,and a method for estimating image boundaries is introduced.The results show that the proposed algorithm can effectively addresses boundary artifacts.By increasing the pixel number of the reconstructed images,the proposed algorithm is capable of recovering more details.Notably,in both simulation and real experiments,the proposed algorithm is demonstrated to achieve subpixel resolution,surpassing the Nyquist sampling limit determined by the camera pixel size.展开更多
A new first-order optimality condition for the basis pursuit denoise (BPDN) problem is derived. This condition provides a new approach to choose the penalty param- eters adaptively for a fixed point iteration algori...A new first-order optimality condition for the basis pursuit denoise (BPDN) problem is derived. This condition provides a new approach to choose the penalty param- eters adaptively for a fixed point iteration algorithm. Meanwhile, the result is extended to matrix completion which is a new field on the heel of the compressed sensing. The numerical experiments of sparse vector recovery and low-rank matrix completion show validity of the theoretic results.展开更多
Over the years,Generative Adversarial Networks(GANs)have revolutionized the medical imaging industry for applications such as image synthesis,denoising,super resolution,data augmentation,and cross-modality translation...Over the years,Generative Adversarial Networks(GANs)have revolutionized the medical imaging industry for applications such as image synthesis,denoising,super resolution,data augmentation,and cross-modality translation.The objective of this review is to evaluate the advances,relevances,and limitations of GANs in medical imaging.An organised literature review was conducted following the guidelines of PRISMA(Preferred Reporting Items for Systematic Reviews and Meta-Analyses).The literature considered included peer-reviewed papers published between 2020 and 2025 across databases including PubMed,IEEE Xplore,and Scopus.The studies related to applications of GAN architectures in medical imaging with reported experimental outcomes and published in English in reputable journals and conferences were considered for the review.Thesis,white papers,communication letters,and non-English articles were not included for the same.CLAIM based quality assessment criteria were applied to the included studies to assess the quality.The study classifies diverse GAN architectures,summarizing their clinical applications,technical performances,and their implementation hardships.Key findings reveal the increasing applications of GANs for enhancing diagnostic accuracy,reducing data scarcity through synthetic data generation,and supporting modality translation.However,concerns such as limited generalizability,lack of clinical validation,and regulatory constraints persist.This review provides a comprehensive study of the prevailing scenario of GANs in medical imaging and highlights crucial research gaps and future directions.Though GANs hold transformative capability for medical imaging,their integration into clinical use demands further validation,interpretability,and regulatory alignment.展开更多
The initial noise present in the depth images obtained with RGB-D sensors is a combination of hardware limitations in addition to the environmental factors,due to the limited capabilities of sensors,which also produce...The initial noise present in the depth images obtained with RGB-D sensors is a combination of hardware limitations in addition to the environmental factors,due to the limited capabilities of sensors,which also produce poor computer vision results.The common image denoising techniques tend to remove significant image details and also remove noise,provided they are based on space and frequency filtering.The updated framework presented in this paper is a novel denoising model that makes use of Boruta-driven feature selection using a Long Short-Term Memory Autoencoder(LSTMAE).The Boruta algorithm identifies the most useful depth features that are used to maximize the spatial structure integrity and reduce redundancy.An LSTMAE is then used to process these selected features and model depth pixel sequences to generate robust,noise-resistant representations.The system uses the encoder to encode the input data into a latent space that has been compressed before it is decoded to retrieve the clean image.Experiments on a benchmark data set show that the suggested technique attains a PSNR of 45 dB and an SSIM of 0.90,which is 10 dB higher than the performance of conventional convolutional autoencoders and 15 times higher than that of the wavelet-based models.Moreover,the feature selection step will decrease the input dimensionality by 40%,resulting in a 37.5%reduction in training time and a real-time inference rate of 200 FPS.Boruta-LSTMAE framework,therefore,offers a highly efficient and scalable system for depth image denoising,with a high potential to be applied to close-range 3D systems,such as robotic manipulation and gesture-based interfaces.展开更多
Inverse design of advanced materials represents a pivotal challenge in materials science.Leveraging the latent space of Variational Autoencoders(VAEs)for material optimization has emerged as a significant advancement ...Inverse design of advanced materials represents a pivotal challenge in materials science.Leveraging the latent space of Variational Autoencoders(VAEs)for material optimization has emerged as a significant advancement in the field of material inverse design.However,VAEs are inherently prone to generating blurred images,posing challenges for precise inverse design and microstructure manufacturing.While increasing the dimensionality of the VAE latent space can mitigate reconstruction blurriness to some extent,it simultaneously imposes a substantial burden on target optimization due to an excessively high search space.To address these limitations,this study adopts a Variational Autoencoder guided Conditional Diffusion Generative Model(VAE-CDGM)framework integrated with Bayesian optimization to achieve the inverse design of composite materials with targeted mechanical properties.The VAE-CDGM model synergizes the strengths of VAEs and Denoising Diffusion Probabilistic Models(DDPM),enabling the generation of high-quality,sharp images while preserving a manipulable latent space.To accommodate varying dimensional requirements of the latent space,two optimization strategies are proposed.When the latent space dimensionality is excessively high,SHapley Additive exPlanations(SHAP)sensitivity analysis is employed to identify critical latent features for optimization within a reduced subspace.Conversely,direct optimization is performed in the low-dimensional latent space of VAE-CDGM when dimensionality is modest.The results demonstrate that both strategies accurately achieve the targeted design of composite materials while circumventing the blurred reconstruction flaws of VAEs,which offers a novel pathway for the precise design of advanced materials.展开更多
Crack detection accuracy in computer vision is often constrained by limited annotated datasets.Although Generative Adversarial Networks(GANs)have been applied for data augmentation,they frequently introduce blurs and ...Crack detection accuracy in computer vision is often constrained by limited annotated datasets.Although Generative Adversarial Networks(GANs)have been applied for data augmentation,they frequently introduce blurs and artifacts.To address this challenge,this study leverages Denoising Diffusion Probabilistic Models(DDPMs)to generate high-quality synthetic crack images,enriching the training set with diverse and structurally consistent samples that enhance the crack segmentation.The proposed framework involves a two-stage pipeline:first,DDPMs are used to synthesize high-fidelity crack images that capture fine structural details.Second,these generated samples are combined with real data to train segmentation networks,thereby improving accuracy and robustness in crack detection.Compared with GAN-based approaches,DDPM achieved the best fidelity,with the highest Structural Similarity Index(SSIM)(0.302)and lowest Learned Perceptual Image Patch Similarity(LPIPS)(0.461),producing artifact-free images that preserve fine crack details.To validate its effectiveness,six segmentation models were tested,among which LinkNet consistently achieved the best performance,excelling in both region-level accuracy and structural continuity.Incorporating DDPM-augmented data further enhanced segmentation outcomes,increasing F1 scores by up to 1.1%and IoU by 1.7%,while also improving boundary alignment and skeleton continuity compared with models trained on real images alone.Experiments with varying augmentation ratios showed consistent improvements,with F1 rising from 0.946(no augmentation)to 0.957 and IoU from 0.897 to 0.913 at the highest ratio.These findings demonstrate the effectiveness of diffusion-based augmentation for complex crack detection in structural health monitoring.展开更多
The cemented tailings backfill(CTB)with initial defects is more prone to destabilization damage under the influence of various unfavorable factors during the mining process.In order to investigate its influence on the...The cemented tailings backfill(CTB)with initial defects is more prone to destabilization damage under the influence of various unfavorable factors during the mining process.In order to investigate its influence on the stability of underground mining engineering,this paper simulates the generation of different degrees of initial defects inside the CTB by adding different contents of air-entraining agent(AEA),investigates the acoustic emission RA/AF eigenvalues of CTB with different contents of AEA under uniaxial compression,and adopts various denoising algorithms(e.g.,moving average smoothing,median filtering,and outlier detection)to improve the accuracy of the data.The variance and autocorrelation coefficients of RA/AF parameters were analyzed in conjunction with the critical slowing down(CSD)theory.The results show that the acoustic emission RA/AF values can be used to characterize the progressive damage evolution of CTB.The denoising algorithm processed the AE signals to reduce the effects of extraneous noise and anomalous spikes.Changes in the variance curves provide clear precursor information,while abrupt changes in the autocorrelation coefficient can be used as an auxiliary localization warning signal.The phenomenon of dramatic increase in the variance and autocorrelation coefficient curves during the compression-tightening stage,which is influenced by the initial defects,can lead to false warnings.As the initial defects of the CTB increase,its instability precursor time and instability time are prolonged,the peak stress decreases,and the time difference between the CTB and the instability damage is smaller.The results provide a new method for real-time monitoring and early warning of CTB instability damage.展开更多
The internal flow fields within a three-dimensional inward-tunning combined inlet are extremely complex,especially during the engine mode transition,where the tunnel changes may impact the flow fields significantly.To...The internal flow fields within a three-dimensional inward-tunning combined inlet are extremely complex,especially during the engine mode transition,where the tunnel changes may impact the flow fields significantly.To develop an efficient flow field reconstruction model for this,we present an Improved Conditional Denoising Diffusion Generative Adversarial Network(ICDDGAN),which integrates Conditional Denoising Diffusion Probabilistic Models(CDDPMs)with Style GAN,and introduce a reconstruction discrimination mechanism and dynamic loss weight learning strategy.We establish the Mach number flow field dataset by numerical simulation at various backpressures for the mode transition process from turbine mode to ejector ramjet mode at Mach number 2.5.The proposed ICDDGAN model,given only sparse parameter information,can rapidly generate high-quality Mach number flow fields without a large number of samples for training.The results show that ICDDGAN is superior to CDDGAN in terms of training convergence and stability.Moreover,the interpolation and extrapolation test results during backpressure conditions show that ICDDGAN can accurately and quickly reconstruct Mach number fields at various tunnel slice shapes,with a Structural Similarity Index Measure(SSIM)of over 0.96 and a Mean-Square Error(MSE)of 0.035%to actual flow fields,reducing time costs by 7-8 orders of magnitude compared to Computational Fluid Dynamics(CFD)calculations.This can provide an efficient means for rapid computation of complex flow fields.展开更多
The availability of pulmonary nodules in CT scan image of lung does not completely specify cancer.The noise in an image and morphology of nodules,like shape and size has an implicit and complex association with cancer...The availability of pulmonary nodules in CT scan image of lung does not completely specify cancer.The noise in an image and morphology of nodules,like shape and size has an implicit and complex association with cancer,and thus,a careful analysis should be mandatory on every suspected nodules and the combination of information of every nodule.In this paper,we introduce a“denoising first”two-path convolutional neural network(DFD-Net)to address this complexity.The introduced model is composed of denoising and detection part in an end to end manner.First,a residual learning denoising model(DR-Net)is employed to remove noise during the preprocessing stage.Then,a two-path convolutional neural network which takes the denoised image by DR-Net as an input to detect lung cancer is employed.The two paths focus on the joint integration of local and global features.To this end,each path employs different receptive field size which aids to model local and global dependencies.To further polish our model performance,in different way from the conventional feature concatenation approaches which directly concatenate two sets of features from different CNN layers,we introduce discriminant correlation analysis to concatenate more representative features.Finally,we also propose a retraining technique that allows us to overcome difficulties associated to the image labels imbalance.We found that this type of model easily first reduce noise in an image,balances the receptive field size effect,affords more representative features,and easily adaptable to the inconsistency among nodule shape and size.Our intensive experimental results achieved competitive results.展开更多
Existing deep learning-based point cloud denoising methods are generally trained in a supervised manner that requires clean data as ground-truth labels.However,in practice,it is not always feasible to obtain clean poi...Existing deep learning-based point cloud denoising methods are generally trained in a supervised manner that requires clean data as ground-truth labels.However,in practice,it is not always feasible to obtain clean point clouds.In this paper,we introduce a novel unsupervised point cloud denoising method that eliminates the need to use clean point clouds as groundtruth labels during training.We demonstrate that it is feasible for neural networks to only take noisy point clouds as input,and learn to approximate and restore their clean versions.In particular,we generate two noise levels for the original point clouds,requiring the second noise level to be twice the amount of the first noise level.With this,we can deduce the relationship between the displacement information that recovers the clean surfaces across the two levels of noise,and thus learn the displacement of each noisy point in order to recover the corresponding clean point.Comprehensive experiments demonstrate that our method achieves outstanding denoising results across various datasets with synthetic and real-world noise,obtaining better performance than previous unsupervised methods and competitive performance to current supervised methods.展开更多
The visual noise of each light intensity area is different when the image is drawn by Monte Carlo method.However,the existing denoising algorithms have limited denoising performance under complex lighting conditions a...The visual noise of each light intensity area is different when the image is drawn by Monte Carlo method.However,the existing denoising algorithms have limited denoising performance under complex lighting conditions and are easy to lose detailed information.So we propose a rendered image denoising method with filtering guided by lighting information.First,we design an image segmentation algorithm based on lighting information to segment the image into different illumination areas.Then,we establish the parameter prediction model guided by lighting information for filtering(PGLF)to predict the filtering parameters of different illumination areas.For different illumination areas,we use these filtering parameters to construct area filters,and the filters are guided by the lighting information to perform sub-area filtering.Finally,the filtering results are fused with auxiliary features to output denoised images for improving the overall denoising effect of the image.Under the physically based rendering tool(PBRT)scene and Tungsten dataset,the experimental results show that compared with other guided filtering denoising methods,our method improves the peak signal-to-noise ratio(PSNR)metrics by 4.2164 dB on average and the structural similarity index(SSIM)metrics by 7.8%on average.This shows that our method can better reduce the noise in complex lighting scenesand improvethe imagequality.展开更多
Air target intent recognition holds significant importance in aiding commanders to assess battlefield situations and secure a competitive edge in decision-making.Progress in this domain has been hindered by challenges...Air target intent recognition holds significant importance in aiding commanders to assess battlefield situations and secure a competitive edge in decision-making.Progress in this domain has been hindered by challenges posed by imbalanced battlefield data and the limited robustness of traditional recognition models.Inspired by the success of diffusion models in addressing visual domain sample imbalances,this paper introduces a new approach that utilizes the Markov Transfer Field(MTF)method for time series data visualization.This visualization,when combined with the Denoising Diffusion Probabilistic Model(DDPM),effectively enhances sample data and mitigates noise within the original dataset.Additionally,a transformer-based model tailored for time series visualization and air target intent recognition is developed.Comprehensive experimental results,encompassing comparative,ablation,and denoising validations,reveal that the proposed method achieves a notable 98.86%accuracy in air target intent recognition while demonstrating exceptional robustness and generalization capabilities.This approach represents a promising avenue for advancing air target intent recognition.展开更多
Accurately identifying building distribution from remote sensing images with complex background information is challenging.The emergence of diffusion models has prompted the innovative idea of employing the reverse de...Accurately identifying building distribution from remote sensing images with complex background information is challenging.The emergence of diffusion models has prompted the innovative idea of employing the reverse denoising process to distill building distribution from these complex backgrounds.Building on this concept,we propose a novel framework,building extraction diffusion model(BEDiff),which meticulously refines the extraction of building footprints from remote sensing images in a stepwise fashion.Our approach begins with the design of booster guidance,a mechanism that extracts structural and semantic features from remote sensing images to serve as priors,thereby providing targeted guidance for the diffusion process.Additionally,we introduce a cross-feature fusion module(CFM)that bridges the semantic gap between different types of features,facilitating the integration of the attributes extracted by booster guidance into the diffusion process more effectively.Our proposed BEDiff marks the first application of diffusion models to the task of building extraction.Empirical evidence from extensive experiments on the Beijing building dataset demonstrates the superior performance of BEDiff,affirming its effectiveness and potential for enhancing the accuracy of building extraction in complex urban landscapes.展开更多
Myocardial perfusion imaging(MPI),which uses single-photon emission computed tomography(SPECT),is a well-known estimating tool for medical diagnosis,employing the classification of images to show situations in coronar...Myocardial perfusion imaging(MPI),which uses single-photon emission computed tomography(SPECT),is a well-known estimating tool for medical diagnosis,employing the classification of images to show situations in coronary artery disease(CAD).The automatic classification of SPECT images for different techniques has achieved near-optimal accuracy when using convolutional neural networks(CNNs).This paper uses a SPECT classification framework with three steps:1)Image denoising,2)Attenuation correction,and 3)Image classification.Image denoising is done by a U-Net architecture that ensures effective image denoising.Attenuation correction is implemented by a convolution neural network model that can remove the attenuation that affects the feature extraction process of classification.Finally,a novel multi-scale diluted convolution(MSDC)network is proposed.It merges the features extracted in different scales and makes the model learn the features more efficiently.Three scales of filters with size 3×3 are used to extract features.All three steps are compared with state-of-the-art methods.The proposed denoising architecture ensures a high-quality image with the highest peak signal-to-noise ratio(PSNR)value of 39.7.The proposed classification method is compared with the five different CNN models,and the proposed method ensures better classification with an accuracy of 96%,precision of 87%,sensitivity of 87%,specificity of 89%,and F1-score of 87%.To demonstrate the importance of preprocessing,the classification model was analyzed without denoising and attenuation correction.展开更多
This paper discusses the principle and procedures of the second-generation wavelet transform and its application to the denoising of seismic data. Based on lifting steps, it is a flexible wavelet construction method u...This paper discusses the principle and procedures of the second-generation wavelet transform and its application to the denoising of seismic data. Based on lifting steps, it is a flexible wavelet construction method using linear and nonlinear spatial prediction and operators to implement the wavelet transform and to make it reversible. The lifting scheme transform -includes three steps: split, predict, and update. Deslauriers-Dubuc (4, 2) wavelet transforms are used to process both synthetic and real data in our second-generation wavelet transform. The processing results show that random noise is effectively suppressed and the signal to noise ratio improves remarkably. The lifting wavelet transform is an efficient algorithm.展开更多
In the field of image processing,the analysis of Synthetic Aperture Radar(SAR)images is crucial due to its broad range of applications.However,SAR images are often affected by coherent speckle noise,which significantl...In the field of image processing,the analysis of Synthetic Aperture Radar(SAR)images is crucial due to its broad range of applications.However,SAR images are often affected by coherent speckle noise,which significantly degrades image quality.Traditional denoising methods,typically based on filter techniques,often face challenges related to inefficiency and limited adaptability.To address these limitations,this study proposes a novel SAR image denoising algorithm based on an enhanced residual network architecture,with the objective of enhancing the utility of SAR imagery in complex electromagnetic environments.The proposed algorithm integrates residual network modules,which directly process the noisy input images to generate denoised outputs.This approach not only reduces computational complexity but also mitigates the difficulties associated with model training.By combining the Transformer module with the residual block,the algorithm enhances the network's ability to extract global features,offering superior feature extraction capabilities compared to CNN-based residual modules.Additionally,the algorithm employs the adaptive activation function Meta-ACON,which dynamically adjusts the activation patterns of neurons,thereby improving the network's feature extraction efficiency.The effectiveness of the proposed denoising method is empirically validated using real SAR images from the RSOD dataset.The proposed algorithm exhibits remarkable performance in terms of EPI,SSIM,and ENL,while achieving a substantial enhancement in PSNR when compared to traditional and deep learning-based algorithms.The PSNR performance is enhanced by over twofold.Moreover,the evaluation of the MSTAR SAR dataset substantiates the algorithm's robustness and applicability in SAR denoising tasks,with a PSNR of 25.2021 being attained.These findings underscore the efficacy of the proposed algorithm in mitigating speckle noise while preserving critical features in SAR imagery,thereby enhancing its quality and usability in practical scenarios.展开更多
Wavelet transforms have been successfully used in seismic data processing with their ability for local time - frequency analysis. However, identification of directionality is limited because wavelet transform coeffici...Wavelet transforms have been successfully used in seismic data processing with their ability for local time - frequency analysis. However, identification of directionality is limited because wavelet transform coefficients reveal only three spatial orientations. Whereas the ridgelet transform has a superior capability for direction detection and the ability to process signals with linearly changing characteristics. In this paper, we present the issue of low signal-to-noise ratio (SNR) seismic data processing based on the ridgelet transform. Actual seismic data with low SNR from south China has been processed using ridgelet transforms to improve the SNR and the continuity of seismic events. The results show that the ridgelet transform is better than the wavelet transform for these tasks.展开更多
To address the issues of peak overlap caused by complex matrices in agricultural product terahertz(THz)spectral signals and the dynamic,nonlinear interference induced by environmental and system noise,this study explo...To address the issues of peak overlap caused by complex matrices in agricultural product terahertz(THz)spectral signals and the dynamic,nonlinear interference induced by environmental and system noise,this study explores the feasibility of adaptive-signal-decomposition-based denoising methods to improve THz spectral quality.THz time-domain spectroscopy(THz-TDS)combined with an attenuated total reflection(ATR)accessory was used to collect THz absorbance spectra from 48 peanut samples.Taking the quantitative prediction model of peanut moisture content based on THz-ATR as an example,wavelet transform(WT),empirical mode decomposition(EMD),local mean decomposition(LMD),and its improved methods-segmented local mean decomposition(SLMD)and piecewise mirror extension local mean decomposition(PME-LMD)-were employed for spectral denoising.The applicability of different denoising methods was evaluated using a support vector regression(SVR)model.Experimental results show that the peanut moisture content prediction model constructed after PME-LMD denoising achieved the best performance,with a root mean square error(RMSE),coefficient of determination(R^(2)),and mean absolute percentage error(MAPE)of 0.010,0.912,and 0.040,respectively.Compared with traditional methods,PME-LMD significantly improved spectral quality and model prediction performance.The PME-LMD denoising strategy proposed in this study effectively suppresses non-uniform noise interference in THz spectral signals,providing an efficient and accurate preprocessing method for THz spectral analysis of agricultural products.This research provides theoretical support and technical guidance for the application of THz technology for detecting agricultural product quality.展开更多
The growing complexity of cyber threats requires innovative machine learning techniques,and image-based malware classification opens up new possibilities.Meanwhile,existing research has largely overlooked the impact o...The growing complexity of cyber threats requires innovative machine learning techniques,and image-based malware classification opens up new possibilities.Meanwhile,existing research has largely overlooked the impact of noise and obfuscation techniques commonly employed by malware authors to evade detection,and there is a critical gap in using noise simulation as a means of replicating real-world malware obfuscation techniques and adopting denoising framework to counteract these challenges.This study introduces an image denoising technique based on a U-Net combined with a GAN framework to address noise interference and obfuscation challenges in image-based malware analysis.The proposed methodology addresses existing classification limitations by introducing noise addition,which simulates obfuscated malware,and denoising strategies to restore robust image representations.To evaluate the approach,we used multiple CNN-based classifiers to assess noise resistance across architectures and datasets,measuring significant performance variation.Our denoising technique demonstrates remarkable performance improvements across two multi-class public datasets,MALIMG and BIG-15.For example,the MALIMG classification accuracy improved from 23.73%to 88.84%with denoising applied after Gaussian noise injection,demonstrating robustness.This approach contributes to improving malware detection by offering a robust framework for noise-resilient classification in noisy conditions.展开更多
基金supported by the National Natural Science Foundation of China(Grant No.12175183)。
文摘Gamma-ray imaging systems are powerful tools in radiographic diagnosis.However,the recorded images suffer from degradations such as noise,blurring,and downsampling,consequently failing to meet high-precision diagnostic requirements.In this paper,we propose a novel single-image super-resolution algorithm to enhance the spatial resolution of gamma-ray imaging systems.A mathematical model of the gamma-ray imaging system is established based on maximum a posteriori estimation.Within the plug-and-play framework,the half-quadratic splitting method is employed to decouple the data fidelit term and the regularization term.An image denoiser using convolutional neural networks is adopted as an implicit image prior,referred to as a deep denoiser prior,eliminating the need to explicitly design a regularization term.Furthermore,the impact of the image boundary condition on reconstruction results is considered,and a method for estimating image boundaries is introduced.The results show that the proposed algorithm can effectively addresses boundary artifacts.By increasing the pixel number of the reconstructed images,the proposed algorithm is capable of recovering more details.Notably,in both simulation and real experiments,the proposed algorithm is demonstrated to achieve subpixel resolution,surpassing the Nyquist sampling limit determined by the camera pixel size.
基金supported by the National Natural Science Foundation of China(No.61271014)the Specialized Research Fund for the Doctoral Program of Higher Education(No.20124301110003)the Graduated Students Innovation Fund of Hunan Province(No.CX2012B238)
文摘A new first-order optimality condition for the basis pursuit denoise (BPDN) problem is derived. This condition provides a new approach to choose the penalty param- eters adaptively for a fixed point iteration algorithm. Meanwhile, the result is extended to matrix completion which is a new field on the heel of the compressed sensing. The numerical experiments of sparse vector recovery and low-rank matrix completion show validity of the theoretic results.
基金supported by Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Research Project under grant number RGP2/540/46.
文摘Over the years,Generative Adversarial Networks(GANs)have revolutionized the medical imaging industry for applications such as image synthesis,denoising,super resolution,data augmentation,and cross-modality translation.The objective of this review is to evaluate the advances,relevances,and limitations of GANs in medical imaging.An organised literature review was conducted following the guidelines of PRISMA(Preferred Reporting Items for Systematic Reviews and Meta-Analyses).The literature considered included peer-reviewed papers published between 2020 and 2025 across databases including PubMed,IEEE Xplore,and Scopus.The studies related to applications of GAN architectures in medical imaging with reported experimental outcomes and published in English in reputable journals and conferences were considered for the review.Thesis,white papers,communication letters,and non-English articles were not included for the same.CLAIM based quality assessment criteria were applied to the included studies to assess the quality.The study classifies diverse GAN architectures,summarizing their clinical applications,technical performances,and their implementation hardships.Key findings reveal the increasing applications of GANs for enhancing diagnostic accuracy,reducing data scarcity through synthetic data generation,and supporting modality translation.However,concerns such as limited generalizability,lack of clinical validation,and regulatory constraints persist.This review provides a comprehensive study of the prevailing scenario of GANs in medical imaging and highlights crucial research gaps and future directions.Though GANs hold transformative capability for medical imaging,their integration into clinical use demands further validation,interpretability,and regulatory alignment.
文摘The initial noise present in the depth images obtained with RGB-D sensors is a combination of hardware limitations in addition to the environmental factors,due to the limited capabilities of sensors,which also produce poor computer vision results.The common image denoising techniques tend to remove significant image details and also remove noise,provided they are based on space and frequency filtering.The updated framework presented in this paper is a novel denoising model that makes use of Boruta-driven feature selection using a Long Short-Term Memory Autoencoder(LSTMAE).The Boruta algorithm identifies the most useful depth features that are used to maximize the spatial structure integrity and reduce redundancy.An LSTMAE is then used to process these selected features and model depth pixel sequences to generate robust,noise-resistant representations.The system uses the encoder to encode the input data into a latent space that has been compressed before it is decoded to retrieve the clean image.Experiments on a benchmark data set show that the suggested technique attains a PSNR of 45 dB and an SSIM of 0.90,which is 10 dB higher than the performance of conventional convolutional autoencoders and 15 times higher than that of the wavelet-based models.Moreover,the feature selection step will decrease the input dimensionality by 40%,resulting in a 37.5%reduction in training time and a real-time inference rate of 200 FPS.Boruta-LSTMAE framework,therefore,offers a highly efficient and scalable system for depth image denoising,with a high potential to be applied to close-range 3D systems,such as robotic manipulation and gesture-based interfaces.
文摘Inverse design of advanced materials represents a pivotal challenge in materials science.Leveraging the latent space of Variational Autoencoders(VAEs)for material optimization has emerged as a significant advancement in the field of material inverse design.However,VAEs are inherently prone to generating blurred images,posing challenges for precise inverse design and microstructure manufacturing.While increasing the dimensionality of the VAE latent space can mitigate reconstruction blurriness to some extent,it simultaneously imposes a substantial burden on target optimization due to an excessively high search space.To address these limitations,this study adopts a Variational Autoencoder guided Conditional Diffusion Generative Model(VAE-CDGM)framework integrated with Bayesian optimization to achieve the inverse design of composite materials with targeted mechanical properties.The VAE-CDGM model synergizes the strengths of VAEs and Denoising Diffusion Probabilistic Models(DDPM),enabling the generation of high-quality,sharp images while preserving a manipulable latent space.To accommodate varying dimensional requirements of the latent space,two optimization strategies are proposed.When the latent space dimensionality is excessively high,SHapley Additive exPlanations(SHAP)sensitivity analysis is employed to identify critical latent features for optimization within a reduced subspace.Conversely,direct optimization is performed in the low-dimensional latent space of VAE-CDGM when dimensionality is modest.The results demonstrate that both strategies accurately achieve the targeted design of composite materials while circumventing the blurred reconstruction flaws of VAEs,which offers a novel pathway for the precise design of advanced materials.
基金the National Natural Science Foundation of China(Grant No.:52508343)the Fundamental Research Funds for the Central Universities(Grant No.:B250201004).
文摘Crack detection accuracy in computer vision is often constrained by limited annotated datasets.Although Generative Adversarial Networks(GANs)have been applied for data augmentation,they frequently introduce blurs and artifacts.To address this challenge,this study leverages Denoising Diffusion Probabilistic Models(DDPMs)to generate high-quality synthetic crack images,enriching the training set with diverse and structurally consistent samples that enhance the crack segmentation.The proposed framework involves a two-stage pipeline:first,DDPMs are used to synthesize high-fidelity crack images that capture fine structural details.Second,these generated samples are combined with real data to train segmentation networks,thereby improving accuracy and robustness in crack detection.Compared with GAN-based approaches,DDPM achieved the best fidelity,with the highest Structural Similarity Index(SSIM)(0.302)and lowest Learned Perceptual Image Patch Similarity(LPIPS)(0.461),producing artifact-free images that preserve fine crack details.To validate its effectiveness,six segmentation models were tested,among which LinkNet consistently achieved the best performance,excelling in both region-level accuracy and structural continuity.Incorporating DDPM-augmented data further enhanced segmentation outcomes,increasing F1 scores by up to 1.1%and IoU by 1.7%,while also improving boundary alignment and skeleton continuity compared with models trained on real images alone.Experiments with varying augmentation ratios showed consistent improvements,with F1 rising from 0.946(no augmentation)to 0.957 and IoU from 0.897 to 0.913 at the highest ratio.These findings demonstrate the effectiveness of diffusion-based augmentation for complex crack detection in structural health monitoring.
基金Projects(52374138,51764013)supported by the National Natural Science Foundation of ChinaProject(20204BCJ22005)supported by the Training Plan for Academic and Technical Leaders of Major Disciplines of Jiangxi Province,China+1 种基金Project(2019M652277)supported by the China Postdoctoral Science FoundationProject(20192ACBL21014)supported by the Natural Science Youth Foundation Key Projects of Jiangxi Province,China。
文摘The cemented tailings backfill(CTB)with initial defects is more prone to destabilization damage under the influence of various unfavorable factors during the mining process.In order to investigate its influence on the stability of underground mining engineering,this paper simulates the generation of different degrees of initial defects inside the CTB by adding different contents of air-entraining agent(AEA),investigates the acoustic emission RA/AF eigenvalues of CTB with different contents of AEA under uniaxial compression,and adopts various denoising algorithms(e.g.,moving average smoothing,median filtering,and outlier detection)to improve the accuracy of the data.The variance and autocorrelation coefficients of RA/AF parameters were analyzed in conjunction with the critical slowing down(CSD)theory.The results show that the acoustic emission RA/AF values can be used to characterize the progressive damage evolution of CTB.The denoising algorithm processed the AE signals to reduce the effects of extraneous noise and anomalous spikes.Changes in the variance curves provide clear precursor information,while abrupt changes in the autocorrelation coefficient can be used as an auxiliary localization warning signal.The phenomenon of dramatic increase in the variance and autocorrelation coefficient curves during the compression-tightening stage,which is influenced by the initial defects,can lead to false warnings.As the initial defects of the CTB increase,its instability precursor time and instability time are prolonged,the peak stress decreases,and the time difference between the CTB and the instability damage is smaller.The results provide a new method for real-time monitoring and early warning of CTB instability damage.
文摘The internal flow fields within a three-dimensional inward-tunning combined inlet are extremely complex,especially during the engine mode transition,where the tunnel changes may impact the flow fields significantly.To develop an efficient flow field reconstruction model for this,we present an Improved Conditional Denoising Diffusion Generative Adversarial Network(ICDDGAN),which integrates Conditional Denoising Diffusion Probabilistic Models(CDDPMs)with Style GAN,and introduce a reconstruction discrimination mechanism and dynamic loss weight learning strategy.We establish the Mach number flow field dataset by numerical simulation at various backpressures for the mode transition process from turbine mode to ejector ramjet mode at Mach number 2.5.The proposed ICDDGAN model,given only sparse parameter information,can rapidly generate high-quality Mach number flow fields without a large number of samples for training.The results show that ICDDGAN is superior to CDDGAN in terms of training convergence and stability.Moreover,the interpolation and extrapolation test results during backpressure conditions show that ICDDGAN can accurately and quickly reconstruct Mach number fields at various tunnel slice shapes,with a Structural Similarity Index Measure(SSIM)of over 0.96 and a Mean-Square Error(MSE)of 0.035%to actual flow fields,reducing time costs by 7-8 orders of magnitude compared to Computational Fluid Dynamics(CFD)calculations.This can provide an efficient means for rapid computation of complex flow fields.
基金This work was partially funded by the national Key research and development program of China(2018YFC0806802 and 2018YFC0832105)and Bule Hora University of Ethiopia.
文摘The availability of pulmonary nodules in CT scan image of lung does not completely specify cancer.The noise in an image and morphology of nodules,like shape and size has an implicit and complex association with cancer,and thus,a careful analysis should be mandatory on every suspected nodules and the combination of information of every nodule.In this paper,we introduce a“denoising first”two-path convolutional neural network(DFD-Net)to address this complexity.The introduced model is composed of denoising and detection part in an end to end manner.First,a residual learning denoising model(DR-Net)is employed to remove noise during the preprocessing stage.Then,a two-path convolutional neural network which takes the denoised image by DR-Net as an input to detect lung cancer is employed.The two paths focus on the joint integration of local and global features.To this end,each path employs different receptive field size which aids to model local and global dependencies.To further polish our model performance,in different way from the conventional feature concatenation approaches which directly concatenate two sets of features from different CNN layers,we introduce discriminant correlation analysis to concatenate more representative features.Finally,we also propose a retraining technique that allows us to overcome difficulties associated to the image labels imbalance.We found that this type of model easily first reduce noise in an image,balances the receptive field size effect,affords more representative features,and easily adaptable to the inconsistency among nodule shape and size.Our intensive experimental results achieved competitive results.
文摘Existing deep learning-based point cloud denoising methods are generally trained in a supervised manner that requires clean data as ground-truth labels.However,in practice,it is not always feasible to obtain clean point clouds.In this paper,we introduce a novel unsupervised point cloud denoising method that eliminates the need to use clean point clouds as groundtruth labels during training.We demonstrate that it is feasible for neural networks to only take noisy point clouds as input,and learn to approximate and restore their clean versions.In particular,we generate two noise levels for the original point clouds,requiring the second noise level to be twice the amount of the first noise level.With this,we can deduce the relationship between the displacement information that recovers the clean surfaces across the two levels of noise,and thus learn the displacement of each noisy point in order to recover the corresponding clean point.Comprehensive experiments demonstrate that our method achieves outstanding denoising results across various datasets with synthetic and real-world noise,obtaining better performance than previous unsupervised methods and competitive performance to current supervised methods.
基金supported by the National Natural Science(No.U19A2063)the Jilin Provincial Development Program of Science and Technology (No.20230201080GX)the Jilin Province Education Department Scientific Research Project (No.JJKH20230851KJ)。
文摘The visual noise of each light intensity area is different when the image is drawn by Monte Carlo method.However,the existing denoising algorithms have limited denoising performance under complex lighting conditions and are easy to lose detailed information.So we propose a rendered image denoising method with filtering guided by lighting information.First,we design an image segmentation algorithm based on lighting information to segment the image into different illumination areas.Then,we establish the parameter prediction model guided by lighting information for filtering(PGLF)to predict the filtering parameters of different illumination areas.For different illumination areas,we use these filtering parameters to construct area filters,and the filters are guided by the lighting information to perform sub-area filtering.Finally,the filtering results are fused with auxiliary features to output denoised images for improving the overall denoising effect of the image.Under the physically based rendering tool(PBRT)scene and Tungsten dataset,the experimental results show that compared with other guided filtering denoising methods,our method improves the peak signal-to-noise ratio(PSNR)metrics by 4.2164 dB on average and the structural similarity index(SSIM)metrics by 7.8%on average.This shows that our method can better reduce the noise in complex lighting scenesand improvethe imagequality.
基金co-supported by the National Natural Science Foundation of China(Nos.61806219,61876189 and 61703426)the Young Talent Fund of University Association for Science and Technology in Shaanxi,China(Nos.20190108 and 20220106)the Innvation Talent Supporting Project of Shaanxi,China(No.2020KJXX-065)。
文摘Air target intent recognition holds significant importance in aiding commanders to assess battlefield situations and secure a competitive edge in decision-making.Progress in this domain has been hindered by challenges posed by imbalanced battlefield data and the limited robustness of traditional recognition models.Inspired by the success of diffusion models in addressing visual domain sample imbalances,this paper introduces a new approach that utilizes the Markov Transfer Field(MTF)method for time series data visualization.This visualization,when combined with the Denoising Diffusion Probabilistic Model(DDPM),effectively enhances sample data and mitigates noise within the original dataset.Additionally,a transformer-based model tailored for time series visualization and air target intent recognition is developed.Comprehensive experimental results,encompassing comparative,ablation,and denoising validations,reveal that the proposed method achieves a notable 98.86%accuracy in air target intent recognition while demonstrating exceptional robustness and generalization capabilities.This approach represents a promising avenue for advancing air target intent recognition.
基金supported by the National Natural Science Foundation of China(Nos.61906168,62202429 and 62272267)the Zhejiang Provincial Natural Science Foundation of China(No.LY23F020023)the Construction of Hubei Provincial Key Laboratory for Intelligent Visual Monitoring of Hydropower Projects(No.2022SDSJ01)。
文摘Accurately identifying building distribution from remote sensing images with complex background information is challenging.The emergence of diffusion models has prompted the innovative idea of employing the reverse denoising process to distill building distribution from these complex backgrounds.Building on this concept,we propose a novel framework,building extraction diffusion model(BEDiff),which meticulously refines the extraction of building footprints from remote sensing images in a stepwise fashion.Our approach begins with the design of booster guidance,a mechanism that extracts structural and semantic features from remote sensing images to serve as priors,thereby providing targeted guidance for the diffusion process.Additionally,we introduce a cross-feature fusion module(CFM)that bridges the semantic gap between different types of features,facilitating the integration of the attributes extracted by booster guidance into the diffusion process more effectively.Our proposed BEDiff marks the first application of diffusion models to the task of building extraction.Empirical evidence from extensive experiments on the Beijing building dataset demonstrates the superior performance of BEDiff,affirming its effectiveness and potential for enhancing the accuracy of building extraction in complex urban landscapes.
基金the Research Grant of Kwangwoon University in 2024.
文摘Myocardial perfusion imaging(MPI),which uses single-photon emission computed tomography(SPECT),is a well-known estimating tool for medical diagnosis,employing the classification of images to show situations in coronary artery disease(CAD).The automatic classification of SPECT images for different techniques has achieved near-optimal accuracy when using convolutional neural networks(CNNs).This paper uses a SPECT classification framework with three steps:1)Image denoising,2)Attenuation correction,and 3)Image classification.Image denoising is done by a U-Net architecture that ensures effective image denoising.Attenuation correction is implemented by a convolution neural network model that can remove the attenuation that affects the feature extraction process of classification.Finally,a novel multi-scale diluted convolution(MSDC)network is proposed.It merges the features extracted in different scales and makes the model learn the features more efficiently.Three scales of filters with size 3×3 are used to extract features.All three steps are compared with state-of-the-art methods.The proposed denoising architecture ensures a high-quality image with the highest peak signal-to-noise ratio(PSNR)value of 39.7.The proposed classification method is compared with the five different CNN models,and the proposed method ensures better classification with an accuracy of 96%,precision of 87%,sensitivity of 87%,specificity of 89%,and F1-score of 87%.To demonstrate the importance of preprocessing,the classification model was analyzed without denoising and attenuation correction.
文摘This paper discusses the principle and procedures of the second-generation wavelet transform and its application to the denoising of seismic data. Based on lifting steps, it is a flexible wavelet construction method using linear and nonlinear spatial prediction and operators to implement the wavelet transform and to make it reversible. The lifting scheme transform -includes three steps: split, predict, and update. Deslauriers-Dubuc (4, 2) wavelet transforms are used to process both synthetic and real data in our second-generation wavelet transform. The processing results show that random noise is effectively suppressed and the signal to noise ratio improves remarkably. The lifting wavelet transform is an efficient algorithm.
文摘In the field of image processing,the analysis of Synthetic Aperture Radar(SAR)images is crucial due to its broad range of applications.However,SAR images are often affected by coherent speckle noise,which significantly degrades image quality.Traditional denoising methods,typically based on filter techniques,often face challenges related to inefficiency and limited adaptability.To address these limitations,this study proposes a novel SAR image denoising algorithm based on an enhanced residual network architecture,with the objective of enhancing the utility of SAR imagery in complex electromagnetic environments.The proposed algorithm integrates residual network modules,which directly process the noisy input images to generate denoised outputs.This approach not only reduces computational complexity but also mitigates the difficulties associated with model training.By combining the Transformer module with the residual block,the algorithm enhances the network's ability to extract global features,offering superior feature extraction capabilities compared to CNN-based residual modules.Additionally,the algorithm employs the adaptive activation function Meta-ACON,which dynamically adjusts the activation patterns of neurons,thereby improving the network's feature extraction efficiency.The effectiveness of the proposed denoising method is empirically validated using real SAR images from the RSOD dataset.The proposed algorithm exhibits remarkable performance in terms of EPI,SSIM,and ENL,while achieving a substantial enhancement in PSNR when compared to traditional and deep learning-based algorithms.The PSNR performance is enhanced by over twofold.Moreover,the evaluation of the MSTAR SAR dataset substantiates the algorithm's robustness and applicability in SAR denoising tasks,with a PSNR of 25.2021 being attained.These findings underscore the efficacy of the proposed algorithm in mitigating speckle noise while preserving critical features in SAR imagery,thereby enhancing its quality and usability in practical scenarios.
基金This paper is supported by China Petrochemical Key Project in the"11th Five-Year"Plan Technology and the Doctorate Fund of Ministry of Education of China (No.20050491504)
文摘Wavelet transforms have been successfully used in seismic data processing with their ability for local time - frequency analysis. However, identification of directionality is limited because wavelet transform coefficients reveal only three spatial orientations. Whereas the ridgelet transform has a superior capability for direction detection and the ability to process signals with linearly changing characteristics. In this paper, we present the issue of low signal-to-noise ratio (SNR) seismic data processing based on the ridgelet transform. Actual seismic data with low SNR from south China has been processed using ridgelet transforms to improve the SNR and the continuity of seismic events. The results show that the ridgelet transform is better than the wavelet transform for these tasks.
基金Supported by the National Key R&D Program of China(2023YFD2101001)National Natural Science Foundation of China(32202144,61807001)。
文摘To address the issues of peak overlap caused by complex matrices in agricultural product terahertz(THz)spectral signals and the dynamic,nonlinear interference induced by environmental and system noise,this study explores the feasibility of adaptive-signal-decomposition-based denoising methods to improve THz spectral quality.THz time-domain spectroscopy(THz-TDS)combined with an attenuated total reflection(ATR)accessory was used to collect THz absorbance spectra from 48 peanut samples.Taking the quantitative prediction model of peanut moisture content based on THz-ATR as an example,wavelet transform(WT),empirical mode decomposition(EMD),local mean decomposition(LMD),and its improved methods-segmented local mean decomposition(SLMD)and piecewise mirror extension local mean decomposition(PME-LMD)-were employed for spectral denoising.The applicability of different denoising methods was evaluated using a support vector regression(SVR)model.Experimental results show that the peanut moisture content prediction model constructed after PME-LMD denoising achieved the best performance,with a root mean square error(RMSE),coefficient of determination(R^(2)),and mean absolute percentage error(MAPE)of 0.010,0.912,and 0.040,respectively.Compared with traditional methods,PME-LMD significantly improved spectral quality and model prediction performance.The PME-LMD denoising strategy proposed in this study effectively suppresses non-uniform noise interference in THz spectral signals,providing an efficient and accurate preprocessing method for THz spectral analysis of agricultural products.This research provides theoretical support and technical guidance for the application of THz technology for detecting agricultural product quality.
文摘The growing complexity of cyber threats requires innovative machine learning techniques,and image-based malware classification opens up new possibilities.Meanwhile,existing research has largely overlooked the impact of noise and obfuscation techniques commonly employed by malware authors to evade detection,and there is a critical gap in using noise simulation as a means of replicating real-world malware obfuscation techniques and adopting denoising framework to counteract these challenges.This study introduces an image denoising technique based on a U-Net combined with a GAN framework to address noise interference and obfuscation challenges in image-based malware analysis.The proposed methodology addresses existing classification limitations by introducing noise addition,which simulates obfuscated malware,and denoising strategies to restore robust image representations.To evaluate the approach,we used multiple CNN-based classifiers to assess noise resistance across architectures and datasets,measuring significant performance variation.Our denoising technique demonstrates remarkable performance improvements across two multi-class public datasets,MALIMG and BIG-15.For example,the MALIMG classification accuracy improved from 23.73%to 88.84%with denoising applied after Gaussian noise injection,demonstrating robustness.This approach contributes to improving malware detection by offering a robust framework for noise-resilient classification in noisy conditions.