This systematic review aims to comprehensively examine and compare deep learning methods for brain tumor segmentation and classification using MRI and other imaging modalities,focusing on recent trends from 2022 to 20...This systematic review aims to comprehensively examine and compare deep learning methods for brain tumor segmentation and classification using MRI and other imaging modalities,focusing on recent trends from 2022 to 2025.The primary objective is to evaluate methodological advancements,model performance,dataset usage,and existing challenges in developing clinically robust AI systems.We included peer-reviewed journal articles and highimpact conference papers published between 2022 and 2025,written in English,that proposed or evaluated deep learning methods for brain tumor segmentation and/or classification.Excluded were non-open-access publications,books,and non-English articles.A structured search was conducted across Scopus,Google Scholar,Wiley,and Taylor&Francis,with the last search performed in August 2025.Risk of bias was not formally quantified but considered during full-text screening based on dataset diversity,validation methods,and availability of performance metrics.We used narrative synthesis and tabular benchmarking to compare performance metrics(e.g.,accuracy,Dice score)across model types(CNN,Transformer,Hybrid),imaging modalities,and datasets.A total of 49 studies were included(43 journal articles and 6 conference papers).These studies spanned over 9 public datasets(e.g.,BraTS,Figshare,REMBRANDT,MOLAB)and utilized a range of imaging modalities,predominantly MRI.Hybrid models,especially ResViT and UNetFormer,consistently achieved high performance,with classification accuracy exceeding 98%and segmentation Dice scores above 0.90 across multiple studies.Transformers and hybrid architectures showed increasing adoption post2023.Many studies lacked external validation and were evaluated only on a few benchmark datasets,raising concerns about generalizability and dataset bias.Few studies addressed clinical interpretability or uncertainty quantification.Despite promising results,particularly for hybrid deep learning models,widespread clinical adoption remains limited due to lack of validation,interpretability concerns,and real-world deployment barriers.展开更多
The rapid growth of biomedical data,particularly multi-omics data including genomes,transcriptomics,proteomics,metabolomics,and epigenomics,medical research and clinical decision-making confront both new opportunities...The rapid growth of biomedical data,particularly multi-omics data including genomes,transcriptomics,proteomics,metabolomics,and epigenomics,medical research and clinical decision-making confront both new opportunities and obstacles.The huge and diversified nature of these datasets cannot always be managed using traditional data analysis methods.As a consequence,deep learning has emerged as a strong tool for analysing numerous omics data due to its ability to handle complex and non-linear relationships.This paper explores the fundamental concepts of deep learning and how they are used in multi-omics medical data mining.We demonstrate how autoencoders,variational autoencoders,multimodal models,attention mechanisms,transformers,and graph neural networks enable pattern analysis and recognition across all omics data.Deep learning has been found to be effective in illness classification,biomarker identification,gene network learning,and therapeutic efficacy prediction.We also consider critical problems like as data quality,model explainability,whether findings can be repeated,and computational power requirements.We now consider future elements of combining omics with clinical and imaging data,explainable AI,federated learning,and real-time diagnostics.Overall,this study emphasises the need of collaborating across disciplines to advance deep learning-based multi-omics research for precision medicine and comprehending complicated disorders.展开更多
BACKGROUND Esophageal cancer is the sixth most common cancer worldwide,with a high mortality rate.Early prognosis of esophageal abnormalities can improve patient survival rates.The progression of esophageal cancer fol...BACKGROUND Esophageal cancer is the sixth most common cancer worldwide,with a high mortality rate.Early prognosis of esophageal abnormalities can improve patient survival rates.The progression of esophageal cancer follows a sequence from esophagitis to non-dysplastic Barrett’s esophagus,dysplastic Barrett’s esophagus,and eventually esophageal adenocarcinoma(EAC).This study explored the application of deep learning technology in the precise diagnosis of pathological classification and staging of EAC to enhance diagnostic accuracy and efficiency.AIM To explore the application of deep learning models,particularly Wave-Vision Transformer(Wave-ViT),in the pathological classification and staging of esophageal cancer to enhance diagnostic accuracy and efficiency.METHODS We applied several deep learning models,including multi-layer perceptron,residual network,transformer,and Wave-ViT,to a dataset of clinically validated esophageal pathology images.The models were trained to identify pathological features and assist in the classification and staging of different stages of esophageal cancer.The models were compared based on accuracy,computational complexity,and efficiency.RESULTS The Wave-ViT model demonstrated the highest accuracy at 88.97%,surpassing the transformer(87.65%),residual network(85.44%),and multi-layer perceptron(81.17%).Additionally,Wave-ViT exhibited low computational complexity with significantly reduced parameter size,making it highly efficient for real-time clinical applications.CONCLUSION Deep learning technology,particularly the Frequency-Domain Transformer model,shows promise in improving the precision of pathological classification and staging of EAC.The application of the Frequency-Domain Transformer model enhances the automation of the diagnostic process and may support early detection and treatment of EAC.Future research may further explore the potential of this model in broader medical image analysis applications,particularly in the field of precision medicine.展开更多
Chronic uncontrolled inflammation is a major risk factor driving the occurrence of hepatocellular carcinoma(HCC),with over half of global cases attributed to hepatitis B virus(HBV)infection.Persistent inflammation fre...Chronic uncontrolled inflammation is a major risk factor driving the occurrence of hepatocellular carcinoma(HCC),with over half of global cases attributed to hepatitis B virus(HBV)infection.Persistent inflammation frequently progresses to cirrhosis and,ultimately,malignancy[1].Monitoring the key risk factors involved in the inflammatory-to-cancerous transformation in HCC is crucial for enabling timely intervention and improving patient survival rates.To address this challenge,we analyzed plasma samples collected from healthy volunteers and patients at various stages of HCC progression.展开更多
Efficient and accurate prediction of ocean surface latent heat fluxes is essential for understanding and modeling climate dynamics.Conventional estimation methods have low resolution and lack accuracy.The transformer ...Efficient and accurate prediction of ocean surface latent heat fluxes is essential for understanding and modeling climate dynamics.Conventional estimation methods have low resolution and lack accuracy.The transformer model,with its self-attention mechanism,effectively captures long-range dependencies,leading to a degradation of accuracy over time.Due to the non-linearity and uncertainty of physical processes,the transformer model encounters the problem of error accumulation,leading to a degradation of accuracy over time.To solve this problem,we combine the Data Assimilation(DA)technique with the transformer model and continuously modify the model state to make it closer to the actual observations.In this paper,we propose a deep learning model called TransNetDA,which integrates transformer,convolutional neural network and DA methods.By combining data-driven and DA methods for spatiotemporal prediction,TransNetDA effectively extracts multi-scale spatial features and significantly improves prediction accuracy.The experimental results indicate that the TransNetDA method surpasses traditional techniques in terms of root mean square error and R2 metrics,showcasing its superior performance in predicting latent heat fluxes at the ocean surface.展开更多
Forecasting landslide deformation is challenging due to influence of various internal and external factors on the occurrence of systemic and localized heterogeneities.Despite the potential to improve landslide predict...Forecasting landslide deformation is challenging due to influence of various internal and external factors on the occurrence of systemic and localized heterogeneities.Despite the potential to improve landslide predictability,deep learning has yet to be sufficiently explored for complex deformation patterns associated with landslides and is inherently opaque.Herein,we developed a holistic landslide deformation forecasting method that considers spatiotemporal correlations of landslide deformation by integrating domain knowledge into interpretable deep learning.By spatially capturing the interconnections between multiple deformations from different observation points,our method contributes to the understanding and forecasting of landslide systematic behavior.By integrating specific domain knowledge relevant to each observation point and merging internal properties with external variables,the local heterogeneity is considered in our method,identifying deformation temporal patterns in different landslide zones.Case studies involving reservoir-induced landslides and creeping landslides demonstrated that our approach(1)enhances the accuracy of landslide deformation forecasting,(2)identifies significant contributing factors and their influence on spatiotemporal deformation characteristics,and(3)demonstrates how identifying these factors and patterns facilitates landslide forecasting.Our research offers a promising and pragmatic pathway toward a deeper understanding and forecasting of complex landslide behaviors.展开更多
The issue of small-angle maneuvering targets inverse synthetic aperture radar(ISAR)imaging has been successfully addressed by popular motion compensation algorithms.However,when the target’s rotational velocity is su...The issue of small-angle maneuvering targets inverse synthetic aperture radar(ISAR)imaging has been successfully addressed by popular motion compensation algorithms.However,when the target’s rotational velocity is sufficiently high during the dwell time of the radar,such compensation algorithms cannot obtain a high quality image.This paper proposes an ISAR imaging algorithm based on keystone transform and deep learning algorithm.The keystone transform is used to coarsely compensate for the target’s rotational motion and translational motion,and the deep learning algorithm is used to achieve a super-resolution image.The uniformly distributed point target data are used as the data set of the training u-net network.In addition,this method does not require estimating the motion parameters of the target,which simplifies the algorithm steps.Finally,several experiments are performed to demonstrate the effectiveness of the proposed algorithm.展开更多
Many traditional denoising methods,such as Gaussian filtering,tend to blur and lose details or edge information while reducing noise.The stationary wavelet packet transform is a multi-scale and multi-band analysis too...Many traditional denoising methods,such as Gaussian filtering,tend to blur and lose details or edge information while reducing noise.The stationary wavelet packet transform is a multi-scale and multi-band analysis tool.Compared with the stationary wavelet transform,it can suppress high-frequency noise while preserving more edge details.Deep learning has significantly progressed in denoising applications.DnCNN,a residual network;FFDNet,an efficient,fl exible network;U-NET,a codec network;and GAN,a generative adversative network,have better denoising effects than BM3D,the most popular conventional denoising method.Therefore,SWP_hFFDNet,a random noise attenuation network based on the stationary wavelet packet transform(SWPT)and modified FFDNet,is proposed.This network combines the advantages of SWPT,Huber norm,and FFDNet.In addition,it has three characteristics:First,SWPT is an eff ective featureextraction tool that can obtain low-and high-frequency features of different scales and frequency bands.Second,because the noise level map is the input of the network,the noise removal performance of diff erent noise levels can be improved.Third,the Huber norm can reduce the sensitivity of the network to abnormal data and enhance its robustness.The network is trained using the Adam algorithm and the BSD500 dataset,which is augmented,noised,and decomposed by SWPT.Experimental and actual data processing results show that the denoising eff ect of the proposed method is almost the same as those of BM3D,DnCNN,and FFDNet networks for low noise.However,for high noise,the proposed method is superior to the aforementioned networks.展开更多
文摘This systematic review aims to comprehensively examine and compare deep learning methods for brain tumor segmentation and classification using MRI and other imaging modalities,focusing on recent trends from 2022 to 2025.The primary objective is to evaluate methodological advancements,model performance,dataset usage,and existing challenges in developing clinically robust AI systems.We included peer-reviewed journal articles and highimpact conference papers published between 2022 and 2025,written in English,that proposed or evaluated deep learning methods for brain tumor segmentation and/or classification.Excluded were non-open-access publications,books,and non-English articles.A structured search was conducted across Scopus,Google Scholar,Wiley,and Taylor&Francis,with the last search performed in August 2025.Risk of bias was not formally quantified but considered during full-text screening based on dataset diversity,validation methods,and availability of performance metrics.We used narrative synthesis and tabular benchmarking to compare performance metrics(e.g.,accuracy,Dice score)across model types(CNN,Transformer,Hybrid),imaging modalities,and datasets.A total of 49 studies were included(43 journal articles and 6 conference papers).These studies spanned over 9 public datasets(e.g.,BraTS,Figshare,REMBRANDT,MOLAB)and utilized a range of imaging modalities,predominantly MRI.Hybrid models,especially ResViT and UNetFormer,consistently achieved high performance,with classification accuracy exceeding 98%and segmentation Dice scores above 0.90 across multiple studies.Transformers and hybrid architectures showed increasing adoption post2023.Many studies lacked external validation and were evaluated only on a few benchmark datasets,raising concerns about generalizability and dataset bias.Few studies addressed clinical interpretability or uncertainty quantification.Despite promising results,particularly for hybrid deep learning models,widespread clinical adoption remains limited due to lack of validation,interpretability concerns,and real-world deployment barriers.
文摘The rapid growth of biomedical data,particularly multi-omics data including genomes,transcriptomics,proteomics,metabolomics,and epigenomics,medical research and clinical decision-making confront both new opportunities and obstacles.The huge and diversified nature of these datasets cannot always be managed using traditional data analysis methods.As a consequence,deep learning has emerged as a strong tool for analysing numerous omics data due to its ability to handle complex and non-linear relationships.This paper explores the fundamental concepts of deep learning and how they are used in multi-omics medical data mining.We demonstrate how autoencoders,variational autoencoders,multimodal models,attention mechanisms,transformers,and graph neural networks enable pattern analysis and recognition across all omics data.Deep learning has been found to be effective in illness classification,biomarker identification,gene network learning,and therapeutic efficacy prediction.We also consider critical problems like as data quality,model explainability,whether findings can be repeated,and computational power requirements.We now consider future elements of combining omics with clinical and imaging data,explainable AI,federated learning,and real-time diagnostics.Overall,this study emphasises the need of collaborating across disciplines to advance deep learning-based multi-omics research for precision medicine and comprehending complicated disorders.
文摘BACKGROUND Esophageal cancer is the sixth most common cancer worldwide,with a high mortality rate.Early prognosis of esophageal abnormalities can improve patient survival rates.The progression of esophageal cancer follows a sequence from esophagitis to non-dysplastic Barrett’s esophagus,dysplastic Barrett’s esophagus,and eventually esophageal adenocarcinoma(EAC).This study explored the application of deep learning technology in the precise diagnosis of pathological classification and staging of EAC to enhance diagnostic accuracy and efficiency.AIM To explore the application of deep learning models,particularly Wave-Vision Transformer(Wave-ViT),in the pathological classification and staging of esophageal cancer to enhance diagnostic accuracy and efficiency.METHODS We applied several deep learning models,including multi-layer perceptron,residual network,transformer,and Wave-ViT,to a dataset of clinically validated esophageal pathology images.The models were trained to identify pathological features and assist in the classification and staging of different stages of esophageal cancer.The models were compared based on accuracy,computational complexity,and efficiency.RESULTS The Wave-ViT model demonstrated the highest accuracy at 88.97%,surpassing the transformer(87.65%),residual network(85.44%),and multi-layer perceptron(81.17%).Additionally,Wave-ViT exhibited low computational complexity with significantly reduced parameter size,making it highly efficient for real-time clinical applications.CONCLUSION Deep learning technology,particularly the Frequency-Domain Transformer model,shows promise in improving the precision of pathological classification and staging of EAC.The application of the Frequency-Domain Transformer model enhances the automation of the diagnostic process and may support early detection and treatment of EAC.Future research may further explore the potential of this model in broader medical image analysis applications,particularly in the field of precision medicine.
基金supported by Liaoning Province Science and Technology Plan Project,China(Grant No.:2022JH2/101300038)Liaoning Provincial Applied Basic Research Project,China(Grant No.:2022020255-JH2/1013)+5 种基金the National Natural Science Foundation of China of China(Grant Nos.:82104379/H3203 and 82104126/H3410)Liaoning Distinguished Professor Project(2017)Liaoning BaiQianWan Talents Program in 2019(A-37)China,Liaoning key Research and Development Program(2018),China,the Natural Science Funds of Liaoning Province,China(Grant No.:2021JH2/10300068)the Basic Scientific Research Project of Higher Education Department of Liaoning Province,China(Grant No.:LJ212410163036)Educational Commission of Liaoning Province of China,China(Grant No.:LJ212410163026).
文摘Chronic uncontrolled inflammation is a major risk factor driving the occurrence of hepatocellular carcinoma(HCC),with over half of global cases attributed to hepatitis B virus(HBV)infection.Persistent inflammation frequently progresses to cirrhosis and,ultimately,malignancy[1].Monitoring the key risk factors involved in the inflammatory-to-cancerous transformation in HCC is crucial for enabling timely intervention and improving patient survival rates.To address this challenge,we analyzed plasma samples collected from healthy volunteers and patients at various stages of HCC progression.
基金The National Natural Science Foundation of China under contract Nos 42176011 and 61931025the Fundamental Research Funds for the Central Universities of China under contract No.24CX03001A.
文摘Efficient and accurate prediction of ocean surface latent heat fluxes is essential for understanding and modeling climate dynamics.Conventional estimation methods have low resolution and lack accuracy.The transformer model,with its self-attention mechanism,effectively captures long-range dependencies,leading to a degradation of accuracy over time.Due to the non-linearity and uncertainty of physical processes,the transformer model encounters the problem of error accumulation,leading to a degradation of accuracy over time.To solve this problem,we combine the Data Assimilation(DA)technique with the transformer model and continuously modify the model state to make it closer to the actual observations.In this paper,we propose a deep learning model called TransNetDA,which integrates transformer,convolutional neural network and DA methods.By combining data-driven and DA methods for spatiotemporal prediction,TransNetDA effectively extracts multi-scale spatial features and significantly improves prediction accuracy.The experimental results indicate that the TransNetDA method surpasses traditional techniques in terms of root mean square error and R2 metrics,showcasing its superior performance in predicting latent heat fluxes at the ocean surface.
基金supported by the Postdoctoral Fellowship Program of CPSF(Grant No.GZB20230685)the National Science Foundation of China(Grant No.42277161).
文摘Forecasting landslide deformation is challenging due to influence of various internal and external factors on the occurrence of systemic and localized heterogeneities.Despite the potential to improve landslide predictability,deep learning has yet to be sufficiently explored for complex deformation patterns associated with landslides and is inherently opaque.Herein,we developed a holistic landslide deformation forecasting method that considers spatiotemporal correlations of landslide deformation by integrating domain knowledge into interpretable deep learning.By spatially capturing the interconnections between multiple deformations from different observation points,our method contributes to the understanding and forecasting of landslide systematic behavior.By integrating specific domain knowledge relevant to each observation point and merging internal properties with external variables,the local heterogeneity is considered in our method,identifying deformation temporal patterns in different landslide zones.Case studies involving reservoir-induced landslides and creeping landslides demonstrated that our approach(1)enhances the accuracy of landslide deformation forecasting,(2)identifies significant contributing factors and their influence on spatiotemporal deformation characteristics,and(3)demonstrates how identifying these factors and patterns facilitates landslide forecasting.Our research offers a promising and pragmatic pathway toward a deeper understanding and forecasting of complex landslide behaviors.
基金This work was supported by the National Natural Science Foundation of China(61571388,61871465,62071414)the Project of Introducing Overseas Students in Hebei Province(C20200367).
文摘The issue of small-angle maneuvering targets inverse synthetic aperture radar(ISAR)imaging has been successfully addressed by popular motion compensation algorithms.However,when the target’s rotational velocity is sufficiently high during the dwell time of the radar,such compensation algorithms cannot obtain a high quality image.This paper proposes an ISAR imaging algorithm based on keystone transform and deep learning algorithm.The keystone transform is used to coarsely compensate for the target’s rotational motion and translational motion,and the deep learning algorithm is used to achieve a super-resolution image.The uniformly distributed point target data are used as the data set of the training u-net network.In addition,this method does not require estimating the motion parameters of the target,which simplifies the algorithm steps.Finally,several experiments are performed to demonstrate the effectiveness of the proposed algorithm.
文摘Many traditional denoising methods,such as Gaussian filtering,tend to blur and lose details or edge information while reducing noise.The stationary wavelet packet transform is a multi-scale and multi-band analysis tool.Compared with the stationary wavelet transform,it can suppress high-frequency noise while preserving more edge details.Deep learning has significantly progressed in denoising applications.DnCNN,a residual network;FFDNet,an efficient,fl exible network;U-NET,a codec network;and GAN,a generative adversative network,have better denoising effects than BM3D,the most popular conventional denoising method.Therefore,SWP_hFFDNet,a random noise attenuation network based on the stationary wavelet packet transform(SWPT)and modified FFDNet,is proposed.This network combines the advantages of SWPT,Huber norm,and FFDNet.In addition,it has three characteristics:First,SWPT is an eff ective featureextraction tool that can obtain low-and high-frequency features of different scales and frequency bands.Second,because the noise level map is the input of the network,the noise removal performance of diff erent noise levels can be improved.Third,the Huber norm can reduce the sensitivity of the network to abnormal data and enhance its robustness.The network is trained using the Adam algorithm and the BSD500 dataset,which is augmented,noised,and decomposed by SWPT.Experimental and actual data processing results show that the denoising eff ect of the proposed method is almost the same as those of BM3D,DnCNN,and FFDNet networks for low noise.However,for high noise,the proposed method is superior to the aforementioned networks.