Accurate prediction of the remaining useful life(RUL)is crucial for the design and management of lithium-ion batteries.Although various machine learning models offer promising predictions,one critical but often overlo...Accurate prediction of the remaining useful life(RUL)is crucial for the design and management of lithium-ion batteries.Although various machine learning models offer promising predictions,one critical but often overlooked challenge is their demand for considerable run-to-failure data for training.Collection of such training data leads to prohibitive testing efforts as the run-to-failure tests can last for years.Here,we propose a semi-supervised representation learning method to enhance prediction accuracy by learning from data without RUL labels.Our approach builds on a sophisticated deep neural network that comprises an encoder and three decoder heads to extract time-dependent representation features from short-term battery operating data regardless of the existence of RUL labels.The approach is validated using three datasets collected from 34 batteries operating under various conditions,encompassing over 19,900 charge and discharge cycles.Our method achieves a root mean squared error(RMSE)within 25 cycles,even when only 1/50 of the training dataset is labelled,representing a reduction of 48%compared to the conventional approach.We also demonstrate the method's robustness with varying numbers of labelled data and different weights assigned to the three decoder heads.The projection of extracted features in low space reveals that our method effectively learns degradation features from unlabelled data.Our approach highlights the promise of utilising semi-supervised learning to reduce the data demand for reliability monitoring of energy devices.展开更多
With the rapid urbanization and exponential population growth in China,two-wheeled vehicles have become a popular mode of transportation,particularly for short-distance travel.However,due to a lack of safety awareness...With the rapid urbanization and exponential population growth in China,two-wheeled vehicles have become a popular mode of transportation,particularly for short-distance travel.However,due to a lack of safety awareness,traffic violations by two-wheeled vehicle riders have become a widespread concern,contributing to urban traffic risks.Currently,significant human and material resources are being allocated to monitor and intercept non-compliant riders to ensure safe driving behavior.To enhance the safety,efficiency,and cost-effectiveness of traffic monitoring,automated detection systems based on image processing algorithms can be employed to identify traffic violations from eye-level video footage.In this study,we propose a robust detection algorithm specifically designed for two-wheeled vehicles,which serves as a fundamental step toward intelligent traffic monitoring.Our approach integrates a novel convolutional and attention mechanism to improve detection accuracy and efficiency.Additionally,we introduce a semi-supervised training strategy that leverages a large number of unlabeled images to enhance the model’s learning capability by extracting valuable background information.This method enables the model to generalize effectively to diverse urban environments and varying lighting conditions.We evaluate our proposed algorithm on a custom-built dataset,and experimental results demonstrate its superior performance,achieving an average precision(AP)of 95%and a recall(R)of 90.6%.Furthermore,the model maintains a computational efficiency of only 25.7 GFLOPs while achieving a high processing speed of 249 FPS,making it highly suitable for deployment on edge devices.Compared to existing detection methods,our approach significantly enhances the accuracy and robustness of two-wheeled vehicle identification while ensuring real-time performance.展开更多
Early identification and treatment of stroke can greatly improve patient outcomes and quality of life.Although clinical tests such as the Cincinnati Pre-hospital Stroke Scale(CPSS)and the Face Arm Speech Test(FAST)are...Early identification and treatment of stroke can greatly improve patient outcomes and quality of life.Although clinical tests such as the Cincinnati Pre-hospital Stroke Scale(CPSS)and the Face Arm Speech Test(FAST)are commonly used for stroke screening,accurate administration is dependent on specialized training.In this study,we proposed a novel multimodal deep learning approach,based on the FAST,for assessing suspected stroke patients exhibiting symptoms such as limb weakness,facial paresis,and speech disorders in acute settings.We collected a dataset comprising videos and audio recordings of emergency room patients performing designated limb movements,facial expressions,and speech tests based on the FAST.We compared the constructed deep learning model,which was designed to process multi-modal datasets,with six prior models that achieved good action classification performance,including the I3D,SlowFast,X3D,TPN,TimeSformer,and MViT.We found that the findings of our deep learning model had a higher clinical value compared with the other approaches.Moreover,the multi-modal model outperformed its single-module variants,highlighting the benefit of utilizing multiple types of patient data,such as action videos and speech audio.These results indicate that a multi-modal deep learning model combined with the FAST could greatly improve the accuracy and sensitivity of early stroke identification of stroke,thus providing a practical and powerful tool for assessing stroke patients in an emergency clinical setting.展开更多
Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay di...Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay differential equations(DDEs)and delay integrodifferential equations(DIDEs)with constant delays,primarily due to their low regularity at delayinduced breaking points.In this paper,a DNN method that combines multi-task learning(MTL)which is proposed to solve both the forward and inverse problems of DIDEs.The core idea of this approach is to divide the original equation into multiple tasks based on the delay,using auxiliary outputs to represent the integral terms,followed by the use of MTL to seamlessly incorporate the properties at the breaking points into the loss function.Furthermore,given the increased training dificulty associated with multiple tasks and outputs,we employ a sequential training scheme to reduce training complexity and provide reference solutions for subsequent tasks.This approach significantly enhances the approximation accuracy of solving DIDEs with DNNs,as demonstrated by comparisons with traditional DNN methods.We validate the effectiveness of this method through several numerical experiments,test various parameter sharing structures in MTL and compare the testing results of these structures.Finally,this method is implemented to solve the inverse problem of nonlinear DIDE and the results show that the unknown parameters of DIDE can be discovered with sparse or noisy data.展开更多
The current deep learning models for braced excavation cannot predict deformation from the beginning of excavation due to the need for a substantial corpus of sufficient historical data for training purposes.To addres...The current deep learning models for braced excavation cannot predict deformation from the beginning of excavation due to the need for a substantial corpus of sufficient historical data for training purposes.To address this issue,this study proposes a transfer learning model based on a sequence-to-sequence twodimensional(2D)convolutional long short-term memory neural network(S2SCL2D).The model can use the existing data from other adjacent similar excavations to achieve wall deflection prediction once a limited amount of monitoring data from the target excavation has been recorded.In the absence of adjacent excavation data,numerical simulation data from the target project can be employed instead.A weight update strategy is proposed to improve the prediction accuracy by integrating the stochastic gradient masking with an early stopping mechanism.To illustrate the proposed methodology,an excavation project in Hangzhou,China is adopted.The proposed deep transfer learning model,which uses either adjacent excavation data or numerical simulation data as the source domain,shows a significant improvement in performance when compared to the non-transfer learning model.Using the simulation data from the target project even leads to better prediction performance than using the actual monitoring data from other adjacent excavations.The results demonstrate that the proposed model can reasonably predict the deformation with limited data from the target project.展开更多
Bearing is an indispensable key component in mechanical equipment,and its working state is directly related to the stability and safety of the whole equipment.In recent years,with the rapid development of artificial i...Bearing is an indispensable key component in mechanical equipment,and its working state is directly related to the stability and safety of the whole equipment.In recent years,with the rapid development of artificial intelligence technology,especially the breakthrough of deep learning technology,it provides a new idea for bearing fault diagnosis.Deep learning can automatically learn features from a large amount of data,has a strong nonlinear modeling ability,and can effectively solve the problems existing in traditional methods.Aiming at the key problems in bearing fault diagnosis,this paper studies the fault diagnosis method based on deep learning,which not only provides a new solution for bearing fault diagnosis but also provides a reference for the application of deep learning in other mechanical fault diagnosis fields.展开更多
In computer vision and artificial intelligence,automatic facial expression-based emotion identification of humans has become a popular research and industry problem.Recent demonstrations and applications in several fi...In computer vision and artificial intelligence,automatic facial expression-based emotion identification of humans has become a popular research and industry problem.Recent demonstrations and applications in several fields,including computer games,smart homes,expression analysis,gesture recognition,surveillance films,depression therapy,patientmonitoring,anxiety,and others,have brought attention to its significant academic and commercial importance.This study emphasizes research that has only employed facial images for face expression recognition(FER),because facial expressions are a basic way that people communicate meaning to each other.The immense achievement of deep learning has resulted in a growing use of its much architecture to enhance efficiency.This review is on machine learning,deep learning,and hybrid methods’use of preprocessing,augmentation techniques,and feature extraction for temporal properties of successive frames of data.The following section gives a brief summary of assessment criteria that are accessible to the public and then compares them with benchmark results the most trustworthy way to assess FER-related research topics statistically.In this review,a brief synopsis of the subject matter may be beneficial for novices in the field of FER as well as seasoned scholars seeking fruitful avenues for further investigation.The information conveys fundamental knowledge and provides a comprehensive understanding of the most recent state-of-the-art research.展开更多
BACKGROUND Hepatocellular carcinoma(HCC)is a prevalent and life-threatening cancer with increasing incidence worldwide.High Ki-67 risk stratification is closely associated with higher recurrence rates and worse outcom...BACKGROUND Hepatocellular carcinoma(HCC)is a prevalent and life-threatening cancer with increasing incidence worldwide.High Ki-67 risk stratification is closely associated with higher recurrence rates and worse outcomes following curative therapies in patients with HCC.However,the performance of radiomic and deep transfer learning(DTL)models derived from biparametric magnetic resonance imaging(bpMRI)in predicting Ki-67 risk stratification and recurrence-free survival(RFS)in patients with HCC remains limited.AIM To develop a nomogram model integrating bpMRI-based radiomic and DTL signatures for predicting Ki-67 risk stratification and RFS in patients with HCC.METHODS This study included 198 patients with histopathologically confirmed HCC who underwent preoperative bpMRI.Ki-67 risk stratification was categorized as high(>20%)or low(≤20%)according to immunohistochemical staining.Radiomic and DTL signatures were extracted from the T2-weighted and arterial-phase images and combined through a random forest algorithm to establish radiomic and DTL models,respectively.Multivariate regression analysis identified clinical risk factors for high Ki-67 risk stratification,and a predictive nomogram model was developed.RESULTS A nonsmooth margin and the absence of an enhanced capsule were independent factors for high Ki-67 risk stratification.The area under the curve(AUC)of the clinical model was 0.77,while those of the radiomic and DTL models were 0.81 and 0.87,respectively,for the prediction of high Ki-67 risk stratification,and the nomogram model achieved a better AUC of 0.92.The median RFS times for patients with high and low Ki-67 risk stratification were 33.00 months and 66.73 months,respectively(P<0.001).Additionally,patients who were predicted to have high Ki-67 risk stratification by the nomogram model had a lower median RFS than those who were predicted to have low Ki-67 risk stratification(33.53 vs 66.74 months,P=0.007).CONCLUSION Our developed nomogram model demonstrated good performance in predicting Ki-67 risk stratification and predicting survival outcomes in patients with HCC.展开更多
Negative logarithm of the acid dissociation constant(pK_(a))significantly influences the absorption,dis-tribution,metabolism,excretion,and toxicity(ADMET)properties of molecules and is a crucial indicator in drug rese...Negative logarithm of the acid dissociation constant(pK_(a))significantly influences the absorption,dis-tribution,metabolism,excretion,and toxicity(ADMET)properties of molecules and is a crucial indicator in drug research.Given the rapid and accurate characteristics of computational methods,their role in predicting drug properties is increasingly important.Although many pK_(a) prediction models currently exist,they often focus on enhancing model precision while neglecting interpretability.In this study,we present GraFpKa,a pK_(a) prediction model using graph neural networks(GNNs)and molecular finger-prints.The results show that our acidic and basic models achieved mean absolute errors(MAEs)of 0.621 and 0.402,respectively,on the test set,demonstrating good predictive performance.Notably,to improve interpretability,GraFpKa also incorporates Integrated Gradients(IGs),providing a clearer visual description of the atoms significantly affecting the pK_(a) values.The high reliability and interpretability of GraFpKa ensure accurate pKa predictions while also facilitating a deeper understanding of the relation-ship between molecular structure and pK_(a) values,making it a valuable tool in the field of pK_(a) prediction.展开更多
Topographic maps,as essential tools and sources of information for geographic research,contain precise spatial locations and rich map features,and they illustrate spatio-temporal information on the distribution and di...Topographic maps,as essential tools and sources of information for geographic research,contain precise spatial locations and rich map features,and they illustrate spatio-temporal information on the distribution and differences of various surface features.Currently,topographic maps are mainly stored in raster and vector formats.Extraction of the spatio-temporal knowledge in the maps—such as spatial distribution patterns,feature relationships,and dynamic evolution—still primarily relies on manual interpretation.However,manual interpretation is time-consuming and laborious,especially for large-scale,long-term map knowledge extraction and application.With the development of artificial intelligence technology,it is possible to improve the automation level of map knowledge interpretation.Therefore,the present study proposes an automatic interpretation method for raster topographic map knowledge based on deep learning.To address the limitations of current data-driven intelligent technology in learning map spatial relations and cognitive logic,we establish a formal description of map knowledge by mapping the relationship between map knowledge and features,thereby ensuring interpretation accuracy.Subsequently,deep learning techniques are employed to extract map features automatically,and the spatio-temporal knowledge is constructed by combining formal descriptions of geographic feature knowledge.Validation experiments demonstrate that the proposed method effectively achieves automatic interpretation of spatio-temporal knowledge of geographic features in maps,with an accuracy exceeding 80%.The findings of the present study contribute to machine understanding of spatio-temporal differences in map knowledge and advances the intelligent interpretation and utilization of cartographic information.展开更多
Multimodal deep learning has emerged as a key paradigm in contemporary medical diagnostics,advancing precision medicine by enabling integration and learning from diverse data sources.The exponential growth of high-dim...Multimodal deep learning has emerged as a key paradigm in contemporary medical diagnostics,advancing precision medicine by enabling integration and learning from diverse data sources.The exponential growth of high-dimensional healthcare data,encompassing genomic,transcriptomic,and other omics profiles,as well as radiological imaging and histopathological slides,makes this approach increasingly important because,when examined separately,these data sources only offer a fragmented picture of intricate disease processes.Multimodal deep learning leverages the complementary properties of multiple data modalities to enable more accurate prognostic modeling,more robust disease characterization,and improved treatment decision-making.This review provides a comprehensive overview of the current state of multimodal deep learning approaches in medical diagnosis.We classify and examine important application domains,such as(1)radiology,where automated report generation and lesion detection are facilitated by image-text integration;(2)histopathology,where fusion models improve tumor classification and grading;and(3)multi-omics,where molecular subtypes and latent biomarkers are revealed through cross-modal learning.We provide an overview of representative research,methodological advancements,and clinical consequences for each domain.Additionally,we critically analyzed the fundamental issues preventing wider adoption,including computational complexity(particularly in training scalable,multi-branch networks),data heterogeneity(resulting from modality-specific noise,resolution variations,and inconsistent annotations),and the challenge of maintaining significant cross-modal correlations during fusion.These problems impede interpretability,which is crucial for clinical trust and use,in addition to performance and generalizability.Lastly,we outline important areas for future research,including the development of standardized protocols for harmonizing data,the creation of lightweight and interpretable fusion architectures,the integration of real-time clinical decision support systems,and the promotion of cooperation for federated multimodal learning.Our goal is to provide researchers and clinicians with a concise overview of the field’s present state,enduring constraints,and exciting directions for further research through this review.展开更多
Liver cancer remains a leading cause of mortality worldwide,and precise diagnostic tools are essential for effective treatment planning.Liver Tumors(LTs)vary significantly in size,shape,and location,and can present wi...Liver cancer remains a leading cause of mortality worldwide,and precise diagnostic tools are essential for effective treatment planning.Liver Tumors(LTs)vary significantly in size,shape,and location,and can present with tissues of similar intensities,making automatically segmenting and classifying LTs from abdominal tomography images crucial and challenging.This review examines recent advancements in Liver Segmentation(LS)and Tumor Segmentation(TS)algorithms,highlighting their strengths and limitations regarding precision,automation,and resilience.Performance metrics are utilized to assess key detection algorithms and analytical methods,emphasizing their effectiveness and relevance in clinical contexts.The review also addresses ongoing challenges in liver tumor segmentation and identification,such as managing high variability in patient data and ensuring robustness across different imaging conditions.It suggests directions for future research,with insights into technological advancements that can enhance surgical planning and diagnostic accuracy by comparing popular methods.This paper contributes to a comprehensive understanding of current liver tumor detection techniques,provides a roadmap for future innovations,and improves diagnostic and therapeutic outcomes for liver cancer by integrating recent progress with remaining challenges.展开更多
App reviews are crucial in influencing user decisions and providing essential feedback for developers to improve their products.Automating the analysis of these reviews is vital for efficient review management.While t...App reviews are crucial in influencing user decisions and providing essential feedback for developers to improve their products.Automating the analysis of these reviews is vital for efficient review management.While traditional machine learning(ML)models rely on basic word-based feature extraction,deep learning(DL)methods,enhanced with advanced word embeddings,have shown superior performance.This research introduces a novel aspectbased sentiment analysis(ABSA)framework to classify app reviews based on key non-functional requirements,focusing on usability factors:effectiveness,efficiency,and satisfaction.We propose a hybrid DL model,combining BERT(Bidirectional Encoder Representations from Transformers)with BiLSTM(Bidirectional Long Short-Term Memory)and CNN(Convolutional Neural Networks)layers,to enhance classification accuracy.Comparative analysis against state-of-the-art models demonstrates that our BERT-BiLSTM-CNN model achieves exceptional performance,with precision,recall,F1-score,and accuracy of 96%,87%,91%,and 94%,respectively.Thesignificant contributions of this work include a refined ABSA-based relabeling framework,the development of a highperformance classifier,and the comprehensive relabeling of the Instagram App Reviews dataset.These advancements provide valuable insights for software developers to enhance usability and drive user-centric application development.展开更多
As more and more devices in Cyber-Physical Systems(CPS)are connected to the Internet,physical components such as programmable logic controller(PLC),sensors,and actuators are facing greater risks of network attacks,and...As more and more devices in Cyber-Physical Systems(CPS)are connected to the Internet,physical components such as programmable logic controller(PLC),sensors,and actuators are facing greater risks of network attacks,and fast and accurate attack detection techniques are crucial.The key problem in distinguishing between normal and abnormal sequences is to model sequential changes in a large and diverse field of time series.To address this issue,we propose an anomaly detection method based on distributed deep learning.Our method uses a bilateral filtering algorithm for sequential sequences to remove noise in the time series,which can maintain the edge of discrete features.We use a distributed linear deep learning model to establish a sequential prediction model and adjust the threshold for anomaly detection based on the prediction error of the validation set.Our method can not only detect abnormal attacks but also locate the sensors that cause anomalies.We conducted experiments on the Secure Water Treatment(SWAT)and Water Distribution(WADI)public datasets.The experimental results show that our method is superior to the baseline method in identifying the types of attacks and detecting efficiency.展开更多
Exo-atmospheric vehicles are constrained by limited maneuverability,which leads to the contradiction between evasive maneuver and precision strike.To address the problem of Integrated Evasion and Impact(IEI)decision u...Exo-atmospheric vehicles are constrained by limited maneuverability,which leads to the contradiction between evasive maneuver and precision strike.To address the problem of Integrated Evasion and Impact(IEI)decision under multi-constraint conditions,a hierarchical intelligent decision-making method based on Deep Reinforcement Learning(DRL)was proposed.First,an intelligent decision-making framework of“DRL evasion decision”+“impact prediction guidance decision”was established:it takes the impact point deviation correction ability as the constraint and the maximum miss distance as the objective,and effectively solves the problem of poor decisionmaking effect caused by the large IEI decision space.Second,to solve the sparse reward problem faced by evasion decision-making,a hierarchical decision-making method consisting of maneuver timing decision and maneuver duration decision was proposed,and the corresponding Markov Decision Process(MDP)was designed.A detailed simulation experiment was designed to analyze the advantages and computational complexity of the proposed method.Simulation results show that the proposed model has good performance and low computational resource requirement.The minimum miss distance is 21.3 m under the condition of guaranteeing the impact point accuracy,and the single decision-making time is 4.086 ms on an STM32F407 single-chip microcomputer,which has engineering application value.展开更多
Lung cancer continues to be a leading cause of cancer-related deaths worldwide,emphasizing the critical need for improved diagnostic techniques.Early detection of lung tumors significantly increases the chances of suc...Lung cancer continues to be a leading cause of cancer-related deaths worldwide,emphasizing the critical need for improved diagnostic techniques.Early detection of lung tumors significantly increases the chances of successful treatment and survival.However,current diagnostic methods often fail to detect tumors at an early stage or to accurately pinpoint their location within the lung tissue.Single-model deep learning technologies for lung cancer detection,while beneficial,cannot capture the full range of features present in medical imaging data,leading to incomplete or inaccurate detection.Furthermore,it may not be robust enough to handle the wide variability in medical images due to different imaging conditions,patient anatomy,and tumor characteristics.To overcome these disadvantages,dual-model or multi-model approaches can be employed.This research focuses on enhancing the detection of lung cancer by utilizing a combination of two learning models:a Convolutional Neural Network(CNN)for categorization and the You Only Look Once(YOLOv8)architecture for real-time identification and pinpointing of tumors.CNNs automatically learn to extract hierarchical features from raw image data,capturing patterns such as edges,textures,and complex structures that are crucial for identifying lung cancer.YOLOv8 incorporates multiscale feature extraction,enabling the detection of tumors of varying sizes and scales within a single image.This is particularly beneficial for identifying small or irregularly shaped tumors that may be challenging to detect.Furthermore,through the utilization of cutting-edge data augmentation methods,such as Deep Convolutional Generative Adversarial Networks(DCGAN),the suggested approach can handle the issue of limited data and boost the models’ability to learn from diverse and comprehensive datasets.The combined method not only improved accuracy and localization but also ensured efficient real-time processing,which is crucial for practical clinical applications.The CNN achieved an accuracy of 97.67%in classifying lung tissues into healthy and cancerous categories.The YOLOv8 model achieved an Intersection over Union(IoU)score of 0.85 for tumor localization,reflecting high precision in detecting and marking tumor boundaries within the images.Finally,the incorporation of synthetic images generated by DCGAN led to a 10%improvement in both the CNN classification accuracy and YOLOv8 detection performance.展开更多
Sporadic E(Es)layers in the ionosphere are characterized by intense plasma irregularities in the E region at altitudes of 90-130 km.Because they can significantly influence radio communications and navigation systems,...Sporadic E(Es)layers in the ionosphere are characterized by intense plasma irregularities in the E region at altitudes of 90-130 km.Because they can significantly influence radio communications and navigation systems,accurate forecasting of Es layers is crucial for ensuring the precision and dependability of navigation satellite systems.In this study,we present Es predictions made by an empirical model and by a deep learning model,and analyze their differences comprehensively by comparing the model predictions to satellite RO measurements and ground-based ionosonde observations.The deep learning model exhibited significantly better performance,as indicated by its high coefficient of correlation(r=0.87)with RO observations and predictions,than did the empirical model(r=0.53).This study highlights the importance of integrating artificial intelligence technology into ionosphere modelling generally,and into predicting Es layer occurrences and characteristics,in particular.展开更多
The fast increase of online communities has brought about an increase in cyber threats inclusive of cyberbullying, hate speech, misinformation, and online harassment, making content moderation a pressing necessity. Tr...The fast increase of online communities has brought about an increase in cyber threats inclusive of cyberbullying, hate speech, misinformation, and online harassment, making content moderation a pressing necessity. Traditional single-modal AI-based detection systems, which analyze both text, photos, or movies in isolation, have established useless at taking pictures multi-modal threats, in which malicious actors spread dangerous content throughout a couple of formats. To cope with these demanding situations, we advise a multi-modal deep mastering framework that integrates Natural Language Processing (NLP), Convolutional Neural Networks (CNNs), and Long Short-Term Memory (LSTM) networks to become aware of and mitigate online threats effectively. Our proposed model combines BERT for text class, ResNet50 for photograph processing, and a hybrid LSTM-3-d CNN community for video content material analysis. We constructed a large-scale dataset comprising 500,000 textual posts, 200,000 offensive images, and 50,000 annotated motion pictures from more than one platform, which includes Twitter, Reddit, YouTube, and online gaming forums. The system became carefully evaluated using trendy gadget mastering metrics which include accuracy, precision, remember, F1-score, and ROC-AUC curves. Experimental outcomes demonstrate that our multi-modal method extensively outperforms single-modal AI classifiers, achieving an accuracy of 92.3%, precision of 91.2%, do not forget of 90.1%, and an AUC rating of 0.95. The findings validate the necessity of integrating multi-modal AI for actual-time, high-accuracy online chance detection and moderation. Future paintings will have consciousness on improving hostile robustness, enhancing scalability for real-world deployment, and addressing ethical worries associated with AI-driven content moderation.展开更多
Deep learning algorithms have been rapidly incorporated into many different applications due to the increase in computational power and the availability of massive amounts of data.Recently,both deep learning and ensem...Deep learning algorithms have been rapidly incorporated into many different applications due to the increase in computational power and the availability of massive amounts of data.Recently,both deep learning and ensemble learning have been used to recognize underlying structures and patterns from high-level features to make predictions/decisions.With the growth in popularity of deep learning and ensemble learning algorithms,they have received significant attention from both scientists and the industrial community due to their superior ability to learn features from big data.Ensemble deep learning has exhibited significant performance in enhancing learning generalization through the use of multiple deep learning algorithms.Although ensemble deep learning has large quantities of training parameters,which results in time and space overheads,it performs much better than traditional ensemble learning.Ensemble deep learning has been successfully used in several areas,such as bioinformatics,finance,and health care.In this paper,we review and investigate recent ensemble deep learning algorithms and techniques in health care domains,medical imaging,health care data analytics,genomics,diagnosis,disease prevention,and drug discovery.We cover several widely used deep learning algorithms along with their architectures,including deep neural networks(DNNs),convolutional neural networks(CNNs),recurrent neural networks(RNNs),and generative adversarial networks(GANs).Common healthcare tasks,such as medical imaging,electronic health records,and genomics,are also demonstrated.Furthermore,in this review,the challenges inherent in reducing the burden on the healthcare system are discussed and explored.Finally,future directions and opportunities for enhancing healthcare model performance are discussed.展开更多
Automated classification of retinal fundus images is essential for identifying eye diseases,though there is earlier research on applying deep learning models designed especially for detecting tessellation in retinal f...Automated classification of retinal fundus images is essential for identifying eye diseases,though there is earlier research on applying deep learning models designed especially for detecting tessellation in retinal fundus images.This study classifies 4 classes of retinal fundus images with 3 diseased fundus images and 1 normal fundus image,by creating a refined VGG16 model to categorize fundus pictures into tessellated,normal,myopia,and choroidal neovascularization groups.The approach utilizes a VGG16 architecture that has been altered with unique fully connected layers and regularization using dropouts,along with data augmentation techniques(rotation,flip,and rescale)on a dataset of 302 photos.Training involves class weighting and critical callbacks(early halting,learning rate reduction,checkpointing)to maximize performance.Gains in accuracy(93.42%training,77.5%validation)and improved class-specific F1 scores are attained.Grad-CAM’s Explainable AI(XAI)highlights areas of the images that are important for each categorization,making it interpretable for better understanding of medical experts.These results highlight the model’s potential as a helpful diagnostic tool in ophthalmology,providing a clear and practical method for the early identification and categorization of retinal disorders,especially in cases such as tessellated fundus images.展开更多
基金supported by the National Natural Science Foundation of China(No.52207229)the Key Research and Development Program of Ningxia Hui Autonomous Region of China(No.2024BEE02003)+1 种基金the financial support from the AEGiS Research Grant 2024,University of Wollongong(No.R6254)the financial support from the China Scholarship Council(No.202207550010).
文摘Accurate prediction of the remaining useful life(RUL)is crucial for the design and management of lithium-ion batteries.Although various machine learning models offer promising predictions,one critical but often overlooked challenge is their demand for considerable run-to-failure data for training.Collection of such training data leads to prohibitive testing efforts as the run-to-failure tests can last for years.Here,we propose a semi-supervised representation learning method to enhance prediction accuracy by learning from data without RUL labels.Our approach builds on a sophisticated deep neural network that comprises an encoder and three decoder heads to extract time-dependent representation features from short-term battery operating data regardless of the existence of RUL labels.The approach is validated using three datasets collected from 34 batteries operating under various conditions,encompassing over 19,900 charge and discharge cycles.Our method achieves a root mean squared error(RMSE)within 25 cycles,even when only 1/50 of the training dataset is labelled,representing a reduction of 48%compared to the conventional approach.We also demonstrate the method's robustness with varying numbers of labelled data and different weights assigned to the three decoder heads.The projection of extracted features in low space reveals that our method effectively learns degradation features from unlabelled data.Our approach highlights the promise of utilising semi-supervised learning to reduce the data demand for reliability monitoring of energy devices.
基金supported by the Natural Science Foundation Project of Fujian Province,China(Grant No.2023J011439 and No.2019J01859).
文摘With the rapid urbanization and exponential population growth in China,two-wheeled vehicles have become a popular mode of transportation,particularly for short-distance travel.However,due to a lack of safety awareness,traffic violations by two-wheeled vehicle riders have become a widespread concern,contributing to urban traffic risks.Currently,significant human and material resources are being allocated to monitor and intercept non-compliant riders to ensure safe driving behavior.To enhance the safety,efficiency,and cost-effectiveness of traffic monitoring,automated detection systems based on image processing algorithms can be employed to identify traffic violations from eye-level video footage.In this study,we propose a robust detection algorithm specifically designed for two-wheeled vehicles,which serves as a fundamental step toward intelligent traffic monitoring.Our approach integrates a novel convolutional and attention mechanism to improve detection accuracy and efficiency.Additionally,we introduce a semi-supervised training strategy that leverages a large number of unlabeled images to enhance the model’s learning capability by extracting valuable background information.This method enables the model to generalize effectively to diverse urban environments and varying lighting conditions.We evaluate our proposed algorithm on a custom-built dataset,and experimental results demonstrate its superior performance,achieving an average precision(AP)of 95%and a recall(R)of 90.6%.Furthermore,the model maintains a computational efficiency of only 25.7 GFLOPs while achieving a high processing speed of 249 FPS,making it highly suitable for deployment on edge devices.Compared to existing detection methods,our approach significantly enhances the accuracy and robustness of two-wheeled vehicle identification while ensuring real-time performance.
基金supported by the Ministry of Science and Technology of China,No.2020AAA0109605(to XL)Meizhou Major Scientific and Technological Innovation PlatformsProjects of Guangdong Provincial Science & Technology Plan Projects,No.2019A0102005(to HW).
文摘Early identification and treatment of stroke can greatly improve patient outcomes and quality of life.Although clinical tests such as the Cincinnati Pre-hospital Stroke Scale(CPSS)and the Face Arm Speech Test(FAST)are commonly used for stroke screening,accurate administration is dependent on specialized training.In this study,we proposed a novel multimodal deep learning approach,based on the FAST,for assessing suspected stroke patients exhibiting symptoms such as limb weakness,facial paresis,and speech disorders in acute settings.We collected a dataset comprising videos and audio recordings of emergency room patients performing designated limb movements,facial expressions,and speech tests based on the FAST.We compared the constructed deep learning model,which was designed to process multi-modal datasets,with six prior models that achieved good action classification performance,including the I3D,SlowFast,X3D,TPN,TimeSformer,and MViT.We found that the findings of our deep learning model had a higher clinical value compared with the other approaches.Moreover,the multi-modal model outperformed its single-module variants,highlighting the benefit of utilizing multiple types of patient data,such as action videos and speech audio.These results indicate that a multi-modal deep learning model combined with the FAST could greatly improve the accuracy and sensitivity of early stroke identification of stroke,thus providing a practical and powerful tool for assessing stroke patients in an emergency clinical setting.
文摘Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay differential equations(DDEs)and delay integrodifferential equations(DIDEs)with constant delays,primarily due to their low regularity at delayinduced breaking points.In this paper,a DNN method that combines multi-task learning(MTL)which is proposed to solve both the forward and inverse problems of DIDEs.The core idea of this approach is to divide the original equation into multiple tasks based on the delay,using auxiliary outputs to represent the integral terms,followed by the use of MTL to seamlessly incorporate the properties at the breaking points into the loss function.Furthermore,given the increased training dificulty associated with multiple tasks and outputs,we employ a sequential training scheme to reduce training complexity and provide reference solutions for subsequent tasks.This approach significantly enhances the approximation accuracy of solving DIDEs with DNNs,as demonstrated by comparisons with traditional DNN methods.We validate the effectiveness of this method through several numerical experiments,test various parameter sharing structures in MTL and compare the testing results of these structures.Finally,this method is implemented to solve the inverse problem of nonlinear DIDE and the results show that the unknown parameters of DIDE can be discovered with sparse or noisy data.
基金supported by the National Key Research and Development Program of China(Grant No.2023YFC3009400)the National Natural Science Foundation of China(Grant Nos.42307218 and U2239251).
文摘The current deep learning models for braced excavation cannot predict deformation from the beginning of excavation due to the need for a substantial corpus of sufficient historical data for training purposes.To address this issue,this study proposes a transfer learning model based on a sequence-to-sequence twodimensional(2D)convolutional long short-term memory neural network(S2SCL2D).The model can use the existing data from other adjacent similar excavations to achieve wall deflection prediction once a limited amount of monitoring data from the target excavation has been recorded.In the absence of adjacent excavation data,numerical simulation data from the target project can be employed instead.A weight update strategy is proposed to improve the prediction accuracy by integrating the stochastic gradient masking with an early stopping mechanism.To illustrate the proposed methodology,an excavation project in Hangzhou,China is adopted.The proposed deep transfer learning model,which uses either adjacent excavation data or numerical simulation data as the source domain,shows a significant improvement in performance when compared to the non-transfer learning model.Using the simulation data from the target project even leads to better prediction performance than using the actual monitoring data from other adjacent excavations.The results demonstrate that the proposed model can reasonably predict the deformation with limited data from the target project.
文摘Bearing is an indispensable key component in mechanical equipment,and its working state is directly related to the stability and safety of the whole equipment.In recent years,with the rapid development of artificial intelligence technology,especially the breakthrough of deep learning technology,it provides a new idea for bearing fault diagnosis.Deep learning can automatically learn features from a large amount of data,has a strong nonlinear modeling ability,and can effectively solve the problems existing in traditional methods.Aiming at the key problems in bearing fault diagnosis,this paper studies the fault diagnosis method based on deep learning,which not only provides a new solution for bearing fault diagnosis but also provides a reference for the application of deep learning in other mechanical fault diagnosis fields.
文摘In computer vision and artificial intelligence,automatic facial expression-based emotion identification of humans has become a popular research and industry problem.Recent demonstrations and applications in several fields,including computer games,smart homes,expression analysis,gesture recognition,surveillance films,depression therapy,patientmonitoring,anxiety,and others,have brought attention to its significant academic and commercial importance.This study emphasizes research that has only employed facial images for face expression recognition(FER),because facial expressions are a basic way that people communicate meaning to each other.The immense achievement of deep learning has resulted in a growing use of its much architecture to enhance efficiency.This review is on machine learning,deep learning,and hybrid methods’use of preprocessing,augmentation techniques,and feature extraction for temporal properties of successive frames of data.The following section gives a brief summary of assessment criteria that are accessible to the public and then compares them with benchmark results the most trustworthy way to assess FER-related research topics statistically.In this review,a brief synopsis of the subject matter may be beneficial for novices in the field of FER as well as seasoned scholars seeking fruitful avenues for further investigation.The information conveys fundamental knowledge and provides a comprehensive understanding of the most recent state-of-the-art research.
基金Supported by Clinical Trials from the Third Affiliated Hospital of Soochow University,No.2024-156Changzhou Science and Technology Program,No.CJ20244017。
文摘BACKGROUND Hepatocellular carcinoma(HCC)is a prevalent and life-threatening cancer with increasing incidence worldwide.High Ki-67 risk stratification is closely associated with higher recurrence rates and worse outcomes following curative therapies in patients with HCC.However,the performance of radiomic and deep transfer learning(DTL)models derived from biparametric magnetic resonance imaging(bpMRI)in predicting Ki-67 risk stratification and recurrence-free survival(RFS)in patients with HCC remains limited.AIM To develop a nomogram model integrating bpMRI-based radiomic and DTL signatures for predicting Ki-67 risk stratification and RFS in patients with HCC.METHODS This study included 198 patients with histopathologically confirmed HCC who underwent preoperative bpMRI.Ki-67 risk stratification was categorized as high(>20%)or low(≤20%)according to immunohistochemical staining.Radiomic and DTL signatures were extracted from the T2-weighted and arterial-phase images and combined through a random forest algorithm to establish radiomic and DTL models,respectively.Multivariate regression analysis identified clinical risk factors for high Ki-67 risk stratification,and a predictive nomogram model was developed.RESULTS A nonsmooth margin and the absence of an enhanced capsule were independent factors for high Ki-67 risk stratification.The area under the curve(AUC)of the clinical model was 0.77,while those of the radiomic and DTL models were 0.81 and 0.87,respectively,for the prediction of high Ki-67 risk stratification,and the nomogram model achieved a better AUC of 0.92.The median RFS times for patients with high and low Ki-67 risk stratification were 33.00 months and 66.73 months,respectively(P<0.001).Additionally,patients who were predicted to have high Ki-67 risk stratification by the nomogram model had a lower median RFS than those who were predicted to have low Ki-67 risk stratification(33.53 vs 66.74 months,P=0.007).CONCLUSION Our developed nomogram model demonstrated good performance in predicting Ki-67 risk stratification and predicting survival outcomes in patients with HCC.
基金upported by the National Key Research and Development Program of China(Grant No.:2023YFF1204904)the National Natural Science Foundation of China(Grant Nos.:U23A20530 and 82173746)Shanghai Frontiers Science Center of Optogenetic Techniques for Cell Metabolism(Shanghai Municipal Education Commission,China).
文摘Negative logarithm of the acid dissociation constant(pK_(a))significantly influences the absorption,dis-tribution,metabolism,excretion,and toxicity(ADMET)properties of molecules and is a crucial indicator in drug research.Given the rapid and accurate characteristics of computational methods,their role in predicting drug properties is increasingly important.Although many pK_(a) prediction models currently exist,they often focus on enhancing model precision while neglecting interpretability.In this study,we present GraFpKa,a pK_(a) prediction model using graph neural networks(GNNs)and molecular finger-prints.The results show that our acidic and basic models achieved mean absolute errors(MAEs)of 0.621 and 0.402,respectively,on the test set,demonstrating good predictive performance.Notably,to improve interpretability,GraFpKa also incorporates Integrated Gradients(IGs),providing a clearer visual description of the atoms significantly affecting the pK_(a) values.The high reliability and interpretability of GraFpKa ensure accurate pKa predictions while also facilitating a deeper understanding of the relation-ship between molecular structure and pK_(a) values,making it a valuable tool in the field of pK_(a) prediction.
基金Deep-time Digital Earth(DDE)Big Science Program(No.GJ-C03-SGF-2025-004)National Natural Science Foundation of China(No.42394063)Sichuan Science and Technology Program(No.2025ZNSFSC0325).
文摘Topographic maps,as essential tools and sources of information for geographic research,contain precise spatial locations and rich map features,and they illustrate spatio-temporal information on the distribution and differences of various surface features.Currently,topographic maps are mainly stored in raster and vector formats.Extraction of the spatio-temporal knowledge in the maps—such as spatial distribution patterns,feature relationships,and dynamic evolution—still primarily relies on manual interpretation.However,manual interpretation is time-consuming and laborious,especially for large-scale,long-term map knowledge extraction and application.With the development of artificial intelligence technology,it is possible to improve the automation level of map knowledge interpretation.Therefore,the present study proposes an automatic interpretation method for raster topographic map knowledge based on deep learning.To address the limitations of current data-driven intelligent technology in learning map spatial relations and cognitive logic,we establish a formal description of map knowledge by mapping the relationship between map knowledge and features,thereby ensuring interpretation accuracy.Subsequently,deep learning techniques are employed to extract map features automatically,and the spatio-temporal knowledge is constructed by combining formal descriptions of geographic feature knowledge.Validation experiments demonstrate that the proposed method effectively achieves automatic interpretation of spatio-temporal knowledge of geographic features in maps,with an accuracy exceeding 80%.The findings of the present study contribute to machine understanding of spatio-temporal differences in map knowledge and advances the intelligent interpretation and utilization of cartographic information.
文摘Multimodal deep learning has emerged as a key paradigm in contemporary medical diagnostics,advancing precision medicine by enabling integration and learning from diverse data sources.The exponential growth of high-dimensional healthcare data,encompassing genomic,transcriptomic,and other omics profiles,as well as radiological imaging and histopathological slides,makes this approach increasingly important because,when examined separately,these data sources only offer a fragmented picture of intricate disease processes.Multimodal deep learning leverages the complementary properties of multiple data modalities to enable more accurate prognostic modeling,more robust disease characterization,and improved treatment decision-making.This review provides a comprehensive overview of the current state of multimodal deep learning approaches in medical diagnosis.We classify and examine important application domains,such as(1)radiology,where automated report generation and lesion detection are facilitated by image-text integration;(2)histopathology,where fusion models improve tumor classification and grading;and(3)multi-omics,where molecular subtypes and latent biomarkers are revealed through cross-modal learning.We provide an overview of representative research,methodological advancements,and clinical consequences for each domain.Additionally,we critically analyzed the fundamental issues preventing wider adoption,including computational complexity(particularly in training scalable,multi-branch networks),data heterogeneity(resulting from modality-specific noise,resolution variations,and inconsistent annotations),and the challenge of maintaining significant cross-modal correlations during fusion.These problems impede interpretability,which is crucial for clinical trust and use,in addition to performance and generalizability.Lastly,we outline important areas for future research,including the development of standardized protocols for harmonizing data,the creation of lightweight and interpretable fusion architectures,the integration of real-time clinical decision support systems,and the promotion of cooperation for federated multimodal learning.Our goal is to provide researchers and clinicians with a concise overview of the field’s present state,enduring constraints,and exciting directions for further research through this review.
基金the“Intelligent Recognition Industry Service Center”as part of the Featured Areas Research Center Program under the Higher Education Sprout Project by the Ministry of Education(MOE)in Taiwan,and the National Science and Technology Council,Taiwan,under grants 113-2221-E-224-041 and 113-2622-E-224-002.Additionally,partial support was provided by Isuzu Optics Corporation.
文摘Liver cancer remains a leading cause of mortality worldwide,and precise diagnostic tools are essential for effective treatment planning.Liver Tumors(LTs)vary significantly in size,shape,and location,and can present with tissues of similar intensities,making automatically segmenting and classifying LTs from abdominal tomography images crucial and challenging.This review examines recent advancements in Liver Segmentation(LS)and Tumor Segmentation(TS)algorithms,highlighting their strengths and limitations regarding precision,automation,and resilience.Performance metrics are utilized to assess key detection algorithms and analytical methods,emphasizing their effectiveness and relevance in clinical contexts.The review also addresses ongoing challenges in liver tumor segmentation and identification,such as managing high variability in patient data and ensuring robustness across different imaging conditions.It suggests directions for future research,with insights into technological advancements that can enhance surgical planning and diagnostic accuracy by comparing popular methods.This paper contributes to a comprehensive understanding of current liver tumor detection techniques,provides a roadmap for future innovations,and improves diagnostic and therapeutic outcomes for liver cancer by integrating recent progress with remaining challenges.
基金supported by the Deanship of Scientific Research(DSR)at King Abdulaziz University,Jeddah,under grant no.(GPIP:13-612-2024).
文摘App reviews are crucial in influencing user decisions and providing essential feedback for developers to improve their products.Automating the analysis of these reviews is vital for efficient review management.While traditional machine learning(ML)models rely on basic word-based feature extraction,deep learning(DL)methods,enhanced with advanced word embeddings,have shown superior performance.This research introduces a novel aspectbased sentiment analysis(ABSA)framework to classify app reviews based on key non-functional requirements,focusing on usability factors:effectiveness,efficiency,and satisfaction.We propose a hybrid DL model,combining BERT(Bidirectional Encoder Representations from Transformers)with BiLSTM(Bidirectional Long Short-Term Memory)and CNN(Convolutional Neural Networks)layers,to enhance classification accuracy.Comparative analysis against state-of-the-art models demonstrates that our BERT-BiLSTM-CNN model achieves exceptional performance,with precision,recall,F1-score,and accuracy of 96%,87%,91%,and 94%,respectively.Thesignificant contributions of this work include a refined ABSA-based relabeling framework,the development of a highperformance classifier,and the comprehensive relabeling of the Instagram App Reviews dataset.These advancements provide valuable insights for software developers to enhance usability and drive user-centric application development.
基金supported in part by the Guangxi Science and Technology Major Program under grant AA22068067the Guangxi Natural Science Foundation under grant 2023GXNSFAA026236 and 2024GXNSFDA010064the National Natural Science Foundation of China under project 62172119.
文摘As more and more devices in Cyber-Physical Systems(CPS)are connected to the Internet,physical components such as programmable logic controller(PLC),sensors,and actuators are facing greater risks of network attacks,and fast and accurate attack detection techniques are crucial.The key problem in distinguishing between normal and abnormal sequences is to model sequential changes in a large and diverse field of time series.To address this issue,we propose an anomaly detection method based on distributed deep learning.Our method uses a bilateral filtering algorithm for sequential sequences to remove noise in the time series,which can maintain the edge of discrete features.We use a distributed linear deep learning model to establish a sequential prediction model and adjust the threshold for anomaly detection based on the prediction error of the validation set.Our method can not only detect abnormal attacks but also locate the sensors that cause anomalies.We conducted experiments on the Secure Water Treatment(SWAT)and Water Distribution(WADI)public datasets.The experimental results show that our method is superior to the baseline method in identifying the types of attacks and detecting efficiency.
基金co-supported by the National Natural Science Foundation of China(No.62103432)the China Postdoctoral Science Foundation(No.284881)the Young Talent fund of University Association for Science and Technology in Shaanxi,China(No.20210108)。
文摘Exo-atmospheric vehicles are constrained by limited maneuverability,which leads to the contradiction between evasive maneuver and precision strike.To address the problem of Integrated Evasion and Impact(IEI)decision under multi-constraint conditions,a hierarchical intelligent decision-making method based on Deep Reinforcement Learning(DRL)was proposed.First,an intelligent decision-making framework of“DRL evasion decision”+“impact prediction guidance decision”was established:it takes the impact point deviation correction ability as the constraint and the maximum miss distance as the objective,and effectively solves the problem of poor decisionmaking effect caused by the large IEI decision space.Second,to solve the sparse reward problem faced by evasion decision-making,a hierarchical decision-making method consisting of maneuver timing decision and maneuver duration decision was proposed,and the corresponding Markov Decision Process(MDP)was designed.A detailed simulation experiment was designed to analyze the advantages and computational complexity of the proposed method.Simulation results show that the proposed model has good performance and low computational resource requirement.The minimum miss distance is 21.3 m under the condition of guaranteeing the impact point accuracy,and the single decision-making time is 4.086 ms on an STM32F407 single-chip microcomputer,which has engineering application value.
文摘Lung cancer continues to be a leading cause of cancer-related deaths worldwide,emphasizing the critical need for improved diagnostic techniques.Early detection of lung tumors significantly increases the chances of successful treatment and survival.However,current diagnostic methods often fail to detect tumors at an early stage or to accurately pinpoint their location within the lung tissue.Single-model deep learning technologies for lung cancer detection,while beneficial,cannot capture the full range of features present in medical imaging data,leading to incomplete or inaccurate detection.Furthermore,it may not be robust enough to handle the wide variability in medical images due to different imaging conditions,patient anatomy,and tumor characteristics.To overcome these disadvantages,dual-model or multi-model approaches can be employed.This research focuses on enhancing the detection of lung cancer by utilizing a combination of two learning models:a Convolutional Neural Network(CNN)for categorization and the You Only Look Once(YOLOv8)architecture for real-time identification and pinpointing of tumors.CNNs automatically learn to extract hierarchical features from raw image data,capturing patterns such as edges,textures,and complex structures that are crucial for identifying lung cancer.YOLOv8 incorporates multiscale feature extraction,enabling the detection of tumors of varying sizes and scales within a single image.This is particularly beneficial for identifying small or irregularly shaped tumors that may be challenging to detect.Furthermore,through the utilization of cutting-edge data augmentation methods,such as Deep Convolutional Generative Adversarial Networks(DCGAN),the suggested approach can handle the issue of limited data and boost the models’ability to learn from diverse and comprehensive datasets.The combined method not only improved accuracy and localization but also ensured efficient real-time processing,which is crucial for practical clinical applications.The CNN achieved an accuracy of 97.67%in classifying lung tissues into healthy and cancerous categories.The YOLOv8 model achieved an Intersection over Union(IoU)score of 0.85 for tumor localization,reflecting high precision in detecting and marking tumor boundaries within the images.Finally,the incorporation of synthetic images generated by DCGAN led to a 10%improvement in both the CNN classification accuracy and YOLOv8 detection performance.
基金supported by the Project of Stable Support for Youth Team in Basic Research Field,CAS(grant No.YSBR-018)the National Natural Science Foundation of China(grant Nos.42188101,42130204)+4 种基金the B-type Strategic Priority Program of CAS(grant no.XDB41000000)the National Natural Science Foundation of China(NSFC)Distinguished Overseas Young Talents Program,Innovation Program for Quantum Science and Technology(2021ZD0300301)the Open Research Project of Large Research Infrastructures of CAS-“Study on the interaction between low/mid-latitude atmosphere and ionosphere based on the Chinese Meridian Project”.The project was supported also by the National Key Laboratory of Deep Space Exploration(Grant No.NKLDSE2023A002)the Open Fund of Anhui Provincial Key Laboratory of Intelligent Underground Detection(Grant No.APKLIUD23KF01)the China National Space Administration(CNSA)pre-research Project on Civil Aerospace Technologies No.D010305,D010301.
文摘Sporadic E(Es)layers in the ionosphere are characterized by intense plasma irregularities in the E region at altitudes of 90-130 km.Because they can significantly influence radio communications and navigation systems,accurate forecasting of Es layers is crucial for ensuring the precision and dependability of navigation satellite systems.In this study,we present Es predictions made by an empirical model and by a deep learning model,and analyze their differences comprehensively by comparing the model predictions to satellite RO measurements and ground-based ionosonde observations.The deep learning model exhibited significantly better performance,as indicated by its high coefficient of correlation(r=0.87)with RO observations and predictions,than did the empirical model(r=0.53).This study highlights the importance of integrating artificial intelligence technology into ionosphere modelling generally,and into predicting Es layer occurrences and characteristics,in particular.
文摘The fast increase of online communities has brought about an increase in cyber threats inclusive of cyberbullying, hate speech, misinformation, and online harassment, making content moderation a pressing necessity. Traditional single-modal AI-based detection systems, which analyze both text, photos, or movies in isolation, have established useless at taking pictures multi-modal threats, in which malicious actors spread dangerous content throughout a couple of formats. To cope with these demanding situations, we advise a multi-modal deep mastering framework that integrates Natural Language Processing (NLP), Convolutional Neural Networks (CNNs), and Long Short-Term Memory (LSTM) networks to become aware of and mitigate online threats effectively. Our proposed model combines BERT for text class, ResNet50 for photograph processing, and a hybrid LSTM-3-d CNN community for video content material analysis. We constructed a large-scale dataset comprising 500,000 textual posts, 200,000 offensive images, and 50,000 annotated motion pictures from more than one platform, which includes Twitter, Reddit, YouTube, and online gaming forums. The system became carefully evaluated using trendy gadget mastering metrics which include accuracy, precision, remember, F1-score, and ROC-AUC curves. Experimental outcomes demonstrate that our multi-modal method extensively outperforms single-modal AI classifiers, achieving an accuracy of 92.3%, precision of 91.2%, do not forget of 90.1%, and an AUC rating of 0.95. The findings validate the necessity of integrating multi-modal AI for actual-time, high-accuracy online chance detection and moderation. Future paintings will have consciousness on improving hostile robustness, enhancing scalability for real-world deployment, and addressing ethical worries associated with AI-driven content moderation.
基金funded by Taif University,Saudi Arabia,project No.(TU-DSPP-2024-263).
文摘Deep learning algorithms have been rapidly incorporated into many different applications due to the increase in computational power and the availability of massive amounts of data.Recently,both deep learning and ensemble learning have been used to recognize underlying structures and patterns from high-level features to make predictions/decisions.With the growth in popularity of deep learning and ensemble learning algorithms,they have received significant attention from both scientists and the industrial community due to their superior ability to learn features from big data.Ensemble deep learning has exhibited significant performance in enhancing learning generalization through the use of multiple deep learning algorithms.Although ensemble deep learning has large quantities of training parameters,which results in time and space overheads,it performs much better than traditional ensemble learning.Ensemble deep learning has been successfully used in several areas,such as bioinformatics,finance,and health care.In this paper,we review and investigate recent ensemble deep learning algorithms and techniques in health care domains,medical imaging,health care data analytics,genomics,diagnosis,disease prevention,and drug discovery.We cover several widely used deep learning algorithms along with their architectures,including deep neural networks(DNNs),convolutional neural networks(CNNs),recurrent neural networks(RNNs),and generative adversarial networks(GANs).Common healthcare tasks,such as medical imaging,electronic health records,and genomics,are also demonstrated.Furthermore,in this review,the challenges inherent in reducing the burden on the healthcare system are discussed and explored.Finally,future directions and opportunities for enhancing healthcare model performance are discussed.
基金support from the"Intelligent Recognition Industry Service Center"as part of the Featured Areas Research Center Program under the Higher Education Sprout Project by the Ministry of Education(MOE)in Taiwan,and the National Science and Technology Council,Taiwan,under grants[113-2622-E-224-002]and[113-2221-E-224-041]support was provided by Isuzu Optics Corporation.
文摘Automated classification of retinal fundus images is essential for identifying eye diseases,though there is earlier research on applying deep learning models designed especially for detecting tessellation in retinal fundus images.This study classifies 4 classes of retinal fundus images with 3 diseased fundus images and 1 normal fundus image,by creating a refined VGG16 model to categorize fundus pictures into tessellated,normal,myopia,and choroidal neovascularization groups.The approach utilizes a VGG16 architecture that has been altered with unique fully connected layers and regularization using dropouts,along with data augmentation techniques(rotation,flip,and rescale)on a dataset of 302 photos.Training involves class weighting and critical callbacks(early halting,learning rate reduction,checkpointing)to maximize performance.Gains in accuracy(93.42%training,77.5%validation)and improved class-specific F1 scores are attained.Grad-CAM’s Explainable AI(XAI)highlights areas of the images that are important for each categorization,making it interpretable for better understanding of medical experts.These results highlight the model’s potential as a helpful diagnostic tool in ophthalmology,providing a clear and practical method for the early identification and categorization of retinal disorders,especially in cases such as tessellated fundus images.