This paper addresses the performance degradation issue in a fast radio burst search pipeline based on deep learning.This issue is caused by the class imbalance of the radio frequency interference samples in the traini...This paper addresses the performance degradation issue in a fast radio burst search pipeline based on deep learning.This issue is caused by the class imbalance of the radio frequency interference samples in the training dataset,and one solution is applied to improve the distribution of the training data by augmenting minority class samples using a deep convolutional generative adversarial network.Experi.mental results demonstrate that retraining the deep learning model with the newly generated dataset leads to a new fast radio burst classifier,which effectively reduces false positives caused by periodic wide-band impulsive radio frequency interference,thereby enhancing the performance of the search pipeline.展开更多
Network architectures assisted by Generative Artificial Intelligence(GAI)are envisioned as foundational elements of sixth-generation(6G)communication system.To deliver ubiquitous intelligent services and meet diverse ...Network architectures assisted by Generative Artificial Intelligence(GAI)are envisioned as foundational elements of sixth-generation(6G)communication system.To deliver ubiquitous intelligent services and meet diverse service requirements,6G network architecture should offer personalized services to various mobile devices.Federated learning(FL)with personalized local training,as a privacypreserving machine learning(ML)approach,can be applied to address these challenges.In this paper,we propose a meta-learning-based personalized FL(PFL)method that improves both communication and computation efficiency by utilizing over-the-air computations.Its“pretraining-and-fine-tuning”principle makes it particularly suitable for enabling edge nodes to access personalized GAI services while preserving local privacy.Experiment results demonstrate the outperformance and efficacy of the proposed algorithm,and notably indicate enhanced communication efficiency without compromising accuracy.展开更多
Hydrogen energy is a crucial support for China’s low-carbon energy transition.With the large-scale integration of renewable energy,the combination of hydrogen and integrated energy systems has become one of the most ...Hydrogen energy is a crucial support for China’s low-carbon energy transition.With the large-scale integration of renewable energy,the combination of hydrogen and integrated energy systems has become one of the most promising directions of development.This paper proposes an optimized schedulingmodel for a hydrogen-coupled electro-heat-gas integrated energy system(HCEHG-IES)using generative adversarial imitation learning(GAIL).The model aims to enhance renewable-energy absorption,reduce carbon emissions,and improve grid-regulation flexibility.First,the optimal scheduling problem of HCEHG-IES under uncertainty is modeled as a Markov decision process(MDP).To overcome the limitations of conventional deep reinforcement learning algorithms—including long optimization time,slow convergence,and subjective reward design—this study augments the PPO algorithm by incorporating a discriminator network and expert data.The newly developed algorithm,termed GAIL,enables the agent to perform imitation learning from expert data.Based on this model,dynamic scheduling decisions are made in continuous state and action spaces,generating optimal energy-allocation and management schemes.Simulation results indicate that,compared with traditional reinforcement-learning algorithms,the proposed algorithmoffers better economic performance.Guided by expert data,the agent avoids blind optimization,shortens the offline training time,and improves convergence performance.In the online phase,the algorithm enables flexible energy utilization,thereby promoting renewable-energy absorption and reducing carbon emissions.展开更多
This study explores a novel educational model of generative AI-empowered interdisciplinary project-based learning(PBL).By analyzing the current applications of generative AI technology in information technology curric...This study explores a novel educational model of generative AI-empowered interdisciplinary project-based learning(PBL).By analyzing the current applications of generative AI technology in information technology curricula,it elucidates its advantages and operational mechanisms in interdisciplinary PBL.Combining case studies and empirical research,the investigation proposes implementation pathways and strategies for the generative AI-enhanced interdisciplinary PBL model,detailing specific applications across three phases:project preparation,implementation,and evaluation.The research demonstrates that generative AI-enabled interdisciplinary project-based learning can effectively enhance students’learning motivation,interdisciplinary thinking capabilities,and innovative competencies,providing new conceptual frameworks and practical approaches for educational model innovation.展开更多
Deep Learning(DL)is such a powerful tool that we have seen tremendous success in areas such as Computer Vision,Speech Recognition,and Natural Language Processing.Since Automated Modulation Classification(AMC)is an imp...Deep Learning(DL)is such a powerful tool that we have seen tremendous success in areas such as Computer Vision,Speech Recognition,and Natural Language Processing.Since Automated Modulation Classification(AMC)is an important part in Cognitive Radio Networks,we try to explore its potential in solving signal modulation recognition problem.It cannot be overlooked that DL model is a complex model,thus making them prone to over-fitting.DL model requires many training data to combat with over-fitting,but adding high quality labels to training data manually is not always cheap and accessible,especially in real-time system,which may counter unprecedented data in dataset.Semi-supervised Learning is a way to exploit unlabeled data effectively to reduce over-fitting in DL.In this paper,we extend Generative Adversarial Networks(GANs)to the semi-supervised learning will show it is a method can be used to create a more dataefficient classifier.展开更多
In recent years,landslide susceptibility mapping has substantially improved with advances in machine learning.However,there are still challenges remain in landslide mapping due to the availability of limited inventory...In recent years,landslide susceptibility mapping has substantially improved with advances in machine learning.However,there are still challenges remain in landslide mapping due to the availability of limited inventory data.In this paper,a novel method that improves the performance of machine learning techniques is presented.The proposed method creates synthetic inventory data using Generative Adversarial Networks(GANs)for improving the prediction of landslides.In this research,landslide inventory data of 156 landslide locations were identified in Cameron Highlands,Malaysia,taken from previous projects the authors worked on.Elevation,slope,aspect,plan curvature,profile curvature,total curvature,lithology,land use and land cover(LULC),distance to the road,distance to the river,stream power index(SPI),sediment transport index(STI),terrain roughness index(TRI),topographic wetness index(TWI)and vegetation density are geo-environmental factors considered in this study based on suggestions from previous works on Cameron Highlands.To show the capability of GANs in improving landslide prediction models,this study tests the proposed GAN model with benchmark models namely Artificial Neural Network(ANN),Support Vector Machine(SVM),Decision Trees(DT),Random Forest(RF)and Bagging ensemble models with ANN and SVM models.These models were validated using the area under the receiver operating characteristic curve(AUROC).The DT,RF,SVM,ANN and Bagging ensemble could achieve the AUROC values of(0.90,0.94,0.86,0.69 and 0.82)for the training;and the AUROC of(0.76,0.81,0.85,0.72 and 0.75)for the test,subsequently.When using additional samples,the same models achieved the AUROC values of(0.92,0.94,0.88,0.75 and 0.84)for the training and(0.78,0.82,0.82,0.78 and 0.80)for the test,respectively.Using the additional samples improved the test accuracy of all the models except SVM.As a result,in data-scarce environments,this research showed that utilizing GANs to generate supplementary samples is promising because it can improve the predictive capability of common landslide prediction models.展开更多
Generative adversarial network(GAN)has achieved great success in many fields such as computer vision,speech processing,and natural language processing,because of its powerful capabilities for generating realistic samp...Generative adversarial network(GAN)has achieved great success in many fields such as computer vision,speech processing,and natural language processing,because of its powerful capabilities for generating realistic samples.In this paper,we introduce GAN into the field of electromagnetic signal classification(ESC).ESC plays an important role in both military and civilian domains.However,in many specific scenarios,we can’t obtain enough labeled data,which cause failure of deep learning methods because they are easy to fall into over-fitting.Fortunately,semi-supervised learning(SSL)can leverage the large amount of unlabeled data to enhance the classification performance of classifiers,especially in scenarios with limited amount of labeled data.We present an SSL framework by incorporating GAN,which can directly process the raw in-phase and quadrature(IQ)signal data.According to the characteristics of the electromagnetic signal,we propose a weighted loss function,leading to an effective classifier to realize the end-to-end classification of the electromagnetic signal.We validate the proposed method on both public RML2016.04c dataset and real-world Aircraft Communications Addressing and Reporting System(ACARS)signal dataset.Extensive experimental results show that the proposed framework obtains a significant increase in classification accuracy compared with the state-of-the-art studies.展开更多
Malaria is a lethal disease responsible for thousands of deaths worldwide every year.Manual methods of malaria diagnosis are timeconsuming that require a great deal of human expertise and efforts.Computerbased automat...Malaria is a lethal disease responsible for thousands of deaths worldwide every year.Manual methods of malaria diagnosis are timeconsuming that require a great deal of human expertise and efforts.Computerbased automated diagnosis of diseases is progressively becoming popular.Although deep learning models show high performance in the medical field,it demands a large volume of data for training which is hard to acquire for medical problems.Similarly,labeling of medical images can be done with the help of medical experts only.Several recent studies have utilized deep learning models to develop efficient malaria diagnostic system,which showed promising results.However,the most common problem with these models is that they need a large amount of data for training.This paper presents a computer-aided malaria diagnosis system that combines a semi-supervised generative adversarial network and transfer learning.The proposed model is trained in a semi-supervised manner and requires less training data than conventional deep learning models.Performance of the proposed model is evaluated on a publicly available dataset of blood smear images(with malariainfected and normal class)and achieved a classification accuracy of 96.6%.展开更多
Robot calligraphy visually reflects the motion capability of robotic manipulators.While traditional researches mainly focus on image generation and the writing of simple calligraphic strokes or characters,this article...Robot calligraphy visually reflects the motion capability of robotic manipulators.While traditional researches mainly focus on image generation and the writing of simple calligraphic strokes or characters,this article presents a generative adversarial network(GAN)-based motion learning method for robotic calligraphy synthesis(Gan2CS)that can enhance the efficiency in writing complex calligraphy words and reproducing classic calligraphy works.The key technologies in the proposed approach include:(1)adopting the GAN to learn the motion parameters from the robot writing operation;(2)converting the learnt motion data into the style font and realising the transition from static calligraphy images to dynamic writing demonstration;(3)reproducing high-precision calligraphy works by synthesising the writing motion data hierarchically.In this study,the motion trajectories of sample calligraphy images are firstly extracted and converted into the robot module.The robot performs the writing with motion planning,and the writing motion parameters of calligraphy strokes are learnt with GANs.Then the motion data of basic strokes is synthesised based on the hierarchical process of‘stroke-radicalpart-character’.And the robot re-writes the synthesised characters whose similarity with the original calligraphy characters is evaluated.Regular calligraphy characters have been tested in the experiments for method validation and the results validated that the robot can actualise the robotic calligraphy synthesis of writing motion data with GAN.展开更多
With the continuous scaling of integrated circuit technologies,design for manufacturability(DFM)is becoming more critical,yet more challenging.Alongside,recent advances in machine learning have provided a new computin...With the continuous scaling of integrated circuit technologies,design for manufacturability(DFM)is becoming more critical,yet more challenging.Alongside,recent advances in machine learning have provided a new computing paradigm with promising applications in VLSI manufacturability.In particular,generative learning-regarded among the most interesting ideas in present-day machine learning-has demonstrated impressive capabilities in a wide range of applications.This paper surveys recent results of using generative learning in VLSI manufacturing modeling and optimization.Specifically,we examine the unique features of generative learning that have been leveraged to improve DFM efficiency in an unprecedented way;hence,paving the way to a new data-driven DFM approach.The state-of-the-art methods are presented,and challenges/opportunities are discussed.展开更多
As energy demands continue to rise in modern society,the development of high-performance lithium-ion batteries(LIBs)has become crucial.However,traditional research methods of material science face challenges such as l...As energy demands continue to rise in modern society,the development of high-performance lithium-ion batteries(LIBs)has become crucial.However,traditional research methods of material science face challenges such as lengthy timelines and complex processes.In recent years,the integration of machine learning(ML)in LIB materials,including electrolytes,solid-state electrolytes,and electrodes,has yielded remarkable achievements.This comprehensive review explores the latest applications of ML in predicting LIB material performance,covering the core principles and recent advancements in three key inverse material design strategies:high-throughput virtual screening,global optimization,and generative models.These strategies have played a pivotal role in fostering LIB material innovations.Meanwhile,the paper briefly discusses the challenges associated with applying ML to materials research and offers insights and directions for future research.展开更多
Learning is no longer regarded as knowledge acquisition but as knowledge construction;therefore,learner is not a passive recipient but an active constructor.In generative and elaborative learning process,the learner’...Learning is no longer regarded as knowledge acquisition but as knowledge construction;therefore,learner is not a passive recipient but an active constructor.In generative and elaborative learning process,the learner’s existing knowledge is considered to be modified while it is used to construct a meaning from the text.Therefore,the focus in learning is on generating relations,rather than on storing information.This paper aims to draw the outlines of the main ideas of generative learning and their related concepts and theoretical significance.展开更多
Efficiently tracking and imaging interested moving targets is crucial across various applications,from autonomous systems to surveillance.However,persistent challenges remain in various fields,including environmental ...Efficiently tracking and imaging interested moving targets is crucial across various applications,from autonomous systems to surveillance.However,persistent challenges remain in various fields,including environmental intricacies,limitations in perceptual technologies,and privacy considerations.We present a teacher-student learning model,the generative adversarial network(GAN)-guided diffractive neural network(DNN),which performs visual tracking and imaging of the interested moving target.The GAN,as a teacher model,empowers efficient acquisition of the skill to differentiate the specific target of interest in the domains of visual tracking and imaging.The DNN-based student model learns to master the skill to differentiate the interested target from the GAN.The process of obtaining a GAN-guided DNN starts with capturing moving objects effectively using an event camera with high temporal resolution and low latency.Then,the generative power of GAN is utilized to generate data with position-tracking capability for the interested moving target,subsequently serving as labels to the training of the DNN.The DNN learns to image the target during training while retaining the target’s positional information.Our experimental demonstration highlights the efficacy of the GAN-guided DNN in visual tracking and imaging of the interested moving target.We expect the GAN-guided DNN can significantly enhance autonomous systems and surveillance.展开更多
The Internet of Things(IoT)is integral to modern infrastructure,enabling connectivity among a wide range of devices from home automation to industrial control systems.With the exponential increase in data generated by...The Internet of Things(IoT)is integral to modern infrastructure,enabling connectivity among a wide range of devices from home automation to industrial control systems.With the exponential increase in data generated by these interconnected devices,robust anomaly detection mechanisms are essential.Anomaly detection in this dynamic environment necessitates methods that can accurately distinguish between normal and anomalous behavior by learning intricate patterns.This paper presents a novel approach utilizing generative adversarial networks(GANs)for anomaly detection in IoT systems.However,optimizing GANs involves tuning hyper-parameters such as learning rate,batch size,and optimization algorithms,which can be challenging due to the non-convex nature of GAN loss functions.To address this,we propose a five-dimensional Gray wolf optimizer(5DGWO)to optimize GAN hyper-parameters.The 5DGWO introduces two new types of wolves:gamma(γ)for improved exploitation and convergence,and theta(θ)for enhanced exploration and escaping local minima.The proposed system framework comprises four key stages:1)preprocessing,2)generative model training,3)autoencoder(AE)training,and 4)predictive model training.The generative models are utilized to assist the AE training,and the final predictive models(including convolutional neural network(CNN),deep belief network(DBN),recurrent neural network(RNN),random forest(RF),and extreme gradient boosting(XGBoost))are trained using the generated data and AE-encoded features.We evaluated the system on three benchmark datasets:NSL-KDD,UNSW-NB15,and IoT-23.Experiments conducted on diverse IoT datasets show that our method outperforms existing anomaly detection strategies and significantly reduces false positives.The 5DGWO-GAN-CNNAE exhibits superior performance in various metrics,including accuracy,recall,precision,root mean square error(RMSE),and convergence trend.The proposed 5DGWO-GAN-CNNAE achieved the lowest RMSE values across the NSL-KDD,UNSW-NB15,and IoT-23 datasets,with values of 0.24,1.10,and 0.09,respectively.Additionally,it attained the highest accuracy,ranging from 94%to 100%.These results suggest a promising direction for future IoT security frameworks,offering a scalable and efficient solution to safeguard against evolving cyber threats.展开更多
Introduction Deep learning(DL),as one of the most transformative technologies in artificial intelligence(AI),is undergoing a pivotal transition from laboratory research to industrial deployment.Advancing at an unprece...Introduction Deep learning(DL),as one of the most transformative technologies in artificial intelligence(AI),is undergoing a pivotal transition from laboratory research to industrial deployment.Advancing at an unprecedented pace,DL is transcending theoretical and application boundaries to penetrate emerging realworld scenarios such as industrial automation,urban management,and health monitoring,thereby driving a new wave of intelligent transformation.In August 2023,Goldman Sachs estimated that global AI investment will reach US$200 billion by 2025[1].However,the increasing complexity and dynamic nature of application scenarios expose critical challenges in traditional deep learning,including data heterogeneity,insufficient model generalization,computational resource constraints,and privacy-security trade-offs.The next generation of deep learning methodologies needs to achieve breakthroughs in multimodal fusion,lightweight design,interpretability enhancement,and cross-disciplinary collaborative optimization,in order to develop more efficient,robust,and practically valuable intelligent systems.展开更多
For control systems with unknown model parameters,this paper proposes a data-driven iterative learning method for fault estimation.First,input and output data from the system under fault-free conditions are collected....For control systems with unknown model parameters,this paper proposes a data-driven iterative learning method for fault estimation.First,input and output data from the system under fault-free conditions are collected.By applying orthogonal triangular decomposition and singular value decomposition,a data-driven realization of the system's kernel representation is derived,based on this representation,a residual generator is constructed.Then,the actuator fault signal is estimated online by analyzing the system's dynamic residual,and an iterative learning algorithm is introduced to continuously optimize the residual-based performance function,thereby enhancing estimation accuracy.The proposed method achieves actuator fault estimation without requiring knowledge of model parameters,eliminating the time-consuming system modeling process,and allowing operators to focus on system optimization and decision-making.Compared with existing fault estimation methods,the proposed method demonstrates superior transient performance,steady-state performance,and real-time capability,reduces the need for manual intervention and lowers operational complexity.Finally,experimental results on a mobile robot verify the effectiveness and advantages of the method.展开更多
Inferring phylogenetic trees from molecular sequences is a cornerstone of evolutionary biology.Many standard phylogenetic methods(such as maximum-likelihood[ML])rely on explicit models of sequence evolution and thus o...Inferring phylogenetic trees from molecular sequences is a cornerstone of evolutionary biology.Many standard phylogenetic methods(such as maximum-likelihood[ML])rely on explicit models of sequence evolution and thus often suffer from model misspecification or inadequacy.The on-rising deep learning(DL)techniques offer a powerful alternative.Deep learning employs multi-layered artificial neural networks to progressively transform input data into more abstract and complex representations.DL methods can autonomously uncover meaningful patterns from data,thereby bypassing potential biases introduced by predefined features(Franklin,2005;Murphy,2012).Recent efforts have aimed to apply deep neural networks(DNNs)to phylogenetics,with a growing number of applications in tree reconstruction(Suvorov et al.,2020;Zou et al.,2020;Nesterenko et al.,2022;Smith and Hahn,2023;Wang et al.,2023),substitution model selection(Abadi et al.,2020;Burgstaller-Muehlbacher et al.,2023),and diversification rate inference(Voznica et al.,2022;Lajaaiti et al.,2023;Lambert et al.,2023).In phylogenetic tree reconstruction,PhyDL(Zou et al.,2020)and Tree_learning(Suvorov et al.,2020)are two notable DNN-based programs designed to infer unrooted quartet trees directly from alignments of four amino acid(AA)and DNA sequences,respectively.展开更多
This study systematically reviews the applications of generative artificial intelligence(GAI)in breast cancer research,focusing on its role in diagnosis and therapeutic development.While GAI has gained significant att...This study systematically reviews the applications of generative artificial intelligence(GAI)in breast cancer research,focusing on its role in diagnosis and therapeutic development.While GAI has gained significant attention across various domains,its utility in breast cancer research has yet to be comprehensively reviewed.This study aims to fill that gap by synthesizing existing research into a unified document.A comprehensive search was conducted following Preferred Reporting Items for Systematic Reviews and Meta-Analyses(PRISMA)guidelines,resulting in the retrieval of 3827 articles,of which 31 were deemed eligible for analysis.The included studies were categorized based on key criteria,such as application types,geographical distribution,contributing organizations,leading journals,publishers,and temporal trends.Keyword co-occurrence mapping and subject profiling further highlighted the major research themes in this field.The findings reveal that GAI models have been applied to improve breast cancer diagnosis,treatment planning,and outcome predictions.Geographical and network analyses showed that most contributions come from a few leading institutions,with limited global collaboration.The review also identifies key challenges in implementing GAI in clinical practice,such as data availability,ethical concerns,and model validation.Despite these challenges,the study highlights GAI’s potential to enhance breast cancer research,particularly in generating synthetic data,improving diagnostic accuracy,and personalizing treatment approaches.This review serves as a valuable resource for researchers and stakeholders,providing insights into current research trends,major contributors,and collaborative networks in GAI-based breast cancer studies.By offering a holistic overview,it aims to support future research directions and encourage broader adoption of GAI technologies in healthcare.Additionally,the study emphasizes the importance of overcoming implementation barriers to fully realizeGAI’s potential in transforming breast cancer management.展开更多
Federated Learning(FL)has recently emerged as a promising paradigm that enables medical institutions to collaboratively train robust models without centralizing sensitive patient information.Data collected from differ...Federated Learning(FL)has recently emerged as a promising paradigm that enables medical institutions to collaboratively train robust models without centralizing sensitive patient information.Data collected from different institutions represent distinct source domains.Consequently,discrepancies in feature distributions can significantly hinder a model’s generalization to unseen domains.While domain generalization(DG)methods have been proposed to address this challenge,many may compromise data privacy in FL by requiring clients to transmit their local feature representations to the server.Furthermore,existing adversarial training methods,commonly used to align marginal feature distributions,fail to ensure the consistency of conditional distributions.This consistency is often critical for accurate predictions in unseen domains.To address these limitations,we propose GPAF,a privacy-preserving federated learning(FL)framework that mitigates both domain and label shifts in healthcare applications.GPAF aligns conditional distributions across clients in the latent space and restricts communication to model parameters.This design preserves class semantics,enhances privacy,and improves communication efficiency.At the server,a global generator learns a conditional feature distribution from clients’feedback.During local training,each client minimizes an adversarial loss to align its local conditional distribution with the global distribution,enabling the FL model to learn robust,domain-invariant representations across all source domains.To evaluate the effectiveness of our approach,experiments on a medical imaging benchmark demonstrate that GPAF outperforms four FL baselines,achieving up to 17%higher classification accuracy and 25%faster convergence in non-IID scenarios.These results highlight GPAF’s capability to generalize across domains while maintaining strict privacy,offering a robust solution for decentralized healthcare challenges.展开更多
This study addresses the pressing challenge of generating realistic strong ground motion data for simulating earthquakes,a crucial component in pre-earthquake risk assessments and post-earthquake disaster evaluations,...This study addresses the pressing challenge of generating realistic strong ground motion data for simulating earthquakes,a crucial component in pre-earthquake risk assessments and post-earthquake disaster evaluations,particularly suited for regions with limited seismic data.Herein,we report a generative adversarial network(GAN)framework capable of simulating strong ground motions under various environmental conditions using only a small set of real earthquake records.The constructed GAN model generates ground motions based on continuous physical variables such as source distance,site conditions,and magnitude,effectively capturing the complexity and diversity of ground motions under different scenarios.This capability allows the proposed model to approximate real seismic data,making it applicable to a wide range of engineering purposes.Using the Shandong Pingyuan earthquake as an example,a specialized dataset was constructed based on regional real ground motion records.The response spectrum at target locations was obtained through inverse distance-weighted interpolation of actual response spectra,followed by continuous wavelet transform to derive the ground motion time histories at these locations.Through iterative parameter adjustments,the constructed GAN model learned the probability distribution of strong-motion data for this event.The trained model generated three-component ground-motion time histories with clear P-wave and S-wave characteristics,accurately reflecting the non-stationary nature of seismic records.Statistical comparisons between synthetic and real response spectra,waveform envelopes,and peak ground acceleration show a high degree of similarity,underscoring the effectiveness of the model in replicating both the statistical and physical characteristics of real ground motions.These findings validate the feasibility of GANs for generating realistic earthquake data in data-scarce regions,providing a reliable approach for enriching regional ground motion databases.Additionally,the results suggest that GAN-based networks are a powerful tool for building predictive models in seismic hazard analysis.展开更多
基金supported by the Chinese Academy of Science"Light of West China"Program(2022-XBQNXZ-015)the National Natural Science Foundation of China(11903071)the Operation,Maintenance and Upgrading Fund for Astronomical Telescopes and Facility Instruments,budgeted from the Ministry of Finance of China and administered by the Chinese Academy of Sciences。
文摘This paper addresses the performance degradation issue in a fast radio burst search pipeline based on deep learning.This issue is caused by the class imbalance of the radio frequency interference samples in the training dataset,and one solution is applied to improve the distribution of the training data by augmenting minority class samples using a deep convolutional generative adversarial network.Experi.mental results demonstrate that retraining the deep learning model with the newly generated dataset leads to a new fast radio burst classifier,which effectively reduces false positives caused by periodic wide-band impulsive radio frequency interference,thereby enhancing the performance of the search pipeline.
基金supported in part by the National Key R&D Program of China under Grant 2024YFE0200700in part by the National Natural Science Foundation of China under Grant 62201504.
文摘Network architectures assisted by Generative Artificial Intelligence(GAI)are envisioned as foundational elements of sixth-generation(6G)communication system.To deliver ubiquitous intelligent services and meet diverse service requirements,6G network architecture should offer personalized services to various mobile devices.Federated learning(FL)with personalized local training,as a privacypreserving machine learning(ML)approach,can be applied to address these challenges.In this paper,we propose a meta-learning-based personalized FL(PFL)method that improves both communication and computation efficiency by utilizing over-the-air computations.Its“pretraining-and-fine-tuning”principle makes it particularly suitable for enabling edge nodes to access personalized GAI services while preserving local privacy.Experiment results demonstrate the outperformance and efficacy of the proposed algorithm,and notably indicate enhanced communication efficiency without compromising accuracy.
基金supported by State Grid Corporation Technology Project(No.522437250003).
文摘Hydrogen energy is a crucial support for China’s low-carbon energy transition.With the large-scale integration of renewable energy,the combination of hydrogen and integrated energy systems has become one of the most promising directions of development.This paper proposes an optimized schedulingmodel for a hydrogen-coupled electro-heat-gas integrated energy system(HCEHG-IES)using generative adversarial imitation learning(GAIL).The model aims to enhance renewable-energy absorption,reduce carbon emissions,and improve grid-regulation flexibility.First,the optimal scheduling problem of HCEHG-IES under uncertainty is modeled as a Markov decision process(MDP).To overcome the limitations of conventional deep reinforcement learning algorithms—including long optimization time,slow convergence,and subjective reward design—this study augments the PPO algorithm by incorporating a discriminator network and expert data.The newly developed algorithm,termed GAIL,enables the agent to perform imitation learning from expert data.Based on this model,dynamic scheduling decisions are made in continuous state and action spaces,generating optimal energy-allocation and management schemes.Simulation results indicate that,compared with traditional reinforcement-learning algorithms,the proposed algorithmoffers better economic performance.Guided by expert data,the agent avoids blind optimization,shortens the offline training time,and improves convergence performance.In the online phase,the algorithm enables flexible energy utilization,thereby promoting renewable-energy absorption and reducing carbon emissions.
文摘This study explores a novel educational model of generative AI-empowered interdisciplinary project-based learning(PBL).By analyzing the current applications of generative AI technology in information technology curricula,it elucidates its advantages and operational mechanisms in interdisciplinary PBL.Combining case studies and empirical research,the investigation proposes implementation pathways and strategies for the generative AI-enhanced interdisciplinary PBL model,detailing specific applications across three phases:project preparation,implementation,and evaluation.The research demonstrates that generative AI-enabled interdisciplinary project-based learning can effectively enhance students’learning motivation,interdisciplinary thinking capabilities,and innovative competencies,providing new conceptual frameworks and practical approaches for educational model innovation.
基金This work is supported by the National Natural Science Foundation of China(Nos.61771154,61603239,61772454,6171101570).
文摘Deep Learning(DL)is such a powerful tool that we have seen tremendous success in areas such as Computer Vision,Speech Recognition,and Natural Language Processing.Since Automated Modulation Classification(AMC)is an important part in Cognitive Radio Networks,we try to explore its potential in solving signal modulation recognition problem.It cannot be overlooked that DL model is a complex model,thus making them prone to over-fitting.DL model requires many training data to combat with over-fitting,but adding high quality labels to training data manually is not always cheap and accessible,especially in real-time system,which may counter unprecedented data in dataset.Semi-supervised Learning is a way to exploit unlabeled data effectively to reduce over-fitting in DL.In this paper,we extend Generative Adversarial Networks(GANs)to the semi-supervised learning will show it is a method can be used to create a more dataefficient classifier.
基金This research is funded by the Centre for Advanced Modeling and Geospatial Information Systems(CAMGIS),Faculty of Engineering and Information Technology,the University of Technology Sydney,Australia.
文摘In recent years,landslide susceptibility mapping has substantially improved with advances in machine learning.However,there are still challenges remain in landslide mapping due to the availability of limited inventory data.In this paper,a novel method that improves the performance of machine learning techniques is presented.The proposed method creates synthetic inventory data using Generative Adversarial Networks(GANs)for improving the prediction of landslides.In this research,landslide inventory data of 156 landslide locations were identified in Cameron Highlands,Malaysia,taken from previous projects the authors worked on.Elevation,slope,aspect,plan curvature,profile curvature,total curvature,lithology,land use and land cover(LULC),distance to the road,distance to the river,stream power index(SPI),sediment transport index(STI),terrain roughness index(TRI),topographic wetness index(TWI)and vegetation density are geo-environmental factors considered in this study based on suggestions from previous works on Cameron Highlands.To show the capability of GANs in improving landslide prediction models,this study tests the proposed GAN model with benchmark models namely Artificial Neural Network(ANN),Support Vector Machine(SVM),Decision Trees(DT),Random Forest(RF)and Bagging ensemble models with ANN and SVM models.These models were validated using the area under the receiver operating characteristic curve(AUROC).The DT,RF,SVM,ANN and Bagging ensemble could achieve the AUROC values of(0.90,0.94,0.86,0.69 and 0.82)for the training;and the AUROC of(0.76,0.81,0.85,0.72 and 0.75)for the test,subsequently.When using additional samples,the same models achieved the AUROC values of(0.92,0.94,0.88,0.75 and 0.84)for the training and(0.78,0.82,0.82,0.78 and 0.80)for the test,respectively.Using the additional samples improved the test accuracy of all the models except SVM.As a result,in data-scarce environments,this research showed that utilizing GANs to generate supplementary samples is promising because it can improve the predictive capability of common landslide prediction models.
基金the National Natural Science Foundation of China(Nos.61771380,U19B2015,U1730109).
文摘Generative adversarial network(GAN)has achieved great success in many fields such as computer vision,speech processing,and natural language processing,because of its powerful capabilities for generating realistic samples.In this paper,we introduce GAN into the field of electromagnetic signal classification(ESC).ESC plays an important role in both military and civilian domains.However,in many specific scenarios,we can’t obtain enough labeled data,which cause failure of deep learning methods because they are easy to fall into over-fitting.Fortunately,semi-supervised learning(SSL)can leverage the large amount of unlabeled data to enhance the classification performance of classifiers,especially in scenarios with limited amount of labeled data.We present an SSL framework by incorporating GAN,which can directly process the raw in-phase and quadrature(IQ)signal data.According to the characteristics of the electromagnetic signal,we propose a weighted loss function,leading to an effective classifier to realize the end-to-end classification of the electromagnetic signal.We validate the proposed method on both public RML2016.04c dataset and real-world Aircraft Communications Addressing and Reporting System(ACARS)signal dataset.Extensive experimental results show that the proposed framework obtains a significant increase in classification accuracy compared with the state-of-the-art studies.
基金The publication of this article is funded by the Qatar National Library.
文摘Malaria is a lethal disease responsible for thousands of deaths worldwide every year.Manual methods of malaria diagnosis are timeconsuming that require a great deal of human expertise and efforts.Computerbased automated diagnosis of diseases is progressively becoming popular.Although deep learning models show high performance in the medical field,it demands a large volume of data for training which is hard to acquire for medical problems.Similarly,labeling of medical images can be done with the help of medical experts only.Several recent studies have utilized deep learning models to develop efficient malaria diagnostic system,which showed promising results.However,the most common problem with these models is that they need a large amount of data for training.This paper presents a computer-aided malaria diagnosis system that combines a semi-supervised generative adversarial network and transfer learning.The proposed model is trained in a semi-supervised manner and requires less training data than conventional deep learning models.Performance of the proposed model is evaluated on a publicly available dataset of blood smear images(with malariainfected and normal class)and achieved a classification accuracy of 96.6%.
基金National Key Research and Development Program of China,Grant/Award Numbers:2021YFB2501301,2019YFB1600704The Science and Technology Development Fund,Grant/Award Numbers:0068/2020/AGJ,SKL‐IOTSC(UM)‐2021‐2023GDST,Grant/Award Numbers:2020B1212030003,MYRG2022‐00192‐FST。
文摘Robot calligraphy visually reflects the motion capability of robotic manipulators.While traditional researches mainly focus on image generation and the writing of simple calligraphic strokes or characters,this article presents a generative adversarial network(GAN)-based motion learning method for robotic calligraphy synthesis(Gan2CS)that can enhance the efficiency in writing complex calligraphy words and reproducing classic calligraphy works.The key technologies in the proposed approach include:(1)adopting the GAN to learn the motion parameters from the robot writing operation;(2)converting the learnt motion data into the style font and realising the transition from static calligraphy images to dynamic writing demonstration;(3)reproducing high-precision calligraphy works by synthesising the writing motion data hierarchically.In this study,the motion trajectories of sample calligraphy images are firstly extracted and converted into the robot module.The robot performs the writing with motion planning,and the writing motion parameters of calligraphy strokes are learnt with GANs.Then the motion data of basic strokes is synthesised based on the hierarchical process of‘stroke-radicalpart-character’.And the robot re-writes the synthesised characters whose similarity with the original calligraphy characters is evaluated.Regular calligraphy characters have been tested in the experiments for method validation and the results validated that the robot can actualise the robotic calligraphy synthesis of writing motion data with GAN.
文摘With the continuous scaling of integrated circuit technologies,design for manufacturability(DFM)is becoming more critical,yet more challenging.Alongside,recent advances in machine learning have provided a new computing paradigm with promising applications in VLSI manufacturability.In particular,generative learning-regarded among the most interesting ideas in present-day machine learning-has demonstrated impressive capabilities in a wide range of applications.This paper surveys recent results of using generative learning in VLSI manufacturing modeling and optimization.Specifically,we examine the unique features of generative learning that have been leveraged to improve DFM efficiency in an unprecedented way;hence,paving the way to a new data-driven DFM approach.The state-of-the-art methods are presented,and challenges/opportunities are discussed.
基金supported by the National Natural Science Foundation of China(Grant Nos.22225801,W2441009,22408228)。
文摘As energy demands continue to rise in modern society,the development of high-performance lithium-ion batteries(LIBs)has become crucial.However,traditional research methods of material science face challenges such as lengthy timelines and complex processes.In recent years,the integration of machine learning(ML)in LIB materials,including electrolytes,solid-state electrolytes,and electrodes,has yielded remarkable achievements.This comprehensive review explores the latest applications of ML in predicting LIB material performance,covering the core principles and recent advancements in three key inverse material design strategies:high-throughput virtual screening,global optimization,and generative models.These strategies have played a pivotal role in fostering LIB material innovations.Meanwhile,the paper briefly discusses the challenges associated with applying ML to materials research and offers insights and directions for future research.
文摘Learning is no longer regarded as knowledge acquisition but as knowledge construction;therefore,learner is not a passive recipient but an active constructor.In generative and elaborative learning process,the learner’s existing knowledge is considered to be modified while it is used to construct a meaning from the text.Therefore,the focus in learning is on generating relations,rather than on storing information.This paper aims to draw the outlines of the main ideas of generative learning and their related concepts and theoretical significance.
基金supported by the National Natural Science Foundation of China(Grant Nos.62422509 and 62405188)the Shanghai Natural Science Foundation(Grant No.23ZR1443700)+3 种基金the Shuguang Program of Shanghai Education Development Foundation and Shanghai Municipal Education Commission(Grant No.23SG41)the Young Elite Scientist Sponsorship Program by CAST(Grant No.20220042)the Science and Technology Commission of Shanghai Municipality(Grant No.21DZ1100500)the Shanghai Municipal Science and Technology Major Project,and the Shanghai Frontiers Science Center Program(2021-2025 No.20).
文摘Efficiently tracking and imaging interested moving targets is crucial across various applications,from autonomous systems to surveillance.However,persistent challenges remain in various fields,including environmental intricacies,limitations in perceptual technologies,and privacy considerations.We present a teacher-student learning model,the generative adversarial network(GAN)-guided diffractive neural network(DNN),which performs visual tracking and imaging of the interested moving target.The GAN,as a teacher model,empowers efficient acquisition of the skill to differentiate the specific target of interest in the domains of visual tracking and imaging.The DNN-based student model learns to master the skill to differentiate the interested target from the GAN.The process of obtaining a GAN-guided DNN starts with capturing moving objects effectively using an event camera with high temporal resolution and low latency.Then,the generative power of GAN is utilized to generate data with position-tracking capability for the interested moving target,subsequently serving as labels to the training of the DNN.The DNN learns to image the target during training while retaining the target’s positional information.Our experimental demonstration highlights the efficacy of the GAN-guided DNN in visual tracking and imaging of the interested moving target.We expect the GAN-guided DNN can significantly enhance autonomous systems and surveillance.
基金described in this paper has been developed with in the project PRESECREL(PID2021-124502OB-C43)。
文摘The Internet of Things(IoT)is integral to modern infrastructure,enabling connectivity among a wide range of devices from home automation to industrial control systems.With the exponential increase in data generated by these interconnected devices,robust anomaly detection mechanisms are essential.Anomaly detection in this dynamic environment necessitates methods that can accurately distinguish between normal and anomalous behavior by learning intricate patterns.This paper presents a novel approach utilizing generative adversarial networks(GANs)for anomaly detection in IoT systems.However,optimizing GANs involves tuning hyper-parameters such as learning rate,batch size,and optimization algorithms,which can be challenging due to the non-convex nature of GAN loss functions.To address this,we propose a five-dimensional Gray wolf optimizer(5DGWO)to optimize GAN hyper-parameters.The 5DGWO introduces two new types of wolves:gamma(γ)for improved exploitation and convergence,and theta(θ)for enhanced exploration and escaping local minima.The proposed system framework comprises four key stages:1)preprocessing,2)generative model training,3)autoencoder(AE)training,and 4)predictive model training.The generative models are utilized to assist the AE training,and the final predictive models(including convolutional neural network(CNN),deep belief network(DBN),recurrent neural network(RNN),random forest(RF),and extreme gradient boosting(XGBoost))are trained using the generated data and AE-encoded features.We evaluated the system on three benchmark datasets:NSL-KDD,UNSW-NB15,and IoT-23.Experiments conducted on diverse IoT datasets show that our method outperforms existing anomaly detection strategies and significantly reduces false positives.The 5DGWO-GAN-CNNAE exhibits superior performance in various metrics,including accuracy,recall,precision,root mean square error(RMSE),and convergence trend.The proposed 5DGWO-GAN-CNNAE achieved the lowest RMSE values across the NSL-KDD,UNSW-NB15,and IoT-23 datasets,with values of 0.24,1.10,and 0.09,respectively.Additionally,it attained the highest accuracy,ranging from 94%to 100%.These results suggest a promising direction for future IoT security frameworks,offering a scalable and efficient solution to safeguard against evolving cyber threats.
基金supported in part by Guangdong Basic and Applied Basic Research Foundation under Grant 2024A1515012485in part by Shenzhen Fundamental Research Program under Grant JCYJ20220810112354002+4 种基金in part by Shenzhen Science and Technology Program under Grant KJZD20230923114111021in part by the Fund for Academic Innovation Teams and Research Platform of South-Central Minzu University under Grant XTZ24003 and Grant PTZ24001in part by the Knowledge Innovation Program of Wuhan-Basic Research through Project 2023010201010151in part by the Research Start-up Funds of South-Central Minzu University under Grant YZZ18006in part by the Spring Sunshine Program of Ministry of Education of the People’s Republic of China under Grant HZKY20220331.
文摘Introduction Deep learning(DL),as one of the most transformative technologies in artificial intelligence(AI),is undergoing a pivotal transition from laboratory research to industrial deployment.Advancing at an unprecedented pace,DL is transcending theoretical and application boundaries to penetrate emerging realworld scenarios such as industrial automation,urban management,and health monitoring,thereby driving a new wave of intelligent transformation.In August 2023,Goldman Sachs estimated that global AI investment will reach US$200 billion by 2025[1].However,the increasing complexity and dynamic nature of application scenarios expose critical challenges in traditional deep learning,including data heterogeneity,insufficient model generalization,computational resource constraints,and privacy-security trade-offs.The next generation of deep learning methodologies needs to achieve breakthroughs in multimodal fusion,lightweight design,interpretability enhancement,and cross-disciplinary collaborative optimization,in order to develop more efficient,robust,and practically valuable intelligent systems.
基金Supported by Shandong Provincial Taishan Scholar Program(Grant No.tsqn202312133)Shandong Provincial Natural Science Foundation(Grant Nos.ZR2022YQ61,ZR2023ZD32)+1 种基金Shandong Provincial Natural Science Foundation(Grant No.ZR2023ZD32)National Natural Science Foundation of China(Grant Nos.61772551 and 62111530052)。
文摘For control systems with unknown model parameters,this paper proposes a data-driven iterative learning method for fault estimation.First,input and output data from the system under fault-free conditions are collected.By applying orthogonal triangular decomposition and singular value decomposition,a data-driven realization of the system's kernel representation is derived,based on this representation,a residual generator is constructed.Then,the actuator fault signal is estimated online by analyzing the system's dynamic residual,and an iterative learning algorithm is introduced to continuously optimize the residual-based performance function,thereby enhancing estimation accuracy.The proposed method achieves actuator fault estimation without requiring knowledge of model parameters,eliminating the time-consuming system modeling process,and allowing operators to focus on system optimization and decision-making.Compared with existing fault estimation methods,the proposed method demonstrates superior transient performance,steady-state performance,and real-time capability,reduces the need for manual intervention and lowers operational complexity.Finally,experimental results on a mobile robot verify the effectiveness and advantages of the method.
基金supported by the National Key R&D Program of China(2022YFD1401600)the National Science Foundation for Distinguished Young Scholars of Zhejang Province,China(LR23C140001)supported by the Key Area Research and Development Program of Guangdong Province,China(2018B020205003 and 2020B0202090001).
文摘Inferring phylogenetic trees from molecular sequences is a cornerstone of evolutionary biology.Many standard phylogenetic methods(such as maximum-likelihood[ML])rely on explicit models of sequence evolution and thus often suffer from model misspecification or inadequacy.The on-rising deep learning(DL)techniques offer a powerful alternative.Deep learning employs multi-layered artificial neural networks to progressively transform input data into more abstract and complex representations.DL methods can autonomously uncover meaningful patterns from data,thereby bypassing potential biases introduced by predefined features(Franklin,2005;Murphy,2012).Recent efforts have aimed to apply deep neural networks(DNNs)to phylogenetics,with a growing number of applications in tree reconstruction(Suvorov et al.,2020;Zou et al.,2020;Nesterenko et al.,2022;Smith and Hahn,2023;Wang et al.,2023),substitution model selection(Abadi et al.,2020;Burgstaller-Muehlbacher et al.,2023),and diversification rate inference(Voznica et al.,2022;Lajaaiti et al.,2023;Lambert et al.,2023).In phylogenetic tree reconstruction,PhyDL(Zou et al.,2020)and Tree_learning(Suvorov et al.,2020)are two notable DNN-based programs designed to infer unrooted quartet trees directly from alignments of four amino acid(AA)and DNA sequences,respectively.
基金financial support from the Fundamental Research Grant Scheme(FRGS)under grant number:FRGS/1/2024/ICT02/TARUMT/02/1from the Ministry of Higher Education Malaysiafunded in part by the internal grant from the Tunku Abdul Rahman University of Management and Technology(TAR UMT)with grant number:UC/I/G2024-00129.
文摘This study systematically reviews the applications of generative artificial intelligence(GAI)in breast cancer research,focusing on its role in diagnosis and therapeutic development.While GAI has gained significant attention across various domains,its utility in breast cancer research has yet to be comprehensively reviewed.This study aims to fill that gap by synthesizing existing research into a unified document.A comprehensive search was conducted following Preferred Reporting Items for Systematic Reviews and Meta-Analyses(PRISMA)guidelines,resulting in the retrieval of 3827 articles,of which 31 were deemed eligible for analysis.The included studies were categorized based on key criteria,such as application types,geographical distribution,contributing organizations,leading journals,publishers,and temporal trends.Keyword co-occurrence mapping and subject profiling further highlighted the major research themes in this field.The findings reveal that GAI models have been applied to improve breast cancer diagnosis,treatment planning,and outcome predictions.Geographical and network analyses showed that most contributions come from a few leading institutions,with limited global collaboration.The review also identifies key challenges in implementing GAI in clinical practice,such as data availability,ethical concerns,and model validation.Despite these challenges,the study highlights GAI’s potential to enhance breast cancer research,particularly in generating synthetic data,improving diagnostic accuracy,and personalizing treatment approaches.This review serves as a valuable resource for researchers and stakeholders,providing insights into current research trends,major contributors,and collaborative networks in GAI-based breast cancer studies.By offering a holistic overview,it aims to support future research directions and encourage broader adoption of GAI technologies in healthcare.Additionally,the study emphasizes the importance of overcoming implementation barriers to fully realizeGAI’s potential in transforming breast cancer management.
文摘Federated Learning(FL)has recently emerged as a promising paradigm that enables medical institutions to collaboratively train robust models without centralizing sensitive patient information.Data collected from different institutions represent distinct source domains.Consequently,discrepancies in feature distributions can significantly hinder a model’s generalization to unseen domains.While domain generalization(DG)methods have been proposed to address this challenge,many may compromise data privacy in FL by requiring clients to transmit their local feature representations to the server.Furthermore,existing adversarial training methods,commonly used to align marginal feature distributions,fail to ensure the consistency of conditional distributions.This consistency is often critical for accurate predictions in unseen domains.To address these limitations,we propose GPAF,a privacy-preserving federated learning(FL)framework that mitigates both domain and label shifts in healthcare applications.GPAF aligns conditional distributions across clients in the latent space and restricts communication to model parameters.This design preserves class semantics,enhances privacy,and improves communication efficiency.At the server,a global generator learns a conditional feature distribution from clients’feedback.During local training,each client minimizes an adversarial loss to align its local conditional distribution with the global distribution,enabling the FL model to learn robust,domain-invariant representations across all source domains.To evaluate the effectiveness of our approach,experiments on a medical imaging benchmark demonstrate that GPAF outperforms four FL baselines,achieving up to 17%higher classification accuracy and 25%faster convergence in non-IID scenarios.These results highlight GPAF’s capability to generalize across domains while maintaining strict privacy,offering a robust solution for decentralized healthcare challenges.
基金Funded by the National Key Research and Development Program(2022YFC3003502).
文摘This study addresses the pressing challenge of generating realistic strong ground motion data for simulating earthquakes,a crucial component in pre-earthquake risk assessments and post-earthquake disaster evaluations,particularly suited for regions with limited seismic data.Herein,we report a generative adversarial network(GAN)framework capable of simulating strong ground motions under various environmental conditions using only a small set of real earthquake records.The constructed GAN model generates ground motions based on continuous physical variables such as source distance,site conditions,and magnitude,effectively capturing the complexity and diversity of ground motions under different scenarios.This capability allows the proposed model to approximate real seismic data,making it applicable to a wide range of engineering purposes.Using the Shandong Pingyuan earthquake as an example,a specialized dataset was constructed based on regional real ground motion records.The response spectrum at target locations was obtained through inverse distance-weighted interpolation of actual response spectra,followed by continuous wavelet transform to derive the ground motion time histories at these locations.Through iterative parameter adjustments,the constructed GAN model learned the probability distribution of strong-motion data for this event.The trained model generated three-component ground-motion time histories with clear P-wave and S-wave characteristics,accurately reflecting the non-stationary nature of seismic records.Statistical comparisons between synthetic and real response spectra,waveform envelopes,and peak ground acceleration show a high degree of similarity,underscoring the effectiveness of the model in replicating both the statistical and physical characteristics of real ground motions.These findings validate the feasibility of GANs for generating realistic earthquake data in data-scarce regions,providing a reliable approach for enriching regional ground motion databases.Additionally,the results suggest that GAN-based networks are a powerful tool for building predictive models in seismic hazard analysis.