期刊文献+
共找到276篇文章
< 1 2 14 >
每页显示 20 50 100
Gradient Optimizer Algorithm with Hybrid Deep Learning Based Failure Detection and Classification in the Industrial Environment
1
作者 Mohamed Zarouan Ibrahim M.Mehedi +1 位作者 Shaikh Abdul Latif Md.Masud Rana 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第2期1341-1364,共24页
Failure detection is an essential task in industrial systems for preventing costly downtime and ensuring the seamlessoperation of the system. Current industrial processes are getting smarter with the emergence of Indu... Failure detection is an essential task in industrial systems for preventing costly downtime and ensuring the seamlessoperation of the system. Current industrial processes are getting smarter with the emergence of Industry 4.0.Specifically, various modernized industrial processes have been equipped with quite a few sensors to collectprocess-based data to find faults arising or prevailing in processes along with monitoring the status of processes.Fault diagnosis of rotating machines serves a main role in the engineering field and industrial production. Dueto the disadvantages of existing fault, diagnosis approaches, which greatly depend on professional experienceand human knowledge, intellectual fault diagnosis based on deep learning (DL) has attracted the researcher’sinterest. DL reaches the desired fault classification and automatic feature learning. Therefore, this article designs a Gradient Optimizer Algorithm with Hybrid Deep Learning-based Failure Detection and Classification (GOAHDLFDC)in the industrial environment. The presented GOAHDL-FDC technique initially applies continuous wavelettransform (CWT) for preprocessing the actual vibrational signals of the rotating machinery. Next, the residualnetwork (ResNet18) model was exploited for the extraction of features from the vibration signals which are thenfed into theHDLmodel for automated fault detection. Finally, theGOA-based hyperparameter tuning is performedtoadjust the parameter valuesof theHDLmodel accurately.The experimental result analysis of the GOAHDL-FD Calgorithm takes place using a series of simulations and the experimentation outcomes highlight the better resultsof the GOAHDL-FDC technique under different aspects. 展开更多
关键词 Fault detection Industry 4.0 gradient optimizer algorithm deep learning rotating machineries artificial intelligence
在线阅读 下载PDF
Efficient and High-quality Recommendations via Momentum-incorporated Parallel Stochastic Gradient Descent-Based Learning 被引量:7
2
作者 Xin Luo Wen Qin +2 位作者 Ani Dong Khaled Sedraoui MengChu Zhou 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2021年第2期402-411,共10页
A recommender system(RS)relying on latent factor analysis usually adopts stochastic gradient descent(SGD)as its learning algorithm.However,owing to its serial mechanism,an SGD algorithm suffers from low efficiency and... A recommender system(RS)relying on latent factor analysis usually adopts stochastic gradient descent(SGD)as its learning algorithm.However,owing to its serial mechanism,an SGD algorithm suffers from low efficiency and scalability when handling large-scale industrial problems.Aiming at addressing this issue,this study proposes a momentum-incorporated parallel stochastic gradient descent(MPSGD)algorithm,whose main idea is two-fold:a)implementing parallelization via a novel datasplitting strategy,and b)accelerating convergence rate by integrating momentum effects into its training process.With it,an MPSGD-based latent factor(MLF)model is achieved,which is capable of performing efficient and high-quality recommendations.Experimental results on four high-dimensional and sparse matrices generated by industrial RS indicate that owing to an MPSGD algorithm,an MLF model outperforms the existing state-of-the-art ones in both computational efficiency and scalability. 展开更多
关键词 Big data industrial application industrial data latent factor analysis machine learning parallel algorithm recommender system(RS) stochastic gradient descent(SGD)
在线阅读 下载PDF
Data-Driven Learning Control Algorithms for Unachievable Tracking Problems 被引量:1
3
作者 Zeyi Zhang Hao Jiang +1 位作者 Dong Shen Samer S.Saab 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第1期205-218,共14页
For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to in... For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to investigate solutions using the Ptype learning control scheme. Initially, we demonstrate the necessity of gradient information for achieving the best approximation.Subsequently, we propose an input-output-driven learning gain design to handle the imprecise gradients of a class of uncertain systems. However, it is discovered that the desired performance may not be attainable when faced with incomplete information.To address this issue, an extended iterative learning control scheme is introduced. In this scheme, the tracking errors are modified through output data sampling, which incorporates lowmemory footprints and offers flexibility in learning gain design.The input sequence is shown to converge towards the desired input, resulting in an output that is closest to the given reference in the least square sense. Numerical simulations are provided to validate the theoretical findings. 展开更多
关键词 Data-driven algorithms incomplete information iterative learning control gradient information unachievable problems
在线阅读 下载PDF
Memetic algorithms-based neural network learning for basic oxygen furnace endpoint prediction
4
作者 Peng CHEN Yong-zai LU 《Journal of Zhejiang University-Science A(Applied Physics & Engineering)》 SCIE EI CAS CSCD 2010年第11期841-848,共8页
Based on the critical position of the endpoint quality prediction for basic oxygen furnaces (BOFs) in steelmaking, and the latest results in computational intelligence (C1), this paper deals with the development ... Based on the critical position of the endpoint quality prediction for basic oxygen furnaces (BOFs) in steelmaking, and the latest results in computational intelligence (C1), this paper deals with the development of a novel memetic algorithm (MA) for neural network (NN) lcarnmg. Included in this is the integration of extremal optimization (EO) and Levenberg-Marquardt (LM) pradicnt search, and its application in BOF endpoint quality prediction. The fundamental analysis reveals that the proposed EO-LM algorithm may provide superior performance in generalization, computation efficiency, and avoid local minima, compared to traditional NN learning methods. Experimental results with production-scale BOF data show that the proposed method can effectively improve the NN model for BOF endpoint quality prediction. 展开更多
关键词 Memetic algorithm (MA) Neural network (NN) learning Back propagation (BP) Extremal optimization (EO) gevenberg-Marquardt (LM) gradient search Basic oxygen furnace (BOF)
原文传递
Chimp Optimization Algorithm Based Feature Selection with Machine Learning for Medical Data Classification
5
作者 Firas Abedi Hayder M.A.Ghanimi +6 位作者 Abeer D.Algarni Naglaa F.Soliman Walid El-Shafai Ali Hashim Abbas Zahraa H.Kareem Hussein Muhi Hariz Ahmed Alkhayyat 《Computer Systems Science & Engineering》 SCIE EI 2023年第12期2791-2814,共24页
Datamining plays a crucial role in extractingmeaningful knowledge fromlarge-scale data repositories,such as data warehouses and databases.Association rule mining,a fundamental process in data mining,involves discoveri... Datamining plays a crucial role in extractingmeaningful knowledge fromlarge-scale data repositories,such as data warehouses and databases.Association rule mining,a fundamental process in data mining,involves discovering correlations,patterns,and causal structures within datasets.In the healthcare domain,association rules offer valuable opportunities for building knowledge bases,enabling intelligent diagnoses,and extracting invaluable information rapidly.This paper presents a novel approach called the Machine Learning based Association Rule Mining and Classification for Healthcare Data Management System(MLARMC-HDMS).The MLARMC-HDMS technique integrates classification and association rule mining(ARM)processes.Initially,the chimp optimization algorithm-based feature selection(COAFS)technique is employed within MLARMC-HDMS to select relevant attributes.Inspired by the foraging behavior of chimpanzees,the COA algorithm mimics their search strategy for food.Subsequently,the classification process utilizes stochastic gradient descent with a multilayer perceptron(SGD-MLP)model,while the Apriori algorithm determines attribute relationships.We propose a COA-based feature selection approach for medical data classification using machine learning techniques.This approach involves selecting pertinent features from medical datasets through COA and training machine learning models using the reduced feature set.We evaluate the performance of our approach on various medical datasets employing diverse machine learning classifiers.Experimental results demonstrate that our proposed approach surpasses alternative feature selection methods,achieving higher accuracy and precision rates in medical data classification tasks.The study showcases the effectiveness and efficiency of the COA-based feature selection approach in identifying relevant features,thereby enhancing the diagnosis and treatment of various diseases.To provide further validation,we conduct detailed experiments on a benchmark medical dataset,revealing the superiority of the MLARMCHDMS model over other methods,with a maximum accuracy of 99.75%.Therefore,this research contributes to the advancement of feature selection techniques in medical data classification and highlights the potential for improving healthcare outcomes through accurate and efficient data analysis.The presented MLARMC-HDMS framework and COA-based feature selection approach offer valuable insights for researchers and practitioners working in the field of healthcare data mining and machine learning. 展开更多
关键词 Association rule mining data classification healthcare data machine learning parameter tuning data mining feature selection MLARMC-HDMS COA stochastic gradient descent Apriori algorithm
在线阅读 下载PDF
基于生成模型的Q-learning二分类算法 被引量:1
6
作者 尚志刚 徐若灏 +2 位作者 乔康加 杨莉芳 李蒙蒙 《计算机应用研究》 CSCD 北大核心 2020年第11期3326-3329,3333,共5页
对于二分类问题,基于判别模型的分类器一般都是寻找一条最优判决边界,容易受到数据波动的影响。针对该问题提出一种基于生成模型的Q-learning二分类算法(BGQ-learning),将状态和动作分开编码,得到对应各类的判决函数,增加了决策空间的... 对于二分类问题,基于判别模型的分类器一般都是寻找一条最优判决边界,容易受到数据波动的影响。针对该问题提出一种基于生成模型的Q-learning二分类算法(BGQ-learning),将状态和动作分开编码,得到对应各类的判决函数,增加了决策空间的灵活性,同时在求解参数时,采用最小二乘时序差分(TD)算法和半梯度下降法的组合优化方法,加速了参数的收敛速度。设计实验对比了BGQ-learning算法与三种经典分类器以及一种新颖的分类器的分类性能,在UCI数据库七个数据集上的测试结果表明,该算法有着优良的稳定性以及良好的分类精确度。 展开更多
关键词 Q-learning 生成模型 二分类 最小二乘时序差分算法 半梯度下降法
在线阅读 下载PDF
Machine learning-based prediction of soil compression modulus with application of ID settlement 被引量:16
7
作者 Dong-ming ZHANG Jin-zhang ZHANG +2 位作者 Hong-wei HUANG Chong-chong QI Chen-yu CHANG 《Journal of Zhejiang University-Science A(Applied Physics & Engineering)》 SCIE EI CAS CSCD 2020年第6期430-444,共15页
The compression modulus(Es)is one of the most significant soil parameters that affects the compressive deformation of geotechnical systems,such as foundations.However,it is difficult and sometime costly to obtain this... The compression modulus(Es)is one of the most significant soil parameters that affects the compressive deformation of geotechnical systems,such as foundations.However,it is difficult and sometime costly to obtain this parameter in engineering practice.In this study,we aimed to develop a non-parametric ensemble artificial intelligence(AI)approach to calculate the Es of soft clay in contrast to the traditional regression models proposed in previous studies.A gradient boosted regression tree(GBRT)algorithm was used to discern the non-linear pattern between input variables and the target response,while a genetic algorithm(GA)was adopted for tuning the GBRT model's hyper-parameters.The model was tested through 10-fold cross validation.A dataset of 221 samples from 65 engineering survey reports from Shanghai infrastructure projects was constructed to evaluate the accuracy of the new model5 s predictions.The mean squared error and correlation coefficient of the optimum GBRT model applied to the testing set were 0.13 and 0.91,respectively,indicating that the proposed machine learning(ML)model has great potential to improve the prediction of Es for soft clay.A comparison of the performance of empirical formulas and the proposed ML method for predicting foundation settlement indicated the rationality of the proposed ML model and its applicability to the compressive deformation of geotechnical systems.This model,however,cannot be directly applied to the prediction of Es in other sites due to its site specificity.This problem can be solved by retraining the model using local data.This study provides a useful reference for future multi-parameter prediction of soil behavior. 展开更多
关键词 Compression modulus prediction Machine learning(ML) gradient boosted regression tree(GBRT) Genetic algorithm(GA) Foundation settlement
原文传递
Adaptive Error Curve Learning Ensemble Model for Improving Energy Consumption Forecasting 被引量:1
8
作者 Prince Waqas Khan Yung-Cheol Byun 《Computers, Materials & Continua》 SCIE EI 2021年第11期1893-1913,共21页
Despite the advancement within the last decades in the field of smart grids,energy consumption forecasting utilizing the metrological features is still challenging.This paper proposes a genetic algorithm-based adaptiv... Despite the advancement within the last decades in the field of smart grids,energy consumption forecasting utilizing the metrological features is still challenging.This paper proposes a genetic algorithm-based adaptive error curve learning ensemble(GA-ECLE)model.The proposed technique copes with the stochastic variations of improving energy consumption forecasting using a machine learning-based ensembled approach.A modified ensemble model based on a utilizing error of model as a feature is used to improve the forecast accuracy.This approach combines three models,namely CatBoost(CB),Gradient Boost(GB),and Multilayer Perceptron(MLP).The ensembled CB-GB-MLP model’s inner mechanism consists of generating a meta-data from Gradient Boosting and CatBoost models to compute the final predictions using the Multilayer Perceptron network.A genetic algorithm is used to obtain the optimal features to be used for the model.To prove the proposed model’s effectiveness,we have used a four-phase technique using Jeju island’s real energy consumption data.In the first phase,we have obtained the results by applying the CB-GB-MLP model.In the second phase,we have utilized a GA-ensembled model with optimal features.The third phase is for the comparison of the energy forecasting result with the proposed ECL-based model.The fourth stage is the final stage,where we have applied the GA-ECLE model.We obtained a mean absolute error of 3.05,and a root mean square error of 5.05.Extensive experimental results are provided,demonstrating the superiority of the proposed GA-ECLE model over traditional ensemble models. 展开更多
关键词 Energy consumption meteorological features error curve learning ensemble model energy forecasting gradient boost catboost multilayer perceptron genetic algorithm
在线阅读 下载PDF
Personalized movie recommendation method based on ensemble learning
9
作者 YANG Kun DUAN Yong 《High Technology Letters》 EI CAS 2022年第1期56-62,共7页
Aiming at the personalized movie recommendation problem,a recommendation algorithm in-tegrating manifold learning and ensemble learning is studied.In this work,manifold learning is used to reduce the dimension of data... Aiming at the personalized movie recommendation problem,a recommendation algorithm in-tegrating manifold learning and ensemble learning is studied.In this work,manifold learning is used to reduce the dimension of data so that both time and space complexities of the model are mitigated.Meanwhile,gradient boosting decision tree(GBDT)is used to train the target user profile prediction model.Based on the recommendation results,Bayesian optimization algorithm is applied to optimize the recommendation model,which can effectively improve the prediction accuracy.The experimental results show that the proposed algorithm can improve the accuracy of movie recommendation. 展开更多
关键词 gradient boosting decision tree(GBDT) recommendation algorithm manifold learn-ing ensemble learning Bayesian optimization
在线阅读 下载PDF
A UAV collaborative defense scheme driven by DDPG algorithm 被引量:3
10
作者 ZHANG Yaozhong WU Zhuoran +1 位作者 XIONG Zhenkai CHEN Long 《Journal of Systems Engineering and Electronics》 SCIE EI CSCD 2023年第5期1211-1224,共14页
The deep deterministic policy gradient(DDPG)algo-rithm is an off-policy method that combines two mainstream reinforcement learning methods based on value iteration and policy iteration.Using the DDPG algorithm,agents ... The deep deterministic policy gradient(DDPG)algo-rithm is an off-policy method that combines two mainstream reinforcement learning methods based on value iteration and policy iteration.Using the DDPG algorithm,agents can explore and summarize the environment to achieve autonomous deci-sions in the continuous state space and action space.In this paper,a cooperative defense with DDPG via swarms of unmanned aerial vehicle(UAV)is developed and validated,which has shown promising practical value in the effect of defending.We solve the sparse rewards problem of reinforcement learning pair in a long-term task by building the reward function of UAV swarms and optimizing the learning process of artificial neural network based on the DDPG algorithm to reduce the vibration in the learning process.The experimental results show that the DDPG algorithm can guide the UAVs swarm to perform the defense task efficiently,meeting the requirements of a UAV swarm for non-centralization,autonomy,and promoting the intelligent development of UAVs swarm as well as the decision-making process. 展开更多
关键词 deep deterministic policy gradient(DDPG)algorithm unmanned aerial vehicles(UAVs)swarm task decision making deep reinforcement learning sparse reward problem
在线阅读 下载PDF
Convergence of gradient method for Elman networks
11
作者 吴微 徐东坡 李正学 《Applied Mathematics and Mechanics(English Edition)》 SCIE EI 2008年第9期1231-1238,共8页
The gradient method for training Elman networks with a finite training sample set is considered. Monotonicity of the error function in the iteration is shown. Weak and strong convergence results are proved, indicating... The gradient method for training Elman networks with a finite training sample set is considered. Monotonicity of the error function in the iteration is shown. Weak and strong convergence results are proved, indicating that the gradient of the error function goes to zero and the weight sequence goes to a fixed point, respectively. A numerical example is given to support the theoretical findings. 展开更多
关键词 Elman network gradient learning algorithm CONVERGENCE MONOTONICITY
在线阅读 下载PDF
Research on three-step accelerated gradient algorithm in deep learning
12
作者 Yongqiang Lian Yincai Tang Shirong Zhou 《Statistical Theory and Related Fields》 2022年第1期40-57,共18页
Gradient descent(GD)algorithm is the widely used optimisation method in training machine learning and deep learning models.In this paper,based on GD,Polyak’s momentum(PM),and Nesterov accelerated gradient(NAG),we giv... Gradient descent(GD)algorithm is the widely used optimisation method in training machine learning and deep learning models.In this paper,based on GD,Polyak’s momentum(PM),and Nesterov accelerated gradient(NAG),we give the convergence of the algorithms from an ini-tial value to the optimal value of an objective function in simple quadratic form.Based on the convergence property of the quadratic function,two sister sequences of NAG’s iteration and par-allel tangent methods in neural networks,the three-step accelerated gradient(TAG)algorithm is proposed,which has three sequences other than two sister sequences.To illustrate the perfor-mance of this algorithm,we compare the proposed algorithm with the three other algorithms in quadratic function,high-dimensional quadratic functions,and nonquadratic function.Then we consider to combine the TAG algorithm to the backpropagation algorithm and the stochastic gradient descent algorithm in deep learning.For conveniently facilitate the proposed algorithms,we rewite the R package‘neuralnet’and extend it to‘supneuralnet’.All kinds of deep learning algorithms in this paper are included in‘supneuralnet’package.Finally,we show our algorithms are superior to other algorithms in four case studies. 展开更多
关键词 Accelerated algorithm backpropagation deep learning learning rate MOMENTUM stochastic gradient descent
原文传递
Interpretable gradient boosting based ensemble learning and African vultures optimization algorithm optimization for estimating deflection induced by excavation
13
作者 Zenglong LIANG Shan LIN +3 位作者 Miao DONG Xitailang CAO Hongwei GUO Hong ZHENG 《Frontiers of Structural and Civil Engineering》 SCIE EI CSCD 2024年第11期1698-1712,共15页
Intelligent construction has become an inevitable trend in the development of the construction industry.In the excavation project,using machine learning methods for early warning can improve construction efficiency an... Intelligent construction has become an inevitable trend in the development of the construction industry.In the excavation project,using machine learning methods for early warning can improve construction efficiency and quality and reduce the chances of damage in the excavation process.An interpretable gradient boosting based ensemble learning framework enhanced by the African Vultures Optimization Algorithm(AVOA)was proposed and evaluated in estimating the diaphragm wall deflections induced by excavation.We investigated and compared the performance of machine learning models in predicting deflections induced by excavation based on a database generated by finite element simulations.First,we exploratively analyzed these data to discover the relationship between features.We used several state-of-the-art intelligent models based on gradient boosting and several simple models for model selection.The hyperparameters for all models in evaluation are optimized using AVOA,and then the optimized models are assembled into a unified framework for fairness assessment.The comprehensive evaluation results show that the AVOA-CatBoost built in this paper performs well(RMSE=1.84,MAE=1.18,R2=0.9993)and cross-validation(RMSE=2.65±1.54,MAE=1.17±0.23,R2=0.998±0.002).In the end,in order to improve the transparency and usefulness of the model,we constructed an interpretable model from both global and local perspectives. 展开更多
关键词 African vultures optimization algorithm gradient boosting ensemble learning interpretable model wall deflection prediction
原文传递
Improving the accuracy of heart disease diagnosis with an augmented back propagation algorithm
14
作者 颜红梅 《Journal of Chongqing University》 CAS 2003年第1期31-34,共4页
A multilayer perceptron neural network system is established to support the diagnosis for five most common heart diseases (coronary heart disease, rheumatic valvular heart disease, hypertension, chronic cor pulmonale ... A multilayer perceptron neural network system is established to support the diagnosis for five most common heart diseases (coronary heart disease, rheumatic valvular heart disease, hypertension, chronic cor pulmonale and congenital heart disease). Momentum term, adaptive learning rate, the forgetting mechanics, and conjugate gradients method are introduced to improve the basic BP algorithm aiming to speed up the convergence of the BP algorithm and enhance the accuracy for diagnosis. A heart disease database consisting of 352 samples is applied to the training and testing courses of the system. The performance of the system is assessed by cross-validation method. It is found that as the basic BP algorithm is improved step by step, the convergence speed and the classification accuracy of the network are enhanced, and the system has great application prospect in supporting heart diseases diagnosis. 展开更多
关键词 multilayer perceptron back propagation algorithm heart disease momentum term adaptive learning rate the forgetting mechanics conjugate gradients method
暂未订购
AN SVAD ALGORITHM BASED ON FNNKD METHOD
15
作者 ChenDong ZhangYan 《Journal of Electronics(China)》 2002年第3期280-288,共9页
The capacity of mobile communication system is improved by using Voice Activity Detection (VAD) technology. In this letter, a novel VAD algorithm, SVAD algorithm based on Fuzzy Neural Network Knowledge Discovery (FNNK... The capacity of mobile communication system is improved by using Voice Activity Detection (VAD) technology. In this letter, a novel VAD algorithm, SVAD algorithm based on Fuzzy Neural Network Knowledge Discovery (FNNKD) method is proposed. The performance of SVAD algorithm is discussed and compared with traditional algorithm recommended by ITU G.729B in different situations. The simulation results show that the SVAD algorithm performs better. 展开更多
关键词 FNNKD VAD Adaptive Multi-Rate (AMR) Heuristic gradient learning algorithm (HGLA)
在线阅读 下载PDF
Policy Gradient算法的研究与实现
16
作者 刘俊利 《现代计算机》 2019年第30期3-5,18,共4页
近年来,由于在多个领域高效率高质量的成功运用,强化学习在机器学习领域和人工智能领域引发极大的关注。越来越多的人开始意识到机器能够自主学习的巨大价值。策略梯度(Policy Gradient)算法是一种基于策略的(policybased)强化学习技术... 近年来,由于在多个领域高效率高质量的成功运用,强化学习在机器学习领域和人工智能领域引发极大的关注。越来越多的人开始意识到机器能够自主学习的巨大价值。策略梯度(Policy Gradient)算法是一种基于策略的(policybased)强化学习技术,是强化学习的经典算法之一。以policy-based方法开始,然后介绍Policy Gradient算法的具体内容,最终利用深度学习框架TensorFlow完成该算法的实现。 展开更多
关键词 TensorFlow 机器学习 强化学习 Policy-Based方法 POLICY gradient算法
在线阅读 下载PDF
一种进化梯度引导的强化学习算法
17
作者 许斌 练元洪 +2 位作者 卞鸿根 刘丹 亓晋 《南京邮电大学学报(自然科学版)》 北大核心 2025年第1期99-105,共7页
进化算法(Evolutionary Algorithm,EA)和深度强化学习(Deep Reinforcement Learning,DRL)的组合被认为能够结合二者的优点,即EA的强大随机搜索能力和DRL的样本效率,实现更好的策略学习。然而,现有的组合方法存在EA引入所导致的策略性能... 进化算法(Evolutionary Algorithm,EA)和深度强化学习(Deep Reinforcement Learning,DRL)的组合被认为能够结合二者的优点,即EA的强大随机搜索能力和DRL的样本效率,实现更好的策略学习。然而,现有的组合方法存在EA引入所导致的策略性能不可预测性问题。提出自适应历史梯度引导机制,其利用历史梯度信息,找到平衡探索和利用的线索,从而获得较为稳定的高质量策略,进一步将此机制融合经典的进化强化学习算法,提出一种进化梯度引导的强化学习算法(Evolutionary Gradient Guided Reinforcement Learning,EGG⁃RL)。在连续控制任务方面的实验表明,EGG⁃RL的性能表现优于其他方法。 展开更多
关键词 CEM⁃RL 深度强化学习 进化算法 历史梯度
在线阅读 下载PDF
基于精英指导和随机搜索的进化强化学习
18
作者 邸剑 万雪 姜丽梅 《系统仿真学报》 北大核心 2025年第11期2877-2887,共11页
针对进化强化学习因样本效率低、耦合方式单一及收敛性差而导致的性能与扩展性受限问题,提出一种基于精英梯度指导和双重随机搜索的改进算法。通过在强化策略训练时引入携带进化信息的精英策略梯度指导,纠正了强化策略梯度更新的方向;... 针对进化强化学习因样本效率低、耦合方式单一及收敛性差而导致的性能与扩展性受限问题,提出一种基于精英梯度指导和双重随机搜索的改进算法。通过在强化策略训练时引入携带进化信息的精英策略梯度指导,纠正了强化策略梯度更新的方向;采用双重随机搜索替换原有的进化组件,降低算法复杂性的同时使得策略搜索在参数空间进行有意义和可控的搜索;引入完全替换信息交易有效地平衡了强化策略和进化策略的学习和探索。实验结果表明:该方法相比于经典的进化强化学习方法在探索力、鲁棒性和收敛性方面具有一定的提升。 展开更多
关键词 进化强化学习 深度强化学习 进化算法 连续控制 精英梯度指导
原文传递
基于深度强化学习的停机位分配
19
作者 向征 吴秋玥 +1 位作者 储同 岳伊杨 《科学技术与工程》 北大核心 2025年第16期6977-6984,共8页
针对停机位分配问题展开系统研究,目标是最小化远机位分配数量以及近机位空闲时间,针对其多目标多约束特性,提出以最小远机位分配数量和最小近机位空闲时间为目标的多目标数学模型,该模型考虑了航班进出港实际时间、机型类别及停机位间... 针对停机位分配问题展开系统研究,目标是最小化远机位分配数量以及近机位空闲时间,针对其多目标多约束特性,提出以最小远机位分配数量和最小近机位空闲时间为目标的多目标数学模型,该模型考虑了航班进出港实际时间、机型类别及停机位间相互关系等参数。结合深度强化学习方法,特别是深度确定性策略梯度算法(deep deterministic policy gradient,DDPG),对停机位分配过程进行优化。为提升算法的寻优能力与性能,设计了改进后的DDPG算法,融入优先级经验回放和多策略融合探索机制。通过对比实验,表明改进后的算法更优,显著减少了最小远机位分配数量并优化了近机位空闲时间,且收敛更快、全局寻优能力更强,充分证实了其有效性。 展开更多
关键词 停机位分配 深度学习 强化学习 深度确定性策略梯度算法(DDPG)
在线阅读 下载PDF
求解多核学习的自适应随机递归梯度下降法
20
作者 王梅 任怡果 +1 位作者 刘勇 王志宝 《计算机技术与发展》 2025年第7期93-99,共7页
针对随机递归梯度法(SARAH)求解多核学习(MKL)的不足之处,如收敛速度缓慢以及计算成本高等问题,该文提出一种改进算法——基于随机Polyak步长(SPS)的小批量随机递归梯度下降算法(SPS-MSARAH)来求解多核学习优化问题。首先将小批量方法... 针对随机递归梯度法(SARAH)求解多核学习(MKL)的不足之处,如收敛速度缓慢以及计算成本高等问题,该文提出一种改进算法——基于随机Polyak步长(SPS)的小批量随机递归梯度下降算法(SPS-MSARAH)来求解多核学习优化问题。首先将小批量方法引入随机方差缩减类算法中,选取一个固定大小的样本集代替单个训练样本计算SARAH的梯度,降低传统随机梯度下降算法使用单个样本计算梯度导致较大的波动和不稳定性所带来的方差。在此基础上,使用随机Polyak步长自适应地更新小批量SARAH的步长,使得优化过程更加灵活和鲁棒,从而解决随机优化算法中步长选取的难题。为了验证该算法的有效性,在标准数据集上进行了详细的数值实验。实验结果显示,在求解大规模多核学习优化问题时,SPS-MSARAH算法不仅显著提高了收敛速度,还有效降低了计算复杂度。此外,对初始参数的敏感性问题也得到了很好的克服,展现出良好的鲁棒性。 展开更多
关键词 多核学习 随机递归梯度下降法 随机Polyak步长 小批量 凸优化
在线阅读 下载PDF
上一页 1 2 14 下一页 到第
使用帮助 返回顶部