期刊文献+
共找到294,211篇文章
< 1 2 250 >
每页显示 20 50 100
Information Diffusion Models and Fuzzing Algorithms for a Privacy-Aware Data Transmission Scheduling in 6G Heterogeneous ad hoc Networks
1
作者 Borja Bordel Sánchez Ramón Alcarria Tomás Robles 《Computer Modeling in Engineering & Sciences》 2026年第2期1214-1234,共21页
In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic h... In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic heterogeneous infrastructures,unstable links and non-uniform hardware capabilities create critical issues regarding security and privacy.Traditional protocols are often too computationally heavy to allow 6G services to achieve their expected Quality-of-Service(QoS).As the transport network is built of ad hoc nodes,there is no guarantee about their trustworthiness or behavior,and transversal functionalities are delegated to the extreme nodes.However,while security can be guaranteed in extreme-to-extreme solutions,privacy cannot,as all intermediate nodes still have to handle the data packets they are transporting.Besides,traditional schemes for private anonymous ad hoc communications are vulnerable against modern intelligent attacks based on learning models.The proposed scheme fulfills this gap.Findings show the probability of a successful intelligent attack reduces by up to 65%compared to ad hoc networks with no privacy protection strategy when used the proposed technology.While congestion probability can remain below 0.001%,as required in 6G services. 展开更多
关键词 6G networks ad hoc networks PRIVACY scheduling algorithms diffusion models fuzzing algorithms
在线阅读 下载PDF
Algorithmically Enhanced Data-Driven Prediction of Shear Strength for Concrete-Filled Steel Tubes
2
作者 Shengkang Zhang Yong Jin +5 位作者 Soon Poh Yap Haoyun Fan Shiyuan Li Ahmed El-Shafie Zainah Ibrahim Amr El-Dieb 《Computer Modeling in Engineering & Sciences》 2026年第1期374-398,共25页
Concrete-filled steel tubes(CFST)are widely utilized in civil engineering due to their superior load-bearing capacity,ductility,and seismic resistance.However,existing design codes,such as AISC and Eurocode 4,tend to ... Concrete-filled steel tubes(CFST)are widely utilized in civil engineering due to their superior load-bearing capacity,ductility,and seismic resistance.However,existing design codes,such as AISC and Eurocode 4,tend to be excessively conservative as they fail to account for the composite action between the steel tube and the concrete core.To address this limitation,this study proposes a hybrid model that integrates XGBoost with the Pied Kingfisher Optimizer(PKO),a nature-inspired algorithm,to enhance the accuracy of shear strength prediction for CFST columns.Additionally,quantile regression is employed to construct prediction intervals for the ultimate shear force,while the Asymmetric Squared Error Loss(ASEL)function is incorporated to mitigate overestimation errors.The computational results demonstrate that the PKO-XGBoost model delivers superior predictive accuracy,achieving a Mean Absolute Percentage Error(MAPE)of 4.431%and R2 of 0.9925 on the test set.Furthermore,the ASEL-PKO-XGBoost model substantially reduces overestimation errors to 28.26%,with negligible impact on predictive performance.Additionally,based on the Genetic Algorithm(GA)and existing equation models,a strength equation model is developed,achieving markedly higher accuracy than existing models(R^(2)=0.934).Lastly,web-based Graphical User Interfaces(GUIs)were developed to enable real-time prediction. 展开更多
关键词 Asymmetric squared error loss genetic algorithm machine learning pied kingfisher optimizer quantile regression
在线阅读 下载PDF
Optimization of the frequency offset increment of FDA-MIMO based on cuckoo search algorithm
3
作者 WANG Bo ZHAO Yu +2 位作者 LI Yonglin YANG Rennong XUE Junjie 《Journal of Systems Engineering and Electronics》 2026年第1期157-170,共14页
Frequency diverse array multiple-input multiple-output(FDA-MIMO)radar has gained considerable research attention due to its ability to effectively counter active repeater deception jamming in complex electromagnetic e... Frequency diverse array multiple-input multiple-output(FDA-MIMO)radar has gained considerable research attention due to its ability to effectively counter active repeater deception jamming in complex electromagnetic environments.The effectiveness of interference suppression by FDA-MIMO is limited by the inherent range-angle coupling issue in the FDA beampattern.Existing literature primarily focuses on control methods for FDA-MIMO radar beam direction under the assumption of static beampatterns,with insufficient exploration of techniques for managing nonstationary beam directions.To address this gap,this paper initially introduces the FDA-MIMO signal model and the calculation formula for the FDA-MIMO array output using the minimum variance distortionless response(MVDR)beamformer.Building on this,the problem of determining the optimal frequency offset for the FDA is rephrased as a convex optimization problem,which is then resolved using the cuckoo search(CS)algorithm.Simulations confirm the effectiveness of the proposed approach,showing that the frequency offsets obtained through the CS algorithm can create a dot-shaped beam direction at the target location while effectively suppressing interference signals within the mainlobe. 展开更多
关键词 frequency diverse array multiple-input multiple-output(Fda-MIMO) convex optimization cuckoo search algorithm beampattern
在线阅读 下载PDF
Study on the destabilizing damage precursors of cemented tailings backfill based on critical slowing down theory combined with multiple denoising algorithms under consideration of initial defect conditions
4
作者 ZHAO Kang ZHONG Jun-cheng +3 位作者 YAN Ya-jing LIU Yang WEN Dao-tan XIAO Wei-ling 《Journal of Central South University》 2026年第1期375-399,共25页
The cemented tailings backfill(CTB)with initial defects is more prone to destabilization damage under the influence of various unfavorable factors during the mining process.In order to investigate its influence on the... The cemented tailings backfill(CTB)with initial defects is more prone to destabilization damage under the influence of various unfavorable factors during the mining process.In order to investigate its influence on the stability of underground mining engineering,this paper simulates the generation of different degrees of initial defects inside the CTB by adding different contents of air-entraining agent(AEA),investigates the acoustic emission RA/AF eigenvalues of CTB with different contents of AEA under uniaxial compression,and adopts various denoising algorithms(e.g.,moving average smoothing,median filtering,and outlier detection)to improve the accuracy of the data.The variance and autocorrelation coefficients of RA/AF parameters were analyzed in conjunction with the critical slowing down(CSD)theory.The results show that the acoustic emission RA/AF values can be used to characterize the progressive damage evolution of CTB.The denoising algorithm processed the AE signals to reduce the effects of extraneous noise and anomalous spikes.Changes in the variance curves provide clear precursor information,while abrupt changes in the autocorrelation coefficient can be used as an auxiliary localization warning signal.The phenomenon of dramatic increase in the variance and autocorrelation coefficient curves during the compression-tightening stage,which is influenced by the initial defects,can lead to false warnings.As the initial defects of the CTB increase,its instability precursor time and instability time are prolonged,the peak stress decreases,and the time difference between the CTB and the instability damage is smaller.The results provide a new method for real-time monitoring and early warning of CTB instability damage. 展开更多
关键词 initial defects cemented tailings backfill critical slowing down acoustic emission RA/AF values denoising algorithms
在线阅读 下载PDF
基于热激励的DAS流量监测试验与数值模拟研究
5
作者 刘均荣 周黎明 +3 位作者 韩艳慧 刘明 姚谦 李志刚 《石油钻探技术》 北大核心 2026年第1期47-57,共11页
为解决现有基于上下游声波信号的分布式光纤声波传感(DAS)流量解释方法在井筒流体流量低时难以识别弱信号的技术难题,引入主动热激励技术,增强信号的差异度。构建室内全尺寸试验装置进行模拟试验,并采用数值模拟验证了该方法的可行性。... 为解决现有基于上下游声波信号的分布式光纤声波传感(DAS)流量解释方法在井筒流体流量低时难以识别弱信号的技术难题,引入主动热激励技术,增强信号的差异度。构建室内全尺寸试验装置进行模拟试验,并采用数值模拟验证了该方法的可行性。模拟试验结果和数值模拟结果表明:主动热激励技术能有效增强DAS低频信号的响应强度,当光纤与热段塞直接接触时,可获得稳定的信号;信号强度主要由热激励强度控制,热激励强度达到10℃即可形成清晰可辨的信号边缘,满足工程应用需求;在流速≥0.0666 m/s(对应流量24 m^(3)/d)工况下,最值追踪法的计算误差可控制在10%以内,但低流速下因热交换充分导致特征点偏移,需进一步优化算法。数值模拟与模拟试验数据在响应趋势和流速计算方面展现出良好的一致性,结构相似性指标验证了数值模型的有效性。研究结果表明,采用10℃热激励强度配合最值追踪法,既可降低系统能耗,又能保障计算精度,主动热激励技术为低产井井下流量监测提供了新的解决方案,但该方法只适用于流量≥24 m^(3)/d的工况,建议下一步优化低流速工况下的特征识别算法及进行现场环境适应性验证。 展开更多
关键词 低流量 热激励 分布式光纤声波传感 物理模拟试验 数值模拟
在线阅读 下载PDF
植物DA1肽酶的调控及其生物学功能研究进展
6
作者 张航 刘源 +4 位作者 陈盈盈 吴丁洁 段若昕 卢振芳 李瑞丽 《植物学报》 北大核心 2026年第1期146-156,共11页
植物DA1肽酶是一种关键的蛋白酶,在植物生长发育和环境适应中扮演着至关重要的角色。近年来,随着基因编辑技术的不断进步,DA1肽酶的作用机制和功能逐渐被阐释。研究表明,DA1基因的表达受转录水平调控,而其蛋白活性则受翻译后修饰调控。... 植物DA1肽酶是一种关键的蛋白酶,在植物生长发育和环境适应中扮演着至关重要的角色。近年来,随着基因编辑技术的不断进步,DA1肽酶的作用机制和功能逐渐被阐释。研究表明,DA1基因的表达受转录水平调控,而其蛋白活性则受翻译后修饰调控。尽管对DA1肽酶生物学功能的研究很多,但是缺少对DA1肽酶不同功能系统而全面的总结。因此,该文主要介绍DA1肽酶的分子结构,重点阐述其在调控器官大小、参与盐胁迫、干旱适应以及免疫激活等胁迫响应中的生物学功能,总结了DA1肽酶在维管形成层活性调节和生长素信号转导中的作用,旨在加深人们对DA1肽酶复杂功能和调控网络的理解。 展开更多
关键词 肽酶 da1 转录调控 翻译后修饰 功能研究
原文传递
Intelligent sequential multi-impulse collision avoidance method for non-cooperative spacecraft based on an improved search tree algorithm 被引量:1
7
作者 Xuyang CAO Xin NING +4 位作者 Zheng WANG Suyi LIU Fei CHENG Wenlong LI Xiaobin LIAN 《Chinese Journal of Aeronautics》 2025年第4期378-393,共16页
The problem of collision avoidance for non-cooperative targets has received significant attention from researchers in recent years.Non-cooperative targets exhibit uncertain states and unpredictable behaviors,making co... The problem of collision avoidance for non-cooperative targets has received significant attention from researchers in recent years.Non-cooperative targets exhibit uncertain states and unpredictable behaviors,making collision avoidance significantly more challenging than that for space debris.Much existing research focuses on the continuous thrust model,whereas the impulsive maneuver model is more appropriate for long-duration and long-distance avoidance missions.Additionally,it is important to minimize the impact on the original mission while avoiding noncooperative targets.On the other hand,the existing avoidance algorithms are computationally complex and time-consuming especially with the limited computing capability of the on-board computer,posing challenges for practical engineering applications.To conquer these difficulties,this paper makes the following key contributions:(A)a turn-based(sequential decision-making)limited-area impulsive collision avoidance model considering the time delay of precision orbit determination is established for the first time;(B)a novel Selection Probability Learning Adaptive Search-depth Search Tree(SPL-ASST)algorithm is proposed for non-cooperative target avoidance,which improves the decision-making efficiency by introducing an adaptive-search-depth mechanism and a neural network into the traditional Monte Carlo Tree Search(MCTS).Numerical simulations confirm the effectiveness and efficiency of the proposed method. 展开更多
关键词 Non-cooperative target Collision avoidance Limited motion area Impulsive maneuver model Search tree algorithm Neural networks
原文传递
Bat algorithm based on kinetic adaptation and elite communication for engineering problems
8
作者 Chong Yuan Dong Zhao +4 位作者 Ali Asghar Heidari Lei Liu Shuihua Wang Huiling Chen Yudong Zhang 《CAAI Transactions on Intelligence Technology》 2025年第4期1174-1200,共27页
The Bat algorithm,a metaheuristic optimization technique inspired by the foraging behaviour of bats,has been employed to tackle optimization problems.Known for its ease of implementation,parameter tunability,and stron... The Bat algorithm,a metaheuristic optimization technique inspired by the foraging behaviour of bats,has been employed to tackle optimization problems.Known for its ease of implementation,parameter tunability,and strong global search capabilities,this algorithm finds application across diverse optimization problem domains.However,in the face of increasingly complex optimization challenges,the Bat algorithm encounters certain limitations,such as slow convergence and sensitivity to initial solutions.In order to tackle these challenges,the present study incorporates a range of optimization compo-nents into the Bat algorithm,thereby proposing a variant called PKEBA.A projection screening strategy is implemented to mitigate its sensitivity to initial solutions,thereby enhancing the quality of the initial solution set.A kinetic adaptation strategy reforms exploration patterns,while an elite communication strategy enhances group interaction,to avoid algorithm from local optima.Subsequently,the effectiveness of the proposed PKEBA is rigorously evaluated.Testing encompasses 30 benchmark functions from IEEE CEC2014,featuring ablation experiments and comparative assessments against classical algorithms and their variants.Moreover,real-world engineering problems are employed as further validation.The results conclusively demonstrate that PKEBA ex-hibits superior convergence and precision compared to existing algorithms. 展开更多
关键词 Bat algorithm engineering optimization global optimization metaheuristic algorithms
在线阅读 下载PDF
An Adaptive Firefly Algorithm for Dependent Task Scheduling in IoT-Fog Computing
9
作者 Adil Yousif 《Computer Modeling in Engineering & Sciences》 2025年第3期2869-2892,共24页
The Internet of Things(IoT)has emerged as an important future technology.IoT-Fog is a new computing paradigm that processes IoT data on servers close to the source of the data.In IoT-Fog computing,resource allocation ... The Internet of Things(IoT)has emerged as an important future technology.IoT-Fog is a new computing paradigm that processes IoT data on servers close to the source of the data.In IoT-Fog computing,resource allocation and independent task scheduling aim to deliver short response time services demanded by the IoT devices and performed by fog servers.The heterogeneity of the IoT-Fog resources and the huge amount of data that needs to be processed by the IoT-Fog tasks make scheduling fog computing tasks a challenging problem.This study proposes an Adaptive Firefly Algorithm(AFA)for dependent task scheduling in IoT-Fog computing.The proposed AFA is a modified version of the standard Firefly Algorithm(FA),considering the execution times of the submitted tasks,the impact of synchronization requirements,and the communication time between dependent tasks.As IoT-Fog computing depends mainly on distributed fog node servers that receive tasks in a dynamic manner,tackling the communications and synchronization issues between dependent tasks is becoming a challenging problem.The proposed AFA aims to address the dynamic nature of IoT-Fog computing environments.The proposed AFA mechanism considers a dynamic light absorption coefficient to control the decrease in attractiveness over iterations.The proposed AFA mechanism performance was benchmarked against the standard Firefly Algorithm(FA),Puma Optimizer(PO),Genetic Algorithm(GA),and Ant Colony Optimization(ACO)through simulations under light,typical,and heavy workload scenarios.In heavy workloads,the proposed AFA mechanism obtained the shortest average execution time,968.98 ms compared to 970.96,1352.87,1247.28,and 1773.62 of FA,PO,GA,and ACO,respectively.The simulation results demonstrate the proposed AFA’s ability to rapidly converge to optimal solutions,emphasizing its adaptability and efficiency in typical and heavy workloads. 展开更多
关键词 Fog computing SCHEDULING resource management firefly algorithm genetic algorithm ant colony optimization
在线阅读 下载PDF
Phasmatodea Population Evolution Algorithm Based on Spiral Mechanism and Its Application to Data Clustering
10
作者 Jeng-Shyang Pan Mengfei Zhang +2 位作者 Shu-Chuan Chu Xingsi Xue Václav Snášel 《Computers, Materials & Continua》 2025年第4期475-496,共22页
Data clustering is an essential technique for analyzing complex datasets and continues to be a central research topic in data analysis.Traditional clustering algorithms,such as K-means,are widely used due to their sim... Data clustering is an essential technique for analyzing complex datasets and continues to be a central research topic in data analysis.Traditional clustering algorithms,such as K-means,are widely used due to their simplicity and efficiency.This paper proposes a novel Spiral Mechanism-Optimized Phasmatodea Population Evolution Algorithm(SPPE)to improve clustering performance.The SPPE algorithm introduces several enhancements to the standard Phasmatodea Population Evolution(PPE)algorithm.Firstly,a Variable Neighborhood Search(VNS)factor is incorporated to strengthen the local search capability and foster population diversity.Secondly,a position update model,incorporating a spiral mechanism,is designed to improve the algorithm’s global exploration and convergence speed.Finally,a dynamic balancing factor,guided by fitness values,adjusts the search process to balance exploration and exploitation effectively.The performance of SPPE is first validated on CEC2013 benchmark functions,where it demonstrates excellent convergence speed and superior optimization results compared to several state-of-the-art metaheuristic algorithms.To further verify its practical applicability,SPPE is combined with the K-means algorithm for data clustering and tested on seven datasets.Experimental results show that SPPE-K-means improves clustering accuracy,reduces dependency on initialization,and outperforms other clustering approaches.This study highlights SPPE’s robustness and efficiency in solving both optimization and clustering challenges,making it a promising tool for complex data analysis tasks. 展开更多
关键词 Phasmatodea population evolution algorithm data clustering meta-heuristic algorithm
在线阅读 下载PDF
Optimal performance design of bat algorithm:An adaptive multi-stage structure
11
作者 Helong Yu Jiuman Song +4 位作者 Chengcheng Chen Ali Asghar Heidari Yuntao Ma Huiling Chen Yudong Zhang 《CAAI Transactions on Intelligence Technology》 2025年第3期755-814,共60页
The bat algorithm(BA)is a metaheuristic algorithm for global optimisation that simulates the echolocation behaviour of bats with varying pulse rates of emission and loudness,which can be used to find the globally opti... The bat algorithm(BA)is a metaheuristic algorithm for global optimisation that simulates the echolocation behaviour of bats with varying pulse rates of emission and loudness,which can be used to find the globally optimal solutions for various optimisation problems.Knowing the recent criticises of the originality of equations,the principle of BA is concise and easy to implement,and its mathematical structure can be seen as a hybrid particle swarm with simulated annealing.In this research,the authors focus on the performance optimisation of BA as a solver rather than discussing its originality issues.In terms of operation effect,BA has an acceptable convergence speed.However,due to the low proportion of time used to explore the search space,it is easy to converge prematurely and fall into the local optima.The authors propose an adaptive multi-stage bat algorithm(AMSBA).By tuning the algorithm's focus at three different stages of the search process,AMSBA can achieve a better balance between exploration and exploitation and improve its exploration ability by enhancing its performance in escaping local optima as well as maintaining a certain convergence speed.Therefore,AMSBA can achieve solutions with better quality.A convergence analysis was conducted to demonstrate the global convergence of AMSBA.The authors also perform simulation experiments on 30 benchmark functions from IEEE CEC 2017 as the objective functions and compare AMSBA with some original and improved swarm-based algorithms.The results verify the effectiveness and superiority of AMSBA.AMSBA is also compared with eight representative optimisation algorithms on 10 benchmark functions derived from IEEE CEC 2020,while this experiment is carried out on five different dimensions of the objective functions respectively.A balance and diversity analysis was performed on AMSBA to demonstrate its improvement over the original BA in terms of balance.AMSBA was also applied to the multi-threshold image segmentation of Citrus Macular disease,which is a bacterial infection that causes lesions on citrus trees.The segmentation results were analysed by comparing each comparative algorithm's peak signal-to-noise ratio,structural similarity index and feature similarity index.The results show that the proposed BA-based algorithm has apparent advantages,and it can effectively segment the disease spots from citrus leaves when the segmentation threshold is at a low level.Based on a comprehensive study,the authors think the proposed optimiser has mitigated the main drawbacks of the BA,and it can be utilised as an effective optimisation tool. 展开更多
关键词 bat-inspired algorithm Citrus Macular disease global optimization multi-threshold image segmentation Otsu algorithm
在线阅读 下载PDF
An Eulerian-Lagrangian parallel algorithm for simulation of particle-laden turbulent flows 被引量:1
12
作者 Harshal P.Mahamure Deekshith I.Poojary +1 位作者 Vagesh D.Narasimhamurthy Lihao Zhao 《Acta Mechanica Sinica》 2026年第1期15-34,共20页
This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in ... This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in turbulent carrier flow.The Eulerian framework numerically resolves turbulent carrier flow using a parallelized,finite-volume DNS solver on a staggered Cartesian grid.Particles are tracked using a point-particle method utilizing a Lagrangian particle tracking(LPT)algorithm.The proposed Eulerian-Lagrangian algorithm is validated using an inertial particle-laden turbulent channel flow for different Stokes number cases.The particle concentration profiles and higher-order statistics of the carrier and dispersed phases agree well with the benchmark results.We investigated the effect of fluid velocity interpolation and numerical integration schemes of particle tracking algorithms on particle dispersion statistics.The suitability of fluid velocity interpolation schemes for predicting the particle dispersion statistics is discussed in the framework of the particle tracking algorithm coupled to the finite-volume solver.In addition,we present parallelization strategies implemented in the algorithm and evaluate their parallel performance. 展开更多
关键词 DNS Eulerian-Lagrangian Particle tracking algorithm Point-particle Parallel software
原文传递
Dynamic Multi-Objective Gannet Optimization(DMGO):An Adaptive Algorithm for Efficient Data Replication in Cloud Systems
13
作者 P.William Ved Prakash Mishra +3 位作者 Osamah Ibrahim Khalaf Arvind Mukundan Yogeesh N Riya Karmakar 《Computers, Materials & Continua》 2025年第9期5133-5156,共24页
Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple dat... Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple data centers poses a significant challenge,especially when balancing opposing goals such as latency,storage costs,energy consumption,and network efficiency.This study introduces a novel Dynamic Optimization Algorithm called Dynamic Multi-Objective Gannet Optimization(DMGO),designed to enhance data replication efficiency in cloud environments.Unlike traditional static replication systems,DMGO adapts dynamically to variations in network conditions,system demand,and resource availability.The approach utilizes multi-objective optimization approaches to efficiently balance data access latency,storage efficiency,and operational costs.DMGO consistently evaluates data center performance and adjusts replication algorithms in real time to guarantee optimal system efficiency.Experimental evaluations conducted in a simulated cloud environment demonstrate that DMGO significantly outperforms conventional static algorithms,achieving faster data access,lower storage overhead,reduced energy consumption,and improved scalability.The proposed methodology offers a robust and adaptable solution for modern cloud systems,ensuring efficient resource consumption while maintaining high performance. 展开更多
关键词 Cloud computing data replication dynamic optimization multi-objective optimization gannet optimization algorithm adaptive algorithms resource efficiency SCALABILITY latency reduction energy-efficient computing
在线阅读 下载PDF
Artificial intelligence in the service of entrepreneurial finance:knowledge structure and the foundational algorithmic paradigm
14
作者 Robert Kudelić Tamara Šmaguc Sherry Robinson 《Financial Innovation》 2025年第1期2021-2063,共43页
The study conducts a bibliometric review of artificial intelligence applications in two areas:the entrepreneurial finance literature,and the corporate finance literature with implications for entrepreneurship.A rigoro... The study conducts a bibliometric review of artificial intelligence applications in two areas:the entrepreneurial finance literature,and the corporate finance literature with implications for entrepreneurship.A rigorous search and screening of the web of science core collection identified 1,890 journal articles for analysis.The bibliometrics provide a detailed view of the knowledge field,indicating underdeveloped research directions.An important contribution comes from insights through artificial intelligence methods in entrepreneurship.The results demonstrate a high representation of artificial neural networks,deep neural networks,and support vector machines across almost all identified topic niches.In contrast,applications of topic modeling,fuzzy neural networks,and growing hierarchical self-organizing maps are rare.Additionally,we take a broader view by addressing the problem of applying artificial intelligence in economic science.Specifically,we present the foundational paradigm and a bespoke demonstration of the Monte Carlo randomized algorithm. 展开更多
关键词 BIBLIOMETRICS Artificial intelligence ENTREPRENEURSHIP FINANCE Randomized algorithm
在线阅读 下载PDF
PID Steering Control Method of Agricultural Robot Based on Fusion of Particle Swarm Optimization and Genetic Algorithm
15
作者 ZHAO Longlian ZHANG Jiachuang +2 位作者 LI Mei DONG Zhicheng LI Junhui 《农业机械学报》 北大核心 2026年第1期358-367,共10页
Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion... Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion algorithm took advantage of the fast optimization ability of PSO to optimize the population screening link of GA.The Simulink simulation results showed that the convergence of the fitness function of the fusion algorithm was accelerated,the system response adjustment time was reduced,and the overshoot was almost zero.Then the algorithm was applied to the steering test of agricultural robot in various scenes.After modeling the steering system of agricultural robot,the steering test results in the unloaded suspended state showed that the PID control based on fusion algorithm reduced the rise time,response adjustment time and overshoot of the system,and improved the response speed and stability of the system,compared with the artificial trial and error PID control and the PID control based on GA.The actual road steering test results showed that the PID control response rise time based on the fusion algorithm was the shortest,about 4.43 s.When the target pulse number was set to 100,the actual mean value in the steady-state regulation stage was about 102.9,which was the closest to the target value among the three control methods,and the overshoot was reduced at the same time.The steering test results under various scene states showed that the PID control based on the proposed fusion algorithm had good anti-interference ability,it can adapt to the changes of environment and load and improve the performance of the control system.It was effective in the steering control of agricultural robot.This method can provide a reference for the precise steering control of other robots. 展开更多
关键词 agricultural robot steering PID control particle swarm optimization algorithm genetic algorithm
在线阅读 下载PDF
Data Empowerment in Precision Marketing: Algorithm Recommendations and Their Associated Risks
16
作者 Di Zhou 《Proceedings of Business and Economic Studies》 2025年第1期111-118,共8页
This paper examines the impact of algorithmic recommendations and data-driven marketing on consumer engagement and business performance.By leveraging large volumes of user data,businesses can deliver personalized cont... This paper examines the impact of algorithmic recommendations and data-driven marketing on consumer engagement and business performance.By leveraging large volumes of user data,businesses can deliver personalized content that enhances user experiences and increases conversion rates.However,the growing reliance on these technologies introduces significant risks,including privacy violations,algorithmic bias,and ethical concerns.This paper explores these challenges and provides recommendations for businesses to mitigate associated risks while optimizing marketing strategies.It highlights the importance of transparency,fairness,and user control in ensuring responsible and effective data-driven marketing. 展开更多
关键词 data-driven marketing algorithmic recommendations Privacy and ethics
在线阅读 下载PDF
Boosting algorithms for predicting end-point temperature in BOF steelmaking using big industrial datasets
17
作者 Jian-bo Zhang Maryam Khaksar Ghalati +3 位作者 Jun Fu Xiao-an Yang G.M.A.M.El-Fallah Hong-biao Dong 《Journal of Iron and Steel Research International》 2025年第7期1856-1868,共13页
The application of machine learning was investigated for predicting end-point temperature in the basic oxygen furnace steelmaking process,addressing gaps in the field,particularly large-scale dataset sizes and the und... The application of machine learning was investigated for predicting end-point temperature in the basic oxygen furnace steelmaking process,addressing gaps in the field,particularly large-scale dataset sizes and the underutilization of boosting algorithms.Utilizing a substantial dataset containing over 20,000 heats,significantly bigger than those in previous studies,a comprehensive evaluation of five advanced machine learning models was conducted.These include four ensemble learning algorithms:XGBoost,LightGBM,CatBoost(three boosting algorithms),along with random forest(a bagging algorithm),as well as a neural network model,namely the multilayer perceptron.Our comparative analysis reveals that Bayesian-optimized boosting models demonstrate exceptional robustness and accuracy,achieving the highest R-squared values,the lowest root mean square error,and lowest mean absolute error,along with the best hit ratio.CatBoost exhibited superior performance,with its test R-squared improving by 4.2%compared to that of the random forest and by 0.8%compared to that of the multilayer perceptron.This highlights the efficacy of boosting algorithms in refining complex industrial processes.Additionally,our investigation into the impact of varying dataset sizes,ranging from 500 to 20,000 heats,on model accuracy underscores the importance of leveraging larger-scale datasets to improve the accuracy and stability of predictive models. 展开更多
关键词 STEELMAKING Basic oxygen furnace Machine learning-Boosting algorithm
原文传递
DCS-SOCP-SVM:A Novel Integrated Sampling and Classification Algorithm for Imbalanced Datasets
18
作者 Xuewen Mu Bingcong Zhao 《Computers, Materials & Continua》 2025年第5期2143-2159,共17页
When dealing with imbalanced datasets,the traditional support vectormachine(SVM)tends to produce a classification hyperplane that is biased towards the majority class,which exhibits poor robustness.This paper proposes... When dealing with imbalanced datasets,the traditional support vectormachine(SVM)tends to produce a classification hyperplane that is biased towards the majority class,which exhibits poor robustness.This paper proposes a high-performance classification algorithm specifically designed for imbalanced datasets.The proposed method first uses a biased second-order cone programming support vectormachine(B-SOCP-SVM)to identify the support vectors(SVs)and non-support vectors(NSVs)in the imbalanced data.Then,it applies the synthetic minority over-sampling technique(SV-SMOTE)to oversample the support vectors of the minority class and uses the random under-sampling technique(NSV-RUS)multiple times to undersample the non-support vectors of the majority class.Combining the above-obtained minority class data set withmultiple majority class datasets can obtainmultiple new balanced data sets.Finally,SOCP-SVM is used to classify each data set,and the final result is obtained through the integrated algorithm.Experimental results demonstrate that the proposed method performs excellently on imbalanced datasets. 展开更多
关键词 DCS-SOCP-SVM imbalanced datasets sampling method ensemble method integrated algorithm
在线阅读 下载PDF
Adaptive Time Synchronization in Time Sensitive-Wireless Sensor Networks Based on Stochastic Gradient Algorithms Framework
19
作者 Ramadan Abdul-Rashid Mohd Amiruddin Abd Rahman +1 位作者 Kar Tim Chan Arun Kumar Sangaiah 《Computer Modeling in Engineering & Sciences》 2025年第3期2585-2616,共32页
This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different... This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different stochastic gradient algorithms can be adopted for adaptive clock frequency adjustments.The study analyzes the pairwise synchronization behavior of the protocol and proves the generalized convergence of the synchronization error and clock frequency.A novel closed-form expression is also derived for a generalized asymptotic error variance steady state.Steady and convergence analyses are then presented for the synchronization,with frequency adaptations done using least mean square(LMS),the Newton search,the gradient descent(GraDes),the normalized LMS(N-LMS),and the Sign-Data LMS algorithms.Results obtained from real-time experiments showed a better performance of our protocols as compared to the Average Proportional-Integral Synchronization Protocol(AvgPISync)regarding the impact of quantization error on synchronization accuracy,precision,and convergence time.This generalized approach to time synchronization allows flexibility in selecting a suitable protocol for different wireless sensor network applications. 展开更多
关键词 Wireless sensor network time synchronization stochastic gradient algorithm MULTI-HOP
在线阅读 下载PDF
Optimizing Cancer Classification and Gene Discovery with an Adaptive Learning Search Algorithm for Microarray Analysis
20
作者 Chiwen Qu Heng Yao +1 位作者 Tingjiang Pan Zenghui Lu 《Journal of Bionic Engineering》 2025年第2期901-930,共30页
DNA microarrays, a cornerstone in biomedicine, measure gene expression across thousands to tens of thousands of genes. Identifying the genes vital for accurate cancer classification is a key challenge. Here, we presen... DNA microarrays, a cornerstone in biomedicine, measure gene expression across thousands to tens of thousands of genes. Identifying the genes vital for accurate cancer classification is a key challenge. Here, we present Fs-LSA (F-score based Learning Search Algorithm), a novel gene selection algorithm designed to enhance the precision and efficiency of target gene identification from microarray data for cancer classification. This algorithm is divided into two phases: the first leverages F-score values to prioritize and select feature genes with the most significant differential expression;the second phase introduces our Learning Search Algorithm (LSA), which harnesses swarm intelligence to identify the optimal subset among the remaining genes. Inspired by human social learning, LSA integrates historical data and collective intelligence for a thorough search, with a dynamic control mechanism that balances exploration and refinement, thereby enhancing the gene selection process. We conducted a rigorous validation of Fs-LSA’s performance using eight publicly available cancer microarray expression datasets. Fs-LSA achieved accuracy, precision, sensitivity, and F1-score values of 0.9932, 0.9923, 0.9962, and 0.994, respectively. Comparative analyses with state-of-the-art algorithms revealed Fs-LSA’s superior performance in terms of simplicity and efficiency. Additionally, we validated the algorithm’s efficacy independently using glioblastoma data from GEO and TCGA databases. It was significantly superior to those of the comparison algorithms. Importantly, the driver genes identified by Fs-LSA were instrumental in developing a predictive model as an independent prognostic indicator for glioblastoma, underscoring Fs-LSA’s transformative potential in genomics and personalized medicine. 展开更多
关键词 Gene selection Learning search algorithm Gene expression data CLASSIFICATION
暂未订购
上一页 1 2 250 下一页 到第
使用帮助 返回顶部