期刊文献+
共找到5篇文章
< 1 >
每页显示 20 50 100
融合RIACO-CPMD的清扫机器人全覆盖路径规划
1
作者 庄园 李艳萍 +1 位作者 刘行俊 卢磊 《自动化与仪表》 2025年第12期41-46,共6页
在复杂环境下的清扫机器人全覆盖路径规划过程中,存在包括路径重复率过高、动态避障性能不足以及算法易陷入局部最优解等问题,该文提出了融合优先运动方向的往复式改进蚁群的全覆盖路径规划算法。为解决死区脱困问题,首先构建栅格地图,... 在复杂环境下的清扫机器人全覆盖路径规划过程中,存在包括路径重复率过高、动态避障性能不足以及算法易陷入局部最优解等问题,该文提出了融合优先运动方向的往复式改进蚁群的全覆盖路径规划算法。为解决死区脱困问题,首先构建栅格地图,并通过A^(*)改进蚁群算法引入估价函数、惩罚函数和路径筛选策略,增强蚁群朝向最短路径搜索的趋势;同时改进优先运动方向可变的往复式算法,实现自由区域全覆盖,并通过归一化等权求和法筛选全局最优路径。仿真结果表明,该算法在复杂环境中具有良好的可行性与有效性,相比现有方法能够生成更短的覆盖路径,显著降低重复覆盖率,有效避免局部最优问题,在实现全区域覆盖的同时提升了避障效率与整体性能。 展开更多
关键词 优先运动方向 A^(*)改进蚁群 清扫机器人 死区脱困 全覆盖路径规划
在线阅读 下载PDF
基于强化学习的煤矿澡堂防灭火大模型研究
2
作者 胡双虎 《山西煤炭》 2025年第3期71-77,共7页
煤矿职工澡堂用电环境复杂,电气设备密集,且长期处于高湿状态,火灾隐患突出。传统火灾监测系统依赖单一参数监测,难以精准捕捉电气故障早期特征信号,导致预警滞后。本文利用大模型的强大数据处理和模式识别能力,对煤矿职工澡堂中的火灾... 煤矿职工澡堂用电环境复杂,电气设备密集,且长期处于高湿状态,火灾隐患突出。传统火灾监测系统依赖单一参数监测,难以精准捕捉电气故障早期特征信号,导致预警滞后。本文利用大模型的强大数据处理和模式识别能力,对煤矿职工澡堂中的火灾隐患进行排查,并给出火灾应急策略;结合基于人类反馈的强化学习(RLHF)技术,使用直接偏好优化算法(DPO)优化模型性能,提高其在实际应用场景中的适应性和准确性;同时,设计并实现了面向煤矿职工澡堂智能防灭辅助决策系统,具备实时交互、历史记录回溯、智能引导等功能,为煤矿职工澡堂的消防安全提供更加有力的保障。 展开更多
关键词 防灭火 大模型 基于人类反馈强化学习 直接偏好优化算法 辅助决策
在线阅读 下载PDF
客户绿色偏好约束下的直运调度优化 被引量:4
3
作者 易宣齐 胡志华 《辽宁工程技术大学学报(自然科学版)》 CAS 北大核心 2013年第2期253-256,共4页
针对分布式OD网络中客户绿色偏好约束下的直运调度优化问题,采用两阶段方法即OD流最短路径算法和带时间窗的车辆路径问题(VRPTW)优化,得到环境污染最少的多车型直运调运方案.在特定算例中,15条OD流一共需要配置1辆能耗型车,4辆环保型车... 针对分布式OD网络中客户绿色偏好约束下的直运调度优化问题,采用两阶段方法即OD流最短路径算法和带时间窗的车辆路径问题(VRPTW)优化,得到环境污染最少的多车型直运调运方案.在特定算例中,15条OD流一共需要配置1辆能耗型车,4辆环保型车,与全部采用能耗型车运输相比能够降低32.6%的污染.通过环保车型混合调运,能够以低成本满足客户需求,并降低物流对环境的污染,实现经济效益与社会效益的统一. 展开更多
关键词 分布式OD网络 绿色物流 绿色偏好 直运 两阶段方法 最短路径算法 VRPTW 调运优化
原文传递
Reference direction based immune clone algorithm for many-objective optimization 被引量:1
4
作者 Ruochen LIU Chenlin MA Fei HE Wenping MA Licheng JIAO 《Frontiers of Computer Science》 SCIE EI CSCD 2014年第4期642-655,共14页
In this paper, a new preference multi-objective optimization algorithm called immune clone algorithm based on reference direction method (RD-ICA) is proposed for solving many-objective optimization problems. First, ... In this paper, a new preference multi-objective optimization algorithm called immune clone algorithm based on reference direction method (RD-ICA) is proposed for solving many-objective optimization problems. First, an intelligent recombination operator, which performs well on the functions comprising many parameters, is introduced into an immune clone algorithm so as to explore the potentially excellent gene segments of all individuals in the antibody pop- ulation. Second, a reference direction method, a very strict ranking based on the desire of decision makers (DMs), is used to guide selection and clone of the active population. Then a light beam search (LBS) is borrowed to pick out a small set of individuals filling the external population. The proposed method has been extensively compared with other recently proposed evolutionary multi-objective optimization (EMO) approaches over DTLZ problems with from 4 to 100 objectives. Experimental results indicate RD-ICA can achieve competitive results. 展开更多
关键词 many-objective optimization preference multiobjective optimization artificial immune system reference direction method light beam search intelligent recombination operator
原文传递
Fine-tuning large language models for domain adaptation:exploration of training strategies,scaling,model merging and synergistic capabilities 被引量:1
5
作者 Wei Lu Rachel K.Luu Markus J.Buehler 《npj Computational Materials》 2025年第1期858-900,共43页
The advancement of Large Language Models (LLMs) for domain applications in fields such as materials science and engineering depends on the development of fine-tuning strategies that adapt models for specialized, techn... The advancement of Large Language Models (LLMs) for domain applications in fields such as materials science and engineering depends on the development of fine-tuning strategies that adapt models for specialized, technical capabilities. In this work, we explore the effects of Continued Pretraining (CPT), Supervised Fine-Tuning (SFT), and various preference-based optimization approaches, including Direct Preference Optimization (DPO) and Odds Ratio Preference Optimization (ORPO), on fine-tuned LLM performance. Our analysis shows how these strategies influence model outcomes and reveals that the merging of multiple fine-tuned models can lead to the emergence of capabilities that surpass the individual contributions of the parent models. We find that model merging is not merely a process of aggregation, but a transformative method that can drive substantial advancements in model capabilities characterized by highly nonlinear interactions between model parameters, resulting in new functionalities that neither parent model could achieve alone, leading to improved performance in domain-specific assessments. We study critical factors that influence the success of model merging, such as the diversity between parent models and the fine-tuning techniques employed. The insights underscore the potential of strategic model merging to unlock novel capabilities in LLMs, offering an effective tool for advancing AI systems to meet complex challenges. Experiments with different model architectures are presented, including the Llama 3.1 8B and Mistral 7B family of models, where similar behaviors are observed. Exploring whether the results hold also for much smaller models, we use a tiny LLM with 1.7 billion parameters and show that very small LLMs do not necessarily feature emergent capabilities under model merging, suggesting that model scaling may be a key component. In open-ended yet consistent chat conversations between a human and AI models, our assessment reveals detailed insights into how different model variants perform, and shows that the smallest model achieves a high intelligence score across key criteria including reasoning depth, creativity, clarity, and quantitative precision. Other experiments include the development of image generation prompts that seek to reason over disparate biological material design concepts, to create new microstructures, architectural concepts, and urban design based on biological materials-inspired construction principles. We conclude with a series of questions about scaling and emergence that could be addressed in future research. 展开更多
关键词 continued pretraining domain adaptation training strategies fine tuning materials science direct preference optimization large language models model merging
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部