期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
Memory replay with unlabeled data for semi-supervised class-incremental learning via temporal consistency
1
作者 Qiang WANG Kele XU +2 位作者 Dawei FENG Bo DING Huaimin WANG 《Frontiers of Computer Science》 2025年第12期183-186,共4页
1 Introduction Current continual learning methods[1–4]can utilize labeled data to alleviate catastrophic forgetting effectively.However,obtaining labeled samples can be difficult and tedious as it may require expert ... 1 Introduction Current continual learning methods[1–4]can utilize labeled data to alleviate catastrophic forgetting effectively.However,obtaining labeled samples can be difficult and tedious as it may require expert knowledge.In many practical application scenarios,labeled and unlabeled samples exist simultaneously,with more unlabeled than labeled samples in streaming data[5,6].Unfortunately,existing class-incremental learning methods face limitations in effectively utilizing unlabeled data,thereby impeding their performance in incremental learning scenarios. 展开更多
关键词 alleviate catastrophic forgetting semi supervised learning temporal consistency memory replay labeled data streaming data unfortunatelyexisting unlabeled data labeled samples
原文传递
Deep-reinforcement-learning-based UAV autonomous navigation and collision avoidance in unknown environments 被引量:4
2
作者 Fei WANG Xiaoping ZHU +1 位作者 Zhou ZHOU Yang TANG 《Chinese Journal of Aeronautics》 SCIE EI CAS CSCD 2024年第3期237-257,共21页
In some military application scenarios,Unmanned Aerial Vehicles(UAVs)need to perform missions with the assistance of on-board cameras when radar is not available and communication is interrupted,which brings challenge... In some military application scenarios,Unmanned Aerial Vehicles(UAVs)need to perform missions with the assistance of on-board cameras when radar is not available and communication is interrupted,which brings challenges for UAV autonomous navigation and collision avoidance.In this paper,an improved deep-reinforcement-learning algorithm,Deep Q-Network with a Faster R-CNN model and a Data Deposit Mechanism(FRDDM-DQN),is proposed.A Faster R-CNN model(FR)is introduced and optimized to obtain the ability to extract obstacle information from images,and a new replay memory Data Deposit Mechanism(DDM)is designed to train an agent with a better performance.During training,a two-part training approach is used to reduce the time spent on training as well as retraining when the scenario changes.In order to verify the performance of the proposed method,a series of experiments,including training experiments,test experiments,and typical episodes experiments,is conducted in a 3D simulation environment.Experimental results show that the agent trained by the proposed FRDDM-DQN has the ability to navigate autonomously and avoid collisions,and performs better compared to the FRDQN,FR-DDQN,FR-Dueling DQN,YOLO-based YDDM-DQN,and original FR outputbased FR-ODQN. 展开更多
关键词 Faster R-CNN model replay memory Data Deposit Mechanism(DDM) Two-part training approach Image-based Autonomous Navigation and Collision Avoidance(ANCA) Unmanned Aerial Vehicle(UAV)
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部