为更精确地预测航班过站时间,将全国机场按照规模差异及不同地理位置所导致的客流量差异和天气差异对航班过站时间造成的不同影响进行分类,基于各类机场航班数据,构建混合轻量级梯度提升机算法(LightGBM)模型对航班过站时间分类预测。...为更精确地预测航班过站时间,将全国机场按照规模差异及不同地理位置所导致的客流量差异和天气差异对航班过站时间造成的不同影响进行分类,基于各类机场航班数据,构建混合轻量级梯度提升机算法(LightGBM)模型对航班过站时间分类预测。引入自适应鲁棒损失函数(adaptive robust loss function,ARLF)改进LightGBM模型损失函数,降低航班数据中存在离群值的影响;通过改进的麻雀搜索算法对改进后的LightGBM模型进行参数寻优,形成混合LightGBM模型。采用全国2019年全年航班数据进行验证,实验结果验证了方法的可行性。展开更多
Learning from demonstration is widely regarded as a promising paradigm for robots to acquire diverse skills.Other than the artificial learning from observation-action pairs for machines,humans can learn to imitate in ...Learning from demonstration is widely regarded as a promising paradigm for robots to acquire diverse skills.Other than the artificial learning from observation-action pairs for machines,humans can learn to imitate in a more versatile and effective manner:acquiring skills through mere“observation”.Video to Command task is widely perceived as a promising approach for task-based learning,which yet faces two key challenges:(1)High redundancy and low frame rate of fine-grained action sequences make it difficult to manipulate objects robustly and accurately.(2)Video to Command models often prioritize accuracy and richness of output commands over physical capabilities,leading to impractical or unsafe instructions for robots.This article presents a novel Video to Command framework that employs multiple data associations and physical constraints.First,we introduce an object-level appearancecontrasting multiple data association strategy to effectively associate manipulated objects in visually complex environments,capturing dynamic changes in video content.Then,we propose a multi-task Video to Command model that utilizes object-level video content changes to compile expert demonstrations into manipulation commands.Finally,a multi-task hybrid loss function is proposed to train a Video to Command model that adheres to the constraints of the physical world and manipulation tasks.Our method achieved over 10%on BLEU_N,METEOR,ROUGE_L,and CIDEr compared to the up-to-date methods.The dual-arm robot prototype was established to demonstrate the whole process of learning from an expert demonstration of multiple skills and then executing the tasks by a robot.展开更多
Segmentation of the retinal vessels in the fundus is crucial for diagnosing ocular diseases.Retinal vessel images often suffer from category imbalance and large scale variations.This ultimately results in incomplete v...Segmentation of the retinal vessels in the fundus is crucial for diagnosing ocular diseases.Retinal vessel images often suffer from category imbalance and large scale variations.This ultimately results in incomplete vessel segmentation and poor continuity.In this study,we propose CT-MFENet to address the aforementioned issues.First,the use of context transformer(CT)allows for the integration of contextual feature information,which helps establish the connection between pixels and solve the problem of incomplete vessel continuity.Second,multi-scale dense residual networks are used instead of traditional CNN to address the issue of inadequate local feature extraction when the model encounters vessels at multiple scales.In the decoding stage,we introduce a local-global fusion module.It enhances the localization of vascular information and reduces the semantic gap between high-and low-level features.To address the class imbalance in retinal images,we propose a hybrid loss function that enhances the segmentation ability of the model for topological structures.We conducted experiments on the publicly available DRIVE,CHASEDB1,STARE,and IOSTAR datasets.The experimental results show that our CT-MFENet performs better than most existing methods,including the baseline U-Net.展开更多
As an important part of the new generation of information technology,the Internet of Things(IoT)has been widely concerned and regarded as an enabling technology of the next generation of health care system.The fundus ...As an important part of the new generation of information technology,the Internet of Things(IoT)has been widely concerned and regarded as an enabling technology of the next generation of health care system.The fundus photography equipment is connected to the cloud platform through the IoT,so as to realize the realtime uploading of fundus images and the rapid issuance of diagnostic suggestions by artificial intelligence.At the same time,important security and privacy issues have emerged.The data uploaded to the cloud platform involves more personal attributes,health status and medical application data of patients.Once leaked,abused or improperly disclosed,personal information security will be violated.Therefore,it is important to address the security and privacy issues of massive medical and healthcare equipment connecting to the infrastructure of IoT healthcare and health systems.To meet this challenge,we propose MIA-UNet,a multi-scale iterative aggregation U-network,which aims to achieve accurate and efficient retinal vessel segmentation for ophthalmic auxiliary diagnosis while ensuring that the network has low computational complexity to adapt to mobile terminals.In this way,users do not need to upload the data to the cloud platform,and can analyze and process the fundus images on their own mobile terminals,thus eliminating the leakage of personal information.Specifically,the interconnection between encoder and decoder,as well as the internal connection between decoder subnetworks in classic U-Net are redefined and redesigned.Furthermore,we propose a hybrid loss function to smooth the gradient and deal with the imbalance between foreground and background.Compared with the UNet,the segmentation performance of the proposed network is significantly improved on the premise that the number of parameters is only increased by 2%.When applied to three publicly available datasets:DRIVE,STARE and CHASE DB1,the proposed network achieves the accuracy/F1-score of 96.33%/84.34%,97.12%/83.17%and 97.06%/84.10%,respectively.The experimental results show that the MIA-UNet is superior to the state-of-the-art methods.展开更多
文摘为更精确地预测航班过站时间,将全国机场按照规模差异及不同地理位置所导致的客流量差异和天气差异对航班过站时间造成的不同影响进行分类,基于各类机场航班数据,构建混合轻量级梯度提升机算法(LightGBM)模型对航班过站时间分类预测。引入自适应鲁棒损失函数(adaptive robust loss function,ARLF)改进LightGBM模型损失函数,降低航班数据中存在离群值的影响;通过改进的麻雀搜索算法对改进后的LightGBM模型进行参数寻优,形成混合LightGBM模型。采用全国2019年全年航班数据进行验证,实验结果验证了方法的可行性。
基金Supported by Zhejiang Provincial Key Research and Development Program(Grant No.2021C04015)。
文摘Learning from demonstration is widely regarded as a promising paradigm for robots to acquire diverse skills.Other than the artificial learning from observation-action pairs for machines,humans can learn to imitate in a more versatile and effective manner:acquiring skills through mere“observation”.Video to Command task is widely perceived as a promising approach for task-based learning,which yet faces two key challenges:(1)High redundancy and low frame rate of fine-grained action sequences make it difficult to manipulate objects robustly and accurately.(2)Video to Command models often prioritize accuracy and richness of output commands over physical capabilities,leading to impractical or unsafe instructions for robots.This article presents a novel Video to Command framework that employs multiple data associations and physical constraints.First,we introduce an object-level appearancecontrasting multiple data association strategy to effectively associate manipulated objects in visually complex environments,capturing dynamic changes in video content.Then,we propose a multi-task Video to Command model that utilizes object-level video content changes to compile expert demonstrations into manipulation commands.Finally,a multi-task hybrid loss function is proposed to train a Video to Command model that adheres to the constraints of the physical world and manipulation tasks.Our method achieved over 10%on BLEU_N,METEOR,ROUGE_L,and CIDEr compared to the up-to-date methods.The dual-arm robot prototype was established to demonstrate the whole process of learning from an expert demonstration of multiple skills and then executing the tasks by a robot.
基金the National Natural Science Foundation of China(No.62266025)。
文摘Segmentation of the retinal vessels in the fundus is crucial for diagnosing ocular diseases.Retinal vessel images often suffer from category imbalance and large scale variations.This ultimately results in incomplete vessel segmentation and poor continuity.In this study,we propose CT-MFENet to address the aforementioned issues.First,the use of context transformer(CT)allows for the integration of contextual feature information,which helps establish the connection between pixels and solve the problem of incomplete vessel continuity.Second,multi-scale dense residual networks are used instead of traditional CNN to address the issue of inadequate local feature extraction when the model encounters vessels at multiple scales.In the decoding stage,we introduce a local-global fusion module.It enhances the localization of vascular information and reduces the semantic gap between high-and low-level features.To address the class imbalance in retinal images,we propose a hybrid loss function that enhances the segmentation ability of the model for topological structures.We conducted experiments on the publicly available DRIVE,CHASEDB1,STARE,and IOSTAR datasets.The experimental results show that our CT-MFENet performs better than most existing methods,including the baseline U-Net.
基金This work was supported in part by the National Natural Science Foundation of China(Nos.62072074,62076054,62027827,61902054)the Frontier Science and Technology Innovation Projects of National Key R&D Program(No.2019QY1405)+2 种基金the Sichuan Science and Technology Innovation Platform and Talent Plan(No.2020JDJQ0020)the Sichuan Science and Technology Support Plan(No.2020YFSY0010)the Natural Science Foundation of Guangdong Province(No.2018A030313354).
文摘As an important part of the new generation of information technology,the Internet of Things(IoT)has been widely concerned and regarded as an enabling technology of the next generation of health care system.The fundus photography equipment is connected to the cloud platform through the IoT,so as to realize the realtime uploading of fundus images and the rapid issuance of diagnostic suggestions by artificial intelligence.At the same time,important security and privacy issues have emerged.The data uploaded to the cloud platform involves more personal attributes,health status and medical application data of patients.Once leaked,abused or improperly disclosed,personal information security will be violated.Therefore,it is important to address the security and privacy issues of massive medical and healthcare equipment connecting to the infrastructure of IoT healthcare and health systems.To meet this challenge,we propose MIA-UNet,a multi-scale iterative aggregation U-network,which aims to achieve accurate and efficient retinal vessel segmentation for ophthalmic auxiliary diagnosis while ensuring that the network has low computational complexity to adapt to mobile terminals.In this way,users do not need to upload the data to the cloud platform,and can analyze and process the fundus images on their own mobile terminals,thus eliminating the leakage of personal information.Specifically,the interconnection between encoder and decoder,as well as the internal connection between decoder subnetworks in classic U-Net are redefined and redesigned.Furthermore,we propose a hybrid loss function to smooth the gradient and deal with the imbalance between foreground and background.Compared with the UNet,the segmentation performance of the proposed network is significantly improved on the premise that the number of parameters is only increased by 2%.When applied to three publicly available datasets:DRIVE,STARE and CHASE DB1,the proposed network achieves the accuracy/F1-score of 96.33%/84.34%,97.12%/83.17%and 97.06%/84.10%,respectively.The experimental results show that the MIA-UNet is superior to the state-of-the-art methods.