In recent years,fungal diseases affecting grape crops have attracted significant attention.Currently,the assessment of black rot severitymainly depends on the ratio of lesion area to leaf surface area.However,effectiv...In recent years,fungal diseases affecting grape crops have attracted significant attention.Currently,the assessment of black rot severitymainly depends on the ratio of lesion area to leaf surface area.However,effectively and accurately segmenting leaf lesions presents considerable challenges.Existing grape leaf lesion segmentationmodels have several limitations,such as a large number of parameters,long training durations,and limited precision in extracting small lesions and boundary details.To address these issues,we propose an enhanced DeepLabv3+model incorporating Strip Pooling,Content-Guided Fusion,and Convolutional Block Attention Module(SFC_DeepLabv3+),an enhanced lesion segmentation method based on DeepLabv3+.This approach uses the lightweight MobileNetv2 backbone to replace the original Xception,incorporates a lightweight convolutional block attention module,and introduces a content-guided feature fusion module to improve the detection accuracy of small lesions and blurred boundaries.Experimental results showthat the enhancedmodel achieves a mean Intersection overUnion(mIoU)of 90.98%,amean Pixel Accuracy(mPA)of 94.33%,and a precision of 95.84%.This represents relative gains of 2.22%,1.78%,and 0.89%respectively compared to the original model.Additionally,its complexity is significantly reduced without sacrificing performance,the parameter count is reduced to 6.27 M,a decrease of 88.5%compared to the original model,floating point of operations(GFLOPs)drops from 83.62 to 29.00 G,a reduction of 65.1%.Additionally,Frames Per Second(FPS)increases from 63.7 to 74.3 FPS,marking an improvement of 16.7%.Compared to other models,the improved architecture shows faster convergence and superior segmentation accuracy,making it highly suitable for applications in resource-constrained environments.展开更多
Research has been conducted to reduce resource consumption in 3D medical image segmentation for diverse resource-constrained environments.However,decreasing the number of parameters to enhance computational efficiency...Research has been conducted to reduce resource consumption in 3D medical image segmentation for diverse resource-constrained environments.However,decreasing the number of parameters to enhance computational efficiency can also lead to performance degradation.Moreover,these methods face challenges in balancing global and local features,increasing the risk of errors in multi-scale segmentation.This issue is particularly pronounced when segmenting small and complex structures within the human body.To address this problem,we propose a multi-stage hierarchical architecture composed of a detector and a segmentor.The detector extracts regions of interest(ROIs)in a 3D image,while the segmentor performs segmentation in the extracted ROI.Removing unnecessary areas in the detector allows the segmentation to be performed on a more compact input.The segmentor is designed with multiple stages,where each stage utilizes different input sizes.It implements a stage-skippingmechanism that deactivates certain stages using the initial input size.This approach minimizes unnecessary computations on segmenting the essential regions to reduce computational overhead.The proposed framework preserves segmentation performance while reducing resource consumption,enabling segmentation even in resource-constrained environments.展开更多
Brain tumor segmentation from Magnetic Resonance Imaging(MRI)supports neurologists and radiologists in analyzing tumors and developing personalized treatment plans,making it a crucial yet challenging task.Supervised m...Brain tumor segmentation from Magnetic Resonance Imaging(MRI)supports neurologists and radiologists in analyzing tumors and developing personalized treatment plans,making it a crucial yet challenging task.Supervised models such as 3D U-Net perform well in this domain,but their accuracy significantly improves with appropriate preprocessing.This paper demonstrates the effectiveness of preprocessing in brain tumor segmentation by applying a pre-segmentation step based on the Generalized Gaussian Mixture Model(GGMM)to T1 contrastenhanced MRI scans from the BraTS 2020 dataset.The Expectation-Maximization(EM)algorithm is employed to estimate parameters for four tissue classes,generating a new pre-segmented channel that enhances the training and performance of the 3DU-Net model.The proposed GGMM+3D U-Net framework achieved a Dice coefficient of 0.88 for whole tumor segmentation,outperforming both the standard multiscale 3D U-Net(0.84)and MMU-Net(0.85).It also delivered higher Intersection over Union(IoU)scores compared to models trained without preprocessing or with simpler GMM-based segmentation.These results,supported by qualitative visualizations,suggest that GGMM-based preprocessing should be integrated into brain tumor segmentation pipelines to optimize performance.展开更多
Accurate and efficient brain tumor segmentation is essential for early diagnosis,treatment planning,and clinical decision-making.However,the complex structure of brain anatomy and the heterogeneous nature of tumors pr...Accurate and efficient brain tumor segmentation is essential for early diagnosis,treatment planning,and clinical decision-making.However,the complex structure of brain anatomy and the heterogeneous nature of tumors present significant challenges for precise anomaly detection.While U-Net-based architectures have demonstrated strong performance in medical image segmentation,there remains room for improvement in feature extraction and localization accuracy.In this study,we propose a novel hybrid model designed to enhance 3D brain tumor segmentation.The architecture incorporates a 3D ResNet encoder known for mitigating the vanishing gradient problem and a 3D U-Net decoder.Additionally,to enhance the model’s generalization ability,Squeeze and Excitation attention mechanism is integrated.We introduce Gabor filter banks into the encoder to further strengthen the model’s ability to extract robust and transformation-invariant features from the complex and irregular shapes typical in medical imaging.This approach,which is not well explored in current U-Net-based segmentation frameworks,provides a unique advantage by enhancing texture-aware feature representation.Specifically,Gabor filters help extract distinctive low-level texture features,reducing the effects of texture interference and facilitating faster convergence during the early stages of training.Our model achieved Dice scores of 0.881,0.846,and 0.819 for Whole Tumor(WT),Tumor Core(TC),and Enhancing Tumor(ET),respectively,on the BraTS 2020 dataset.Cross-validation on the BraTS 2021 dataset further confirmed the model’s robustness,yielding Dice score values of 0.887 for WT,0.856 for TC,and 0.824 for ET.The proposed model outperforms several state-of-the-art existing models,particularly in accurately identifying small and complex tumor regions.Extensive evaluations suggest integrating advanced preprocessing with an attention-augmented hybrid architecture offers significant potential for reliable and clinically valuable brain tumor segmentation.展开更多
Aiming to address the Unmanned Aerial Vehicle(UAV) formation collision avoidance problem in Three-Dimensional(3-D) low-altitude environments where dense various obstacles exist, a fluid-based path planning framework n...Aiming to address the Unmanned Aerial Vehicle(UAV) formation collision avoidance problem in Three-Dimensional(3-D) low-altitude environments where dense various obstacles exist, a fluid-based path planning framework named the Formation Interfered Fluid Dynamical System(FIFDS) with Moderate Evasive Maneuver Strategy(MEMS) is proposed in this study.First, the UAV formation collision avoidance problem including quantifiable performance indexes is formulated. Second, inspired by the phenomenon of fluids continuously flowing while bypassing objects, the FIFDS for multiple UAVs is presented, which contains a Parallel Streamline Tracking(PST) method for formation keeping and the traditional IFDS for collision avoidance. Third, to rationally balance flight safety and collision avoidance cost, MEMS is proposed to generate moderate evasive maneuvers that match up with collision risks. Comprehensively containing the time and distance safety information, the 3-D dynamic collision regions are modeled for collision prediction. Then, the moderate evasive maneuver principle is refined, which provides criterions of the maneuver amplitude and direction. On this basis, an analytical parameter mapping mechanism is designed to online optimize IFDS parameters. Finally, the performance of the proposed method is validated by comparative simulation results and real flight experiments using fixed-wing UAVs.展开更多
Facial wound segmentation plays a crucial role in preoperative planning and optimizing patient outcomes in various medical applications.In this paper,we propose an efficient approach for automating 3D facial wound seg...Facial wound segmentation plays a crucial role in preoperative planning and optimizing patient outcomes in various medical applications.In this paper,we propose an efficient approach for automating 3D facial wound segmentation using a two-stream graph convolutional network.Our method leverages the Cir3D-FaIR dataset and addresses the challenge of data imbalance through extensive experimentation with different loss functions.To achieve accurate segmentation,we conducted thorough experiments and selected a high-performing model from the trainedmodels.The selectedmodel demonstrates exceptional segmentation performance for complex 3D facial wounds.Furthermore,based on the segmentation model,we propose an improved approach for extracting 3D facial wound fillers and compare it to the results of the previous study.Our method achieved a remarkable accuracy of 0.9999993% on the test suite,surpassing the performance of the previous method.From this result,we use 3D printing technology to illustrate the shape of the wound filling.The outcomes of this study have significant implications for physicians involved in preoperative planning and intervention design.By automating facial wound segmentation and improving the accuracy ofwound-filling extraction,our approach can assist in carefully assessing and optimizing interventions,leading to enhanced patient outcomes.Additionally,it contributes to advancing facial reconstruction techniques by utilizing machine learning and 3D bioprinting for printing skin tissue implants.Our source code is available at https://github.com/SIMOGroup/WoundFilling3D.展开更多
This paper focuses on the effective utilization of data augmentation techniques for 3Dlidar point clouds to enhance the performance of neural network models.These point clouds,which represent spatial information throu...This paper focuses on the effective utilization of data augmentation techniques for 3Dlidar point clouds to enhance the performance of neural network models.These point clouds,which represent spatial information through a collection of 3D coordinates,have found wide-ranging applications.Data augmentation has emerged as a potent solution to the challenges posed by limited labeled data and the need to enhance model generalization capabilities.Much of the existing research is devoted to crafting novel data augmentation methods specifically for 3D lidar point clouds.However,there has been a lack of focus on making the most of the numerous existing augmentation techniques.Addressing this deficiency,this research investigates the possibility of combining two fundamental data augmentation strategies.The paper introduces PolarMix andMix3D,two commonly employed augmentation techniques,and presents a new approach,named RandomFusion.Instead of using a fixed or predetermined combination of augmentation methods,RandomFusion randomly chooses one method from a pool of options for each instance or sample.This innovative data augmentation technique randomly augments each point in the point cloud with either PolarMix or Mix3D.The crux of this strategy is the random choice between PolarMix and Mix3Dfor the augmentation of each point within the point cloud data set.The results of the experiments conducted validate the efficacy of the RandomFusion strategy in enhancing the performance of neural network models for 3D lidar point cloud semantic segmentation tasks.This is achieved without compromising computational efficiency.By examining the potential of merging different augmentation techniques,the research contributes significantly to a more comprehensive understanding of how to utilize existing augmentation methods for 3D lidar point clouds.RandomFusion data augmentation technique offers a simple yet effective method to leverage the diversity of augmentation techniques and boost the robustness of models.The insights gained from this research can pave the way for future work aimed at developing more advanced and efficient data augmentation strategies for 3D lidar point cloud analysis.展开更多
X-ray computed tomography(CT)has been an important technology in paleontology for several decades.It helps researchers to acquire detailed anatomical structures of fossils non-destructively.Despite its widespread appl...X-ray computed tomography(CT)has been an important technology in paleontology for several decades.It helps researchers to acquire detailed anatomical structures of fossils non-destructively.Despite its widespread application,developing an efficient and user-friendly method for segmenting CT data continues to be a formidable challenge in the field.Most CT data segmentation software operates on 2D interfaces,which limits flexibility for real-time adjustments in 3D segmentation.Here,we introduce Curves Mode in Drishti Paint 3.2,an open-source tool for CT data segmentation.Drishti Paint 3.2 allows users to manually or semi-automatically segment the CT data in both 2D and 3D environments,providing a novel solution for revisualizing CT data in paleontological studies.展开更多
In recent years,semantic segmentation on 3D point cloud data has attracted much attention.Unlike 2D images where pixels distribute regularly in the image domain,3D point clouds in non-Euclidean space are irregular and...In recent years,semantic segmentation on 3D point cloud data has attracted much attention.Unlike 2D images where pixels distribute regularly in the image domain,3D point clouds in non-Euclidean space are irregular and inherently sparse.Therefore,it is very difficult to extract long-range contexts and effectively aggregate local features for semantic segmentation in 3D point cloud space.Most current methods either focus on local feature aggregation or long-range context dependency,but fail to directly establish a global-local feature extractor to complete the point cloud semantic segmentation tasks.In this paper,we propose a Transformer-based stratified graph convolutional network(SGT-Net),which enlarges the effective receptive field and builds direct long-range dependency.Specifically,we first propose a novel dense-sparse sampling strategy that provides dense local vertices and sparse long-distance vertices for subsequent graph convolutional network(GCN).Secondly,we propose a multi-key self-attention mechanism based on the Transformer to further weight augmentation for crucial neighboring relationships and enlarge the effective receptive field.In addition,to further improve the efficiency of the network,we propose a similarity measurement module to determine whether the neighborhood near the center point is effective.We demonstrate the validity and superiority of our method on the S3DIS and ShapeNet datasets.Through ablation experiments and segmentation visualization,we verify that the SGT model can improve the performance of the point cloud semantic segmentation.展开更多
3D object recognition is a challenging task for intelligent and robot systems in industrial and home indoor environments.It is critical for such systems to recognize and segment the 3D object instances that they encou...3D object recognition is a challenging task for intelligent and robot systems in industrial and home indoor environments.It is critical for such systems to recognize and segment the 3D object instances that they encounter on a frequent basis.The computer vision,graphics,and machine learning fields have all given it a lot of attention.Traditionally,3D segmentation was done with hand-crafted features and designed approaches that didn’t achieve acceptable performance and couldn’t be generalized to large-scale data.Deep learning approaches have lately become the preferred method for 3D segmentation challenges by their great success in 2D computer vision.However,the task of instance segmentation is currently less explored.In this paper,we propose a novel approach for efficient 3D instance segmentation using red green blue and depth(RGB-D)data based on deep learning.The 2D region based convolutional neural networks(Mask R-CNN)deep learning model with point based rending module is adapted to integrate with depth information to recognize and segment 3D instances of objects.In order to generate 3D point cloud coordinates(x,y,z),segmented 2D pixels(u,v)of recognized object regions in the RGB image are merged into(u,v)points of the depth image.Moreover,we conducted an experiment and analysis to compare our proposed method from various points of view and distances.The experimentation shows the proposed 3D object recognition and instance segmentation are sufficiently beneficial to support object handling in robotic and intelligent systems.展开更多
目的探讨(1-3)-β-D葡聚糖联合降钙素原(procalcitonin,PCT)、CD4^(+)T淋巴细胞多指标在艾滋病患者马尔尼菲篮状菌感染早期诊断临床研究。方法回顾性选取我院2020年1月—2022年6月住院的120例艾滋病患者为研究对象。依据实验室结果,将...目的探讨(1-3)-β-D葡聚糖联合降钙素原(procalcitonin,PCT)、CD4^(+)T淋巴细胞多指标在艾滋病患者马尔尼菲篮状菌感染早期诊断临床研究。方法回顾性选取我院2020年1月—2022年6月住院的120例艾滋病患者为研究对象。依据实验室结果,将其分为马尔尼菲篮状菌感染确诊组(血或组织液培育养出马尔尼菲篮状菌),简称A组(62例),及马尔尼菲篮状菌感染临床诊断组[根据临床症状、体征、血常规及(1-3)-β-D葡聚糖、PCT、CD4^(+)T淋巴细胞多指标诊断],简称B组(58例)。检测患者(1-3)-β-D葡聚糖、PCT、CD4^(+)T淋巴细胞的表达水平,采用受试者工作特征(receiver-operating characteristic,ROC)曲线下面积(area under the curve,AUC)评估上述指标联合检测对艾滋病患者感染马尔尼菲篮状菌的诊断效能。结果A组的(1-3)-β-D葡聚糖和PCT水平均高于B组,CD4^(+)T淋巴细胞个数低于B组(P<0.05);(1-3)-β-D葡聚糖、PCT、CD4^(+)T淋巴细胞联合检测的AUC为0.933,(1-3)-β-D葡聚糖单独检测的AUC是0.812,PCT单独检测的AUC为0.883,CD4^(+)T淋巴细胞单独检测的AUC是0.810,(1-3)-β-D葡聚糖、PCT和CD4^(+)T淋巴细胞联合检测的AUC皆优于三项单独检测,表明(1-3)-β-D葡聚糖、PCT和CD4^(+)T淋巴细胞联合检测的诊断价值皆优于单一指标诊断,且联合检测的特异度、约登指数分别为92.43%和0.580,均高于三项单独检测。结论(1-3)-β-D葡聚糖联合PCT和CD4^(+)T淋巴细胞多指标对艾滋病马尔尼菲篮状菌感染具有非常高的临床诊断价值,能够帮助医生分析出高危风险患者,及时制定治疗方案,同时也承担预后效果的判断依据,对治疗艾滋病马尔尼菲篮状菌感染具有非常重要的研究价值。展开更多
In the study of automatic driving,understanding the road scene is a key to improve driving safety.The semantic segmentation method could divide the image into different areas associated with semantic categories in acc...In the study of automatic driving,understanding the road scene is a key to improve driving safety.The semantic segmentation method could divide the image into different areas associated with semantic categories in accordance with the pixel level,so as to help vehicles to perceive and obtain the surrounding road environment information,which would improve driving safety.Deeplabv3+is the current popular semantic segmentation model.There are phenomena that small targets are missed and similar objects are easily misjudged during its semantic segmentation tasks,which leads to rough segmentation boundary and reduces semantic accuracy.This study focuses on the issue,based on the Deeplabv3+network structure and combined with the attention mechanism,to increase the weight of the segmentation area,and then proposes an improved Deeplabv3+fusion attention mechanism for road scene semantic segmentation method.First,a group of parallel position attention module and channel attention module are introduced on the Deeplabv3+encoding end to capture more spatial context information and high-level semantic information.Then,an attention mechanism is introduced to restore the spatial detail information,and the data shall be normalized in order to accelerate the convergence speed of the model at the decoding end.The effects of model segmentation with different attention-introducing mechanisms are compared and tested on CamVid and Cityscapes datasets.The experimental results show that the mean Intersection over Unons of the improved model segmentation accuracies on the two datasets are boosted by 6.88%and 2.58%,respectively,which is better than using Deeplabv3+.This method does not significantly increase the amount of network calculation and complexity,and has a good balance of speed and accuracy.展开更多
基金supported by the following grants:Zhejiang A&F University Research Development Fund(Talent Initiation Project No.2021LFR048)and 2023 University-Enterprise Joint Research Program(Grant No.LHYFZ2302)from the Modern Agricultural and Forestry Artificial Intelligence Industry Academy.
文摘In recent years,fungal diseases affecting grape crops have attracted significant attention.Currently,the assessment of black rot severitymainly depends on the ratio of lesion area to leaf surface area.However,effectively and accurately segmenting leaf lesions presents considerable challenges.Existing grape leaf lesion segmentationmodels have several limitations,such as a large number of parameters,long training durations,and limited precision in extracting small lesions and boundary details.To address these issues,we propose an enhanced DeepLabv3+model incorporating Strip Pooling,Content-Guided Fusion,and Convolutional Block Attention Module(SFC_DeepLabv3+),an enhanced lesion segmentation method based on DeepLabv3+.This approach uses the lightweight MobileNetv2 backbone to replace the original Xception,incorporates a lightweight convolutional block attention module,and introduces a content-guided feature fusion module to improve the detection accuracy of small lesions and blurred boundaries.Experimental results showthat the enhancedmodel achieves a mean Intersection overUnion(mIoU)of 90.98%,amean Pixel Accuracy(mPA)of 94.33%,and a precision of 95.84%.This represents relative gains of 2.22%,1.78%,and 0.89%respectively compared to the original model.Additionally,its complexity is significantly reduced without sacrificing performance,the parameter count is reduced to 6.27 M,a decrease of 88.5%compared to the original model,floating point of operations(GFLOPs)drops from 83.62 to 29.00 G,a reduction of 65.1%.Additionally,Frames Per Second(FPS)increases from 63.7 to 74.3 FPS,marking an improvement of 16.7%.Compared to other models,the improved architecture shows faster convergence and superior segmentation accuracy,making it highly suitable for applications in resource-constrained environments.
文摘Research has been conducted to reduce resource consumption in 3D medical image segmentation for diverse resource-constrained environments.However,decreasing the number of parameters to enhance computational efficiency can also lead to performance degradation.Moreover,these methods face challenges in balancing global and local features,increasing the risk of errors in multi-scale segmentation.This issue is particularly pronounced when segmenting small and complex structures within the human body.To address this problem,we propose a multi-stage hierarchical architecture composed of a detector and a segmentor.The detector extracts regions of interest(ROIs)in a 3D image,while the segmentor performs segmentation in the extracted ROI.Removing unnecessary areas in the detector allows the segmentation to be performed on a more compact input.The segmentor is designed with multiple stages,where each stage utilizes different input sizes.It implements a stage-skippingmechanism that deactivates certain stages using the initial input size.This approach minimizes unnecessary computations on segmenting the essential regions to reduce computational overhead.The proposed framework preserves segmentation performance while reducing resource consumption,enabling segmentation even in resource-constrained environments.
基金Princess Nourah Bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R826),Princess Nourah Bint Abdulrahman University,Riyadh,Saudi ArabiaNorthern Border University,Saudi Arabia,for supporting this work through project number(NBU-CRP-2025-2933).
文摘Brain tumor segmentation from Magnetic Resonance Imaging(MRI)supports neurologists and radiologists in analyzing tumors and developing personalized treatment plans,making it a crucial yet challenging task.Supervised models such as 3D U-Net perform well in this domain,but their accuracy significantly improves with appropriate preprocessing.This paper demonstrates the effectiveness of preprocessing in brain tumor segmentation by applying a pre-segmentation step based on the Generalized Gaussian Mixture Model(GGMM)to T1 contrastenhanced MRI scans from the BraTS 2020 dataset.The Expectation-Maximization(EM)algorithm is employed to estimate parameters for four tissue classes,generating a new pre-segmented channel that enhances the training and performance of the 3DU-Net model.The proposed GGMM+3D U-Net framework achieved a Dice coefficient of 0.88 for whole tumor segmentation,outperforming both the standard multiscale 3D U-Net(0.84)and MMU-Net(0.85).It also delivered higher Intersection over Union(IoU)scores compared to models trained without preprocessing or with simpler GMM-based segmentation.These results,supported by qualitative visualizations,suggest that GGMM-based preprocessing should be integrated into brain tumor segmentation pipelines to optimize performance.
基金the National Science and Technology Council(NSTC)of the Republic of China,Taiwan,for financially supporting this research under Contract No.NSTC 112-2637-M-131-001.
文摘Accurate and efficient brain tumor segmentation is essential for early diagnosis,treatment planning,and clinical decision-making.However,the complex structure of brain anatomy and the heterogeneous nature of tumors present significant challenges for precise anomaly detection.While U-Net-based architectures have demonstrated strong performance in medical image segmentation,there remains room for improvement in feature extraction and localization accuracy.In this study,we propose a novel hybrid model designed to enhance 3D brain tumor segmentation.The architecture incorporates a 3D ResNet encoder known for mitigating the vanishing gradient problem and a 3D U-Net decoder.Additionally,to enhance the model’s generalization ability,Squeeze and Excitation attention mechanism is integrated.We introduce Gabor filter banks into the encoder to further strengthen the model’s ability to extract robust and transformation-invariant features from the complex and irregular shapes typical in medical imaging.This approach,which is not well explored in current U-Net-based segmentation frameworks,provides a unique advantage by enhancing texture-aware feature representation.Specifically,Gabor filters help extract distinctive low-level texture features,reducing the effects of texture interference and facilitating faster convergence during the early stages of training.Our model achieved Dice scores of 0.881,0.846,and 0.819 for Whole Tumor(WT),Tumor Core(TC),and Enhancing Tumor(ET),respectively,on the BraTS 2020 dataset.Cross-validation on the BraTS 2021 dataset further confirmed the model’s robustness,yielding Dice score values of 0.887 for WT,0.856 for TC,and 0.824 for ET.The proposed model outperforms several state-of-the-art existing models,particularly in accurately identifying small and complex tumor regions.Extensive evaluations suggest integrating advanced preprocessing with an attention-augmented hybrid architecture offers significant potential for reliable and clinically valuable brain tumor segmentation.
基金supported in part by the National Natural Science Foundations of China(Nos.61175084,61673042 and 62203046)the China Postdoctoral Science Foundation(No.2022M713006).
文摘Aiming to address the Unmanned Aerial Vehicle(UAV) formation collision avoidance problem in Three-Dimensional(3-D) low-altitude environments where dense various obstacles exist, a fluid-based path planning framework named the Formation Interfered Fluid Dynamical System(FIFDS) with Moderate Evasive Maneuver Strategy(MEMS) is proposed in this study.First, the UAV formation collision avoidance problem including quantifiable performance indexes is formulated. Second, inspired by the phenomenon of fluids continuously flowing while bypassing objects, the FIFDS for multiple UAVs is presented, which contains a Parallel Streamline Tracking(PST) method for formation keeping and the traditional IFDS for collision avoidance. Third, to rationally balance flight safety and collision avoidance cost, MEMS is proposed to generate moderate evasive maneuvers that match up with collision risks. Comprehensively containing the time and distance safety information, the 3-D dynamic collision regions are modeled for collision prediction. Then, the moderate evasive maneuver principle is refined, which provides criterions of the maneuver amplitude and direction. On this basis, an analytical parameter mapping mechanism is designed to online optimize IFDS parameters. Finally, the performance of the proposed method is validated by comparative simulation results and real flight experiments using fixed-wing UAVs.
文摘Facial wound segmentation plays a crucial role in preoperative planning and optimizing patient outcomes in various medical applications.In this paper,we propose an efficient approach for automating 3D facial wound segmentation using a two-stream graph convolutional network.Our method leverages the Cir3D-FaIR dataset and addresses the challenge of data imbalance through extensive experimentation with different loss functions.To achieve accurate segmentation,we conducted thorough experiments and selected a high-performing model from the trainedmodels.The selectedmodel demonstrates exceptional segmentation performance for complex 3D facial wounds.Furthermore,based on the segmentation model,we propose an improved approach for extracting 3D facial wound fillers and compare it to the results of the previous study.Our method achieved a remarkable accuracy of 0.9999993% on the test suite,surpassing the performance of the previous method.From this result,we use 3D printing technology to illustrate the shape of the wound filling.The outcomes of this study have significant implications for physicians involved in preoperative planning and intervention design.By automating facial wound segmentation and improving the accuracy ofwound-filling extraction,our approach can assist in carefully assessing and optimizing interventions,leading to enhanced patient outcomes.Additionally,it contributes to advancing facial reconstruction techniques by utilizing machine learning and 3D bioprinting for printing skin tissue implants.Our source code is available at https://github.com/SIMOGroup/WoundFilling3D.
基金funded in part by the Key Project of Nature Science Research for Universities of Anhui Province of China(No.2022AH051720)in part by the Science and Technology Development Fund,Macao SAR(Grant Nos.0093/2022/A2,0076/2022/A2 and 0008/2022/AGJ)in part by the China University Industry-University-Research Collaborative Innovation Fund(No.2021FNA04017).
文摘This paper focuses on the effective utilization of data augmentation techniques for 3Dlidar point clouds to enhance the performance of neural network models.These point clouds,which represent spatial information through a collection of 3D coordinates,have found wide-ranging applications.Data augmentation has emerged as a potent solution to the challenges posed by limited labeled data and the need to enhance model generalization capabilities.Much of the existing research is devoted to crafting novel data augmentation methods specifically for 3D lidar point clouds.However,there has been a lack of focus on making the most of the numerous existing augmentation techniques.Addressing this deficiency,this research investigates the possibility of combining two fundamental data augmentation strategies.The paper introduces PolarMix andMix3D,two commonly employed augmentation techniques,and presents a new approach,named RandomFusion.Instead of using a fixed or predetermined combination of augmentation methods,RandomFusion randomly chooses one method from a pool of options for each instance or sample.This innovative data augmentation technique randomly augments each point in the point cloud with either PolarMix or Mix3D.The crux of this strategy is the random choice between PolarMix and Mix3Dfor the augmentation of each point within the point cloud data set.The results of the experiments conducted validate the efficacy of the RandomFusion strategy in enhancing the performance of neural network models for 3D lidar point cloud semantic segmentation tasks.This is achieved without compromising computational efficiency.By examining the potential of merging different augmentation techniques,the research contributes significantly to a more comprehensive understanding of how to utilize existing augmentation methods for 3D lidar point clouds.RandomFusion data augmentation technique offers a simple yet effective method to leverage the diversity of augmentation techniques and boost the robustness of models.The insights gained from this research can pave the way for future work aimed at developing more advanced and efficient data augmentation strategies for 3D lidar point cloud analysis.
文摘X-ray computed tomography(CT)has been an important technology in paleontology for several decades.It helps researchers to acquire detailed anatomical structures of fossils non-destructively.Despite its widespread application,developing an efficient and user-friendly method for segmenting CT data continues to be a formidable challenge in the field.Most CT data segmentation software operates on 2D interfaces,which limits flexibility for real-time adjustments in 3D segmentation.Here,we introduce Curves Mode in Drishti Paint 3.2,an open-source tool for CT data segmentation.Drishti Paint 3.2 allows users to manually or semi-automatically segment the CT data in both 2D and 3D environments,providing a novel solution for revisualizing CT data in paleontological studies.
基金supported in part by the National Natural Science Foundation of China under Grant Nos.U20A20197,62306187the Foundation of Ministry of Industry and Information Technology TC220H05X-04.
文摘In recent years,semantic segmentation on 3D point cloud data has attracted much attention.Unlike 2D images where pixels distribute regularly in the image domain,3D point clouds in non-Euclidean space are irregular and inherently sparse.Therefore,it is very difficult to extract long-range contexts and effectively aggregate local features for semantic segmentation in 3D point cloud space.Most current methods either focus on local feature aggregation or long-range context dependency,but fail to directly establish a global-local feature extractor to complete the point cloud semantic segmentation tasks.In this paper,we propose a Transformer-based stratified graph convolutional network(SGT-Net),which enlarges the effective receptive field and builds direct long-range dependency.Specifically,we first propose a novel dense-sparse sampling strategy that provides dense local vertices and sparse long-distance vertices for subsequent graph convolutional network(GCN).Secondly,we propose a multi-key self-attention mechanism based on the Transformer to further weight augmentation for crucial neighboring relationships and enlarge the effective receptive field.In addition,to further improve the efficiency of the network,we propose a similarity measurement module to determine whether the neighborhood near the center point is effective.We demonstrate the validity and superiority of our method on the S3DIS and ShapeNet datasets.Through ablation experiments and segmentation visualization,we verify that the SGT model can improve the performance of the point cloud semantic segmentation.
文摘3D object recognition is a challenging task for intelligent and robot systems in industrial and home indoor environments.It is critical for such systems to recognize and segment the 3D object instances that they encounter on a frequent basis.The computer vision,graphics,and machine learning fields have all given it a lot of attention.Traditionally,3D segmentation was done with hand-crafted features and designed approaches that didn’t achieve acceptable performance and couldn’t be generalized to large-scale data.Deep learning approaches have lately become the preferred method for 3D segmentation challenges by their great success in 2D computer vision.However,the task of instance segmentation is currently less explored.In this paper,we propose a novel approach for efficient 3D instance segmentation using red green blue and depth(RGB-D)data based on deep learning.The 2D region based convolutional neural networks(Mask R-CNN)deep learning model with point based rending module is adapted to integrate with depth information to recognize and segment 3D instances of objects.In order to generate 3D point cloud coordinates(x,y,z),segmented 2D pixels(u,v)of recognized object regions in the RGB image are merged into(u,v)points of the depth image.Moreover,we conducted an experiment and analysis to compare our proposed method from various points of view and distances.The experimentation shows the proposed 3D object recognition and instance segmentation are sufficiently beneficial to support object handling in robotic and intelligent systems.
文摘目的探讨(1-3)-β-D葡聚糖联合降钙素原(procalcitonin,PCT)、CD4^(+)T淋巴细胞多指标在艾滋病患者马尔尼菲篮状菌感染早期诊断临床研究。方法回顾性选取我院2020年1月—2022年6月住院的120例艾滋病患者为研究对象。依据实验室结果,将其分为马尔尼菲篮状菌感染确诊组(血或组织液培育养出马尔尼菲篮状菌),简称A组(62例),及马尔尼菲篮状菌感染临床诊断组[根据临床症状、体征、血常规及(1-3)-β-D葡聚糖、PCT、CD4^(+)T淋巴细胞多指标诊断],简称B组(58例)。检测患者(1-3)-β-D葡聚糖、PCT、CD4^(+)T淋巴细胞的表达水平,采用受试者工作特征(receiver-operating characteristic,ROC)曲线下面积(area under the curve,AUC)评估上述指标联合检测对艾滋病患者感染马尔尼菲篮状菌的诊断效能。结果A组的(1-3)-β-D葡聚糖和PCT水平均高于B组,CD4^(+)T淋巴细胞个数低于B组(P<0.05);(1-3)-β-D葡聚糖、PCT、CD4^(+)T淋巴细胞联合检测的AUC为0.933,(1-3)-β-D葡聚糖单独检测的AUC是0.812,PCT单独检测的AUC为0.883,CD4^(+)T淋巴细胞单独检测的AUC是0.810,(1-3)-β-D葡聚糖、PCT和CD4^(+)T淋巴细胞联合检测的AUC皆优于三项单独检测,表明(1-3)-β-D葡聚糖、PCT和CD4^(+)T淋巴细胞联合检测的诊断价值皆优于单一指标诊断,且联合检测的特异度、约登指数分别为92.43%和0.580,均高于三项单独检测。结论(1-3)-β-D葡聚糖联合PCT和CD4^(+)T淋巴细胞多指标对艾滋病马尔尼菲篮状菌感染具有非常高的临床诊断价值,能够帮助医生分析出高危风险患者,及时制定治疗方案,同时也承担预后效果的判断依据,对治疗艾滋病马尔尼菲篮状菌感染具有非常重要的研究价值。
基金National Natural Science Foundation of China(Nos.61941109,62061023)Distinguished Young Scholars of Gansu Province of China(No.21JR7RA345)。
文摘In the study of automatic driving,understanding the road scene is a key to improve driving safety.The semantic segmentation method could divide the image into different areas associated with semantic categories in accordance with the pixel level,so as to help vehicles to perceive and obtain the surrounding road environment information,which would improve driving safety.Deeplabv3+is the current popular semantic segmentation model.There are phenomena that small targets are missed and similar objects are easily misjudged during its semantic segmentation tasks,which leads to rough segmentation boundary and reduces semantic accuracy.This study focuses on the issue,based on the Deeplabv3+network structure and combined with the attention mechanism,to increase the weight of the segmentation area,and then proposes an improved Deeplabv3+fusion attention mechanism for road scene semantic segmentation method.First,a group of parallel position attention module and channel attention module are introduced on the Deeplabv3+encoding end to capture more spatial context information and high-level semantic information.Then,an attention mechanism is introduced to restore the spatial detail information,and the data shall be normalized in order to accelerate the convergence speed of the model at the decoding end.The effects of model segmentation with different attention-introducing mechanisms are compared and tested on CamVid and Cityscapes datasets.The experimental results show that the mean Intersection over Unons of the improved model segmentation accuracies on the two datasets are boosted by 6.88%and 2.58%,respectively,which is better than using Deeplabv3+.This method does not significantly increase the amount of network calculation and complexity,and has a good balance of speed and accuracy.