We present Depth-Informed Crop Segmentation(DepthCropSeg),an almost unsupervised crop segmentation approach without manual pixel-level annotations.Crop segmentation is a fundamental vision task in agriculture,which be...We present Depth-Informed Crop Segmentation(DepthCropSeg),an almost unsupervised crop segmentation approach without manual pixel-level annotations.Crop segmentation is a fundamental vision task in agriculture,which benefits a number of downstream applications such as crop growth monitoring and yield estimation.Over the past decade,image-based crop segmentation approaches have shifted from classic color-based paradigms to recent deep learning-based ones.The latter,however,rely heavily on large amounts of data with high-quality manual annotation such that considerable human labor and time are spent.In this work,we leverage Depth Anything V2,a vision foundation model,to produce high-quality pseudo crop masks for training segmentation models.We compile a dataset of 17,199 images from six public plant segmentation sources,generating pseudo masks from depth maps after normalization and thresholding.After a coarse-to-fine manual screening,1378 images with reliable masks are selected.We compare four semantic segmentation models and enhance the top-performing one with depth-informed two-stage self-training and depth-informed post-processing.To evaluate the feasibility and robustness of DepthCropSeg,we benchmark the segmentation performance on 10 public crop segmentation testing sets and a self-collect dataset covering in-field,laboratory,and unmanned aerial vehicle(UAV)scenarios.Experimental results show that our DepthCropSeg approach can achieve crop segmentation performance comparable to the fully supervised model trained with manually annotated data(86.91 vs.87.10).For the first time,we demonstrate almost unsupervised,close-to-full-supervision crop segmentation successfully.展开更多
基金supported by the Chinese Academy of Sciences"Strategic Priority Research Program"Under Grant No.XDA24040201Central Government's Guidance Fund for Local Science and Technology Development Under Grant No.2024ZY-CGZY-19+2 种基金National Natural Science Foundation of China Under Grant No.32370435National Key R&D Program of China Under Grant No.2023YFF1001502Changchun Science and Technology Development Programme Under Grant No.23SH 18.
文摘We present Depth-Informed Crop Segmentation(DepthCropSeg),an almost unsupervised crop segmentation approach without manual pixel-level annotations.Crop segmentation is a fundamental vision task in agriculture,which benefits a number of downstream applications such as crop growth monitoring and yield estimation.Over the past decade,image-based crop segmentation approaches have shifted from classic color-based paradigms to recent deep learning-based ones.The latter,however,rely heavily on large amounts of data with high-quality manual annotation such that considerable human labor and time are spent.In this work,we leverage Depth Anything V2,a vision foundation model,to produce high-quality pseudo crop masks for training segmentation models.We compile a dataset of 17,199 images from six public plant segmentation sources,generating pseudo masks from depth maps after normalization and thresholding.After a coarse-to-fine manual screening,1378 images with reliable masks are selected.We compare four semantic segmentation models and enhance the top-performing one with depth-informed two-stage self-training and depth-informed post-processing.To evaluate the feasibility and robustness of DepthCropSeg,we benchmark the segmentation performance on 10 public crop segmentation testing sets and a self-collect dataset covering in-field,laboratory,and unmanned aerial vehicle(UAV)scenarios.Experimental results show that our DepthCropSeg approach can achieve crop segmentation performance comparable to the fully supervised model trained with manually annotated data(86.91 vs.87.10).For the first time,we demonstrate almost unsupervised,close-to-full-supervision crop segmentation successfully.