Daytime star images captured by dedicated near-space star sensors are characterized by short exposures,high noise,and low Signal-to-Noise Ratios(SNRs).Such imaging is also affected by atmospheric turbulence,causing op...Daytime star images captured by dedicated near-space star sensors are characterized by short exposures,high noise,and low Signal-to-Noise Ratios(SNRs).Such imaging is also affected by atmospheric turbulence,causing optical phenomena,such as scintillation,distortion,and jitter.This causes difficulty in recording high-precision star images during the daytime.This study proposes an adaptive star point extraction method based on dynamically predicting stars'positions.First,it predicts the approximate position of stars based on the star catalog,sensor attitude,observation time,and other information,improving the extraction accuracy.Second,it employs a regional SNR sorting method that adaptively selects star images with higher SNRs,suppressing the scintillation effect and enhancing the SNR of star images.Third,depending on the star's motion trajectory characteristics on the image plane,it utilizes the centroid smoothing method for extraction,thus overcoming the impact of star drift.Field experiments demonstrate that the proposed method can effectively overcome star scintillation,drift,and irregular imaging caused by atmospheric turbulence,achieving a 100%success rate.Moreover,the extraction accuracy improves by more than 80%compared to traditional adaptive methods,attaining a value of 0.05 pixels(0.5"),thereby meeting the requirements of daytime astronomical attitude determination and positioning.展开更多
Feature extraction of signals plays an important role in classification problems because of data dimension reduction property and potential improvement of a classification accuracy rate. Principal component analysis (...Feature extraction of signals plays an important role in classification problems because of data dimension reduction property and potential improvement of a classification accuracy rate. Principal component analysis (PCA), wavelets transform or Fourier transform methods are often used for feature extraction. In this paper, we propose a multi-scale PCA, which combines discrete wavelet transform, and PCA for feature extraction of signals in both the spatial and temporal domains. Our study shows that the multi-scale PCA combined with the proposed new classification methods leads to high classification accuracy for the considered signals.展开更多
Tea leaf picking is a crucial stage in tea production that directly influences the quality and value of the tea.Traditional tea-picking machines may compromise the quality of the tea leaves.High-quality teas are often...Tea leaf picking is a crucial stage in tea production that directly influences the quality and value of the tea.Traditional tea-picking machines may compromise the quality of the tea leaves.High-quality teas are often handpicked and need more delicate operations in intelligent picking machines.Compared with traditional image processing techniques,deep learning models have stronger feature extraction capabilities,and better generalization and are more suitable for practical tea shoot harvesting.However,current research mostly focuses on shoot detection and cannot directly accomplish end-to-end shoot segmentation tasks.We propose a tea shoot instance segmentation model based on multi-scale mixed attention(Mask2FusionNet)using a dataset from the tea garden in Hangzhou.We further analyzed the characteristics of the tea shoot dataset,where the proportion of small to medium-sized targets is 89.9%.Our algorithm is compared with several mainstream object segmentation algorithms,and the results demonstrate that our model achieves an accuracy of 82%in recognizing the tea shoots,showing a better performance compared to other models.Through ablation experiments,we found that ResNet50,PointRend strategy,and the Feature Pyramid Network(FPN)architecture can improve performance by 1.6%,1.4%,and 2.4%,respectively.These experiments demonstrated that our proposed multi-scale and point selection strategy optimizes the feature extraction capability for overlapping small targets.The results indicate that the proposed Mask2FusionNet model can perform the shoot segmentation in unstructured environments,realizing the individual distinction of tea shoots,and complete extraction of the shoot edge contours with a segmentation accuracy of 82.0%.The research results can provide algorithmic support for the segmentation and intelligent harvesting of premium tea shoots at different scales.展开更多
In order to extract the richer feature information of ship targets from sea clutter, and address the high dimensional data problem, a method termed as multi-scale fusion kernel sparse preserving projection(MSFKSPP) ba...In order to extract the richer feature information of ship targets from sea clutter, and address the high dimensional data problem, a method termed as multi-scale fusion kernel sparse preserving projection(MSFKSPP) based on the maximum margin criterion(MMC) is proposed for recognizing the class of ship targets utilizing the high-resolution range profile(HRRP). Multi-scale fusion is introduced to capture the local and detailed information in small-scale features, and the global and contour information in large-scale features, offering help to extract the edge information from sea clutter and further improving the target recognition accuracy. The proposed method can maximally preserve the multi-scale fusion sparse of data and maximize the class separability in the reduced dimensionality by reproducing kernel Hilbert space. Experimental results on the measured radar data show that the proposed method can effectively extract the features of ship target from sea clutter, further reduce the feature dimensionality, and improve target recognition performance.展开更多
In this study,an underwater image enhancement method based on multi-scale adversarial network was proposed to solve the problem of detail blur and color distortion in underwater images.Firstly,the local features of ea...In this study,an underwater image enhancement method based on multi-scale adversarial network was proposed to solve the problem of detail blur and color distortion in underwater images.Firstly,the local features of each layer were enhanced into the global features by the proposed residual dense block,which ensured that the generated images retain more details.Secondly,a multi-scale structure was adopted to extract multi-scale semantic features of the original images.Finally,the features obtained from the dual channels were fused by an adaptive fusion module to further optimize the features.The discriminant network adopted the structure of the Markov discriminator.In addition,by constructing mean square error,structural similarity,and perceived color loss function,the generated image is consistent with the reference image in structure,color,and content.The experimental results showed that the enhanced underwater image deblurring effect of the proposed algorithm was good and the problem of underwater image color bias was effectively improved.In both subjective and objective evaluation indexes,the experimental results of the proposed algorithm are better than those of the comparison algorithm.展开更多
Aiming at the difficulty of fault identification caused by manual extraction of fault features of rotating machinery,a one-dimensional multi-scale convolutional auto-encoder fault diagnosis model is proposed,based on ...Aiming at the difficulty of fault identification caused by manual extraction of fault features of rotating machinery,a one-dimensional multi-scale convolutional auto-encoder fault diagnosis model is proposed,based on the standard convolutional auto-encoder.In this model,the parallel convolutional and deconvolutional kernels of different scales are used to extract the features from the input signal and reconstruct the input signal;then the feature map extracted by multi-scale convolutional kernels is used as the input of the classifier;and finally the parameters of the whole model are fine-tuned using labeled data.Experiments on one set of simulation fault data and two sets of rolling bearing fault data are conducted to validate the proposed method.The results show that the model can achieve 99.75%,99.3%and 100%diagnostic accuracy,respectively.In addition,the diagnostic accuracy and reconstruction error of the one-dimensional multi-scale convolutional auto-encoder are compared with traditional machine learning,convolutional neural networks and a traditional convolutional auto-encoder.The final results show that the proposed model has a better recognition effect for rolling bearing fault data.展开更多
A modified Fourier descriptor was presented. Information from a local space can be used more efficiently. After the boundary pixel set of an object was computed, centroid distance approach was used to compute shape si...A modified Fourier descriptor was presented. Information from a local space can be used more efficiently. After the boundary pixel set of an object was computed, centroid distance approach was used to compute shape signature in the local space. A pair of shape signature and boundary pixel gray was used as a point in a feature space. Then, Fourier transform was used for composition of point information in the feature space so that the shape features could be computed. It is proved theoretically that the shape features from modified Fourier descriptors are invariant to translation, rotation, scaling, and change of start point. It is also testified by measuring the retrieval performance of the systems that the shape features from modified Fourier oescriptors are more discriminative than those from other Fourier descriptors.展开更多
针对合作靶标特征点成像在大跨度动态测量中的特征提取精度下降、实时性不足问题,提出了一种融合帧间运动预测与改进亚像素边缘检测的自适应光斑质心提取方法。基于合作靶标测量运动连贯性特性,构建了动态感兴趣区域(Region of Interest...针对合作靶标特征点成像在大跨度动态测量中的特征提取精度下降、实时性不足问题,提出了一种融合帧间运动预测与改进亚像素边缘检测的自适应光斑质心提取方法。基于合作靶标测量运动连贯性特性,构建了动态感兴趣区域(Region of Interest,ROI)特征参数模型,以帧间运动预测实现ROI的快速定位,结合大律法阈值优化策略实现自适应Canny边缘检测,在提升计算效率的同时有效解决了不同测量距离下的降噪问题。然后,采用多方向Sobel算子与强度斜坡改进的Zernike矩相结合改进了边缘点定位算法,并基于高斯牛顿迭代改进鲁棒最小二乘圆拟合法,实现质心坐标计算。实验结果表明:在仿真测试中,本方法在不同噪声水平下的质心定位误差为0.001~0.025像素;实际测试中,ROI预测算法可满足加速度8.75 m/s^(2)以内的测量场景需求,10~30 m测量距离内的光斑重复性定位误差稳定在0.016~0.040像素,优于传统方法;光斑提取速度提升约75.5%,显著增强了系统的实时处理能力。本研究可为合作靶标的测量应用提供有效技术保障。展开更多
In the field of Weakly Supervised Semantic Segmentation(WSSS),methods based on image-level annotation face challenges in accurately capturing objects of varying sizes,lacking sensitivity to image details,and having hi...In the field of Weakly Supervised Semantic Segmentation(WSSS),methods based on image-level annotation face challenges in accurately capturing objects of varying sizes,lacking sensitivity to image details,and having high computational costs.To address these issues,we improve the dual-branch architecture of the Conformer as the fundamental network for generating class activation graphs,proposing a multi-scale efficient weakly-supervised semantic segmentation method based on the improved Conformer.In the Convolution Neural Network(CNN)branch,a cross-scale feature integration convolution module is designed,incorporating multi-receptive field convolution layers to enhance the model’s ability to capture long-range dependencies and improve sensitivity to multi-scale objects.In the Vision Transformer(ViT)branch,an efficient multi-head self-attention module is developed,reducing unnecessary computation through spatial compression and feature partitioning,thereby improving overall network efficiency.Finally,a multi-feature coupling module is introduced to complement the features generated by both branches.This design retains the strength of Convolution Neural Network in extracting local details while harnessing the strength of Vision Transformer to capture comprehensive global features.Experimental results show that the mean Intersection over Union of the image segmentation results of the proposed method on the validation and test sets of the PASCAL VOC 2012 datasets are improved by 2.9%and 3.6%,respectively,over the TransCAM algorithm.Besides,the improved model demonstrates a 1.3%increase of the mean Intersections over Union on the COCO 2014 datasets.Additionally,the number of parameters and the floating-point operations are reduced by 16.2%and 12.9%.However,the proposed method still has limitations of poor performance when dealing with complex scenarios.There is a need for further enhancing the performance of this method to address this issue.展开更多
Semantic segmentation in street scenes is a crucial technology for autonomous driving to analyze the surrounding environment.In street scenes,issues such as high image resolution caused by a large viewpoints and diffe...Semantic segmentation in street scenes is a crucial technology for autonomous driving to analyze the surrounding environment.In street scenes,issues such as high image resolution caused by a large viewpoints and differences in object scales lead to a decline in real-time performance and difficulties in multi-scale feature extraction.To address this,we propose a bilateral-branch real-time semantic segmentationmethod based on semantic information distillation(BSDNet)for street scene images.The BSDNet consists of a Feature Conversion Convolutional Block(FCB),a Semantic Information Distillation Module(SIDM),and a Deep Aggregation Atrous Convolution Pyramid Pooling(DASP).FCB reduces the semantic gap between the backbone and the semantic branch.SIDM extracts high-quality semantic information fromthe Transformer branch to reduce computational costs.DASP aggregates information lost in atrous convolutions,effectively capturingmulti-scale objects.Extensive experiments conducted on Cityscapes,CamVid,and ADE20K,achieving an accuracy of 81.7% Mean Intersection over Union(mIoU)at 70.6 Frames Per Second(FPS)on Cityscapes,demonstrate that our method achieves a better balance between accuracy and inference speed.展开更多
Neuronal soma segmentation plays a crucial role in neuroscience applications.However,the fine structure,such as boundaries,small-volume neuronal somata and fibers,are commonly present in cell images,which pose a chall...Neuronal soma segmentation plays a crucial role in neuroscience applications.However,the fine structure,such as boundaries,small-volume neuronal somata and fibers,are commonly present in cell images,which pose a challenge for accurate segmentation.In this paper,we propose a 3D semantic segmentation network for neuronal soma segmentation to address this issue.Using an encoding-decoding structure,we introduce a Multi-Scale feature extraction and Adaptive Weighting fusion module(MSAW)after each encoding block.The MSAW module can not only emphasize the fine structures via an upsampling strategy,but also provide pixel-wise weights to measure the importance of the multi-scale features.Additionally,a dynamic convolution instead of normal convolution is employed to better adapt the network to input data with different distributions.The proposed MSAW-based semantic segmentation network(MSAW-Net)was evaluated on three neuronal soma images from mouse brain and one neuronal soma image from macaque brain,demonstrating the efficiency of the proposed method.It achieved an F1 score of 91.8%on Fezf2-2A-CreER dataset,97.1%on LSL-H2B-GFP dataset,82.8%on Thy1-EGFP-Mline dataset,and 86.9%on macaque dataset,achieving improvements over the 3D U-Net model by 3.1%,3.3%,3.9%,and 2.3%,respectively.展开更多
Current aero-engine life prediction areas typically focus on single-scale degradation features,and the existing methods are not comprehensive enough to capture the relationship within time series data.To address this ...Current aero-engine life prediction areas typically focus on single-scale degradation features,and the existing methods are not comprehensive enough to capture the relationship within time series data.To address this problem,we propose a novel remaining useful life(RUL)estimation method based on the attention mechanism.Our approach designs a two-layer multi-scale feature extraction module that integrates degradation features at different scales.These features are then processed in parallel by a self-attention module and a three-layer long short-term memory(LSTM)network,which together capture long-term dependencies and adaptively weigh important feature.The integration of degradation patterns from both components into the attention module enhances the model’s ability to capture long-term dependencies.Visualizing the attention module’s weight matrices further improves model interpretability.Experimental results on the C-MAPSS dataset demonstrate that our approach outperforms the existing state-of-the-art methods.展开更多
基金funded by the National Natural Science Foundation of China(Nos.42374011,42074013)through the Natural Science Foundation’s Outstanding Youth Fund Program of Henan Province,China(Nos.242300421150,242300421151)。
文摘Daytime star images captured by dedicated near-space star sensors are characterized by short exposures,high noise,and low Signal-to-Noise Ratios(SNRs).Such imaging is also affected by atmospheric turbulence,causing optical phenomena,such as scintillation,distortion,and jitter.This causes difficulty in recording high-precision star images during the daytime.This study proposes an adaptive star point extraction method based on dynamically predicting stars'positions.First,it predicts the approximate position of stars based on the star catalog,sensor attitude,observation time,and other information,improving the extraction accuracy.Second,it employs a regional SNR sorting method that adaptively selects star images with higher SNRs,suppressing the scintillation effect and enhancing the SNR of star images.Third,depending on the star's motion trajectory characteristics on the image plane,it utilizes the centroid smoothing method for extraction,thus overcoming the impact of star drift.Field experiments demonstrate that the proposed method can effectively overcome star scintillation,drift,and irregular imaging caused by atmospheric turbulence,achieving a 100%success rate.Moreover,the extraction accuracy improves by more than 80%compared to traditional adaptive methods,attaining a value of 0.05 pixels(0.5"),thereby meeting the requirements of daytime astronomical attitude determination and positioning.
文摘Feature extraction of signals plays an important role in classification problems because of data dimension reduction property and potential improvement of a classification accuracy rate. Principal component analysis (PCA), wavelets transform or Fourier transform methods are often used for feature extraction. In this paper, we propose a multi-scale PCA, which combines discrete wavelet transform, and PCA for feature extraction of signals in both the spatial and temporal domains. Our study shows that the multi-scale PCA combined with the proposed new classification methods leads to high classification accuracy for the considered signals.
基金This research was supported by the National Natural Science Foundation of China No.62276086the National Key R&D Program of China No.2022YFD2000100Zhejiang Provincial Natural Science Foundation of China under Grant No.LTGN23D010002.
文摘Tea leaf picking is a crucial stage in tea production that directly influences the quality and value of the tea.Traditional tea-picking machines may compromise the quality of the tea leaves.High-quality teas are often handpicked and need more delicate operations in intelligent picking machines.Compared with traditional image processing techniques,deep learning models have stronger feature extraction capabilities,and better generalization and are more suitable for practical tea shoot harvesting.However,current research mostly focuses on shoot detection and cannot directly accomplish end-to-end shoot segmentation tasks.We propose a tea shoot instance segmentation model based on multi-scale mixed attention(Mask2FusionNet)using a dataset from the tea garden in Hangzhou.We further analyzed the characteristics of the tea shoot dataset,where the proportion of small to medium-sized targets is 89.9%.Our algorithm is compared with several mainstream object segmentation algorithms,and the results demonstrate that our model achieves an accuracy of 82%in recognizing the tea shoots,showing a better performance compared to other models.Through ablation experiments,we found that ResNet50,PointRend strategy,and the Feature Pyramid Network(FPN)architecture can improve performance by 1.6%,1.4%,and 2.4%,respectively.These experiments demonstrated that our proposed multi-scale and point selection strategy optimizes the feature extraction capability for overlapping small targets.The results indicate that the proposed Mask2FusionNet model can perform the shoot segmentation in unstructured environments,realizing the individual distinction of tea shoots,and complete extraction of the shoot edge contours with a segmentation accuracy of 82.0%.The research results can provide algorithmic support for the segmentation and intelligent harvesting of premium tea shoots at different scales.
基金supported by the National Natural Science Foundation of China (62271255,61871218)the Fundamental Research Funds for the Central University (3082019NC2019002)+1 种基金the Aeronautical Science Foundation (ASFC-201920007002)the Program of Remote Sensing Intelligent Monitoring and Emergency Services for Regional Security Elements。
文摘In order to extract the richer feature information of ship targets from sea clutter, and address the high dimensional data problem, a method termed as multi-scale fusion kernel sparse preserving projection(MSFKSPP) based on the maximum margin criterion(MMC) is proposed for recognizing the class of ship targets utilizing the high-resolution range profile(HRRP). Multi-scale fusion is introduced to capture the local and detailed information in small-scale features, and the global and contour information in large-scale features, offering help to extract the edge information from sea clutter and further improving the target recognition accuracy. The proposed method can maximally preserve the multi-scale fusion sparse of data and maximize the class separability in the reduced dimensionality by reproducing kernel Hilbert space. Experimental results on the measured radar data show that the proposed method can effectively extract the features of ship target from sea clutter, further reduce the feature dimensionality, and improve target recognition performance.
文摘In this study,an underwater image enhancement method based on multi-scale adversarial network was proposed to solve the problem of detail blur and color distortion in underwater images.Firstly,the local features of each layer were enhanced into the global features by the proposed residual dense block,which ensured that the generated images retain more details.Secondly,a multi-scale structure was adopted to extract multi-scale semantic features of the original images.Finally,the features obtained from the dual channels were fused by an adaptive fusion module to further optimize the features.The discriminant network adopted the structure of the Markov discriminator.In addition,by constructing mean square error,structural similarity,and perceived color loss function,the generated image is consistent with the reference image in structure,color,and content.The experimental results showed that the enhanced underwater image deblurring effect of the proposed algorithm was good and the problem of underwater image color bias was effectively improved.In both subjective and objective evaluation indexes,the experimental results of the proposed algorithm are better than those of the comparison algorithm.
基金The National Natural Science Foundation of China(No.51675098)
文摘Aiming at the difficulty of fault identification caused by manual extraction of fault features of rotating machinery,a one-dimensional multi-scale convolutional auto-encoder fault diagnosis model is proposed,based on the standard convolutional auto-encoder.In this model,the parallel convolutional and deconvolutional kernels of different scales are used to extract the features from the input signal and reconstruct the input signal;then the feature map extracted by multi-scale convolutional kernels is used as the input of the classifier;and finally the parameters of the whole model are fine-tuned using labeled data.Experiments on one set of simulation fault data and two sets of rolling bearing fault data are conducted to validate the proposed method.The results show that the model can achieve 99.75%,99.3%and 100%diagnostic accuracy,respectively.In addition,the diagnostic accuracy and reconstruction error of the one-dimensional multi-scale convolutional auto-encoder are compared with traditional machine learning,convolutional neural networks and a traditional convolutional auto-encoder.The final results show that the proposed model has a better recognition effect for rolling bearing fault data.
基金Project(60873010)supported by the National Natural Science Foundation of ChinaProject supported by the Doctor Startup Foundation of Shenyang University of Technology,China
文摘A modified Fourier descriptor was presented. Information from a local space can be used more efficiently. After the boundary pixel set of an object was computed, centroid distance approach was used to compute shape signature in the local space. A pair of shape signature and boundary pixel gray was used as a point in a feature space. Then, Fourier transform was used for composition of point information in the feature space so that the shape features could be computed. It is proved theoretically that the shape features from modified Fourier descriptors are invariant to translation, rotation, scaling, and change of start point. It is also testified by measuring the retrieval performance of the systems that the shape features from modified Fourier oescriptors are more discriminative than those from other Fourier descriptors.
文摘针对合作靶标特征点成像在大跨度动态测量中的特征提取精度下降、实时性不足问题,提出了一种融合帧间运动预测与改进亚像素边缘检测的自适应光斑质心提取方法。基于合作靶标测量运动连贯性特性,构建了动态感兴趣区域(Region of Interest,ROI)特征参数模型,以帧间运动预测实现ROI的快速定位,结合大律法阈值优化策略实现自适应Canny边缘检测,在提升计算效率的同时有效解决了不同测量距离下的降噪问题。然后,采用多方向Sobel算子与强度斜坡改进的Zernike矩相结合改进了边缘点定位算法,并基于高斯牛顿迭代改进鲁棒最小二乘圆拟合法,实现质心坐标计算。实验结果表明:在仿真测试中,本方法在不同噪声水平下的质心定位误差为0.001~0.025像素;实际测试中,ROI预测算法可满足加速度8.75 m/s^(2)以内的测量场景需求,10~30 m测量距离内的光斑重复性定位误差稳定在0.016~0.040像素,优于传统方法;光斑提取速度提升约75.5%,显著增强了系统的实时处理能力。本研究可为合作靶标的测量应用提供有效技术保障。
文摘In the field of Weakly Supervised Semantic Segmentation(WSSS),methods based on image-level annotation face challenges in accurately capturing objects of varying sizes,lacking sensitivity to image details,and having high computational costs.To address these issues,we improve the dual-branch architecture of the Conformer as the fundamental network for generating class activation graphs,proposing a multi-scale efficient weakly-supervised semantic segmentation method based on the improved Conformer.In the Convolution Neural Network(CNN)branch,a cross-scale feature integration convolution module is designed,incorporating multi-receptive field convolution layers to enhance the model’s ability to capture long-range dependencies and improve sensitivity to multi-scale objects.In the Vision Transformer(ViT)branch,an efficient multi-head self-attention module is developed,reducing unnecessary computation through spatial compression and feature partitioning,thereby improving overall network efficiency.Finally,a multi-feature coupling module is introduced to complement the features generated by both branches.This design retains the strength of Convolution Neural Network in extracting local details while harnessing the strength of Vision Transformer to capture comprehensive global features.Experimental results show that the mean Intersection over Union of the image segmentation results of the proposed method on the validation and test sets of the PASCAL VOC 2012 datasets are improved by 2.9%and 3.6%,respectively,over the TransCAM algorithm.Besides,the improved model demonstrates a 1.3%increase of the mean Intersections over Union on the COCO 2014 datasets.Additionally,the number of parameters and the floating-point operations are reduced by 16.2%and 12.9%.However,the proposed method still has limitations of poor performance when dealing with complex scenarios.There is a need for further enhancing the performance of this method to address this issue.
基金supported in part by the National Natural Science Foundation of China[Grant number 62471075]the Major Science and Technology Project Grant of the Chongqing Municipal Education Commission[Grant number KJZD-M202301901]Graduate Innovation Fund of Chongqing[gzlcx20253235].
文摘Semantic segmentation in street scenes is a crucial technology for autonomous driving to analyze the surrounding environment.In street scenes,issues such as high image resolution caused by a large viewpoints and differences in object scales lead to a decline in real-time performance and difficulties in multi-scale feature extraction.To address this,we propose a bilateral-branch real-time semantic segmentationmethod based on semantic information distillation(BSDNet)for street scene images.The BSDNet consists of a Feature Conversion Convolutional Block(FCB),a Semantic Information Distillation Module(SIDM),and a Deep Aggregation Atrous Convolution Pyramid Pooling(DASP).FCB reduces the semantic gap between the backbone and the semantic branch.SIDM extracts high-quality semantic information fromthe Transformer branch to reduce computational costs.DASP aggregates information lost in atrous convolutions,effectively capturingmulti-scale objects.Extensive experiments conducted on Cityscapes,CamVid,and ADE20K,achieving an accuracy of 81.7% Mean Intersection over Union(mIoU)at 70.6 Frames Per Second(FPS)on Cityscapes,demonstrate that our method achieves a better balance between accuracy and inference speed.
基金supported by the STI2030-Major-Projects(No.2021ZD0200104)the National Natural Science Foundations of China under Grant 61771437.
文摘Neuronal soma segmentation plays a crucial role in neuroscience applications.However,the fine structure,such as boundaries,small-volume neuronal somata and fibers,are commonly present in cell images,which pose a challenge for accurate segmentation.In this paper,we propose a 3D semantic segmentation network for neuronal soma segmentation to address this issue.Using an encoding-decoding structure,we introduce a Multi-Scale feature extraction and Adaptive Weighting fusion module(MSAW)after each encoding block.The MSAW module can not only emphasize the fine structures via an upsampling strategy,but also provide pixel-wise weights to measure the importance of the multi-scale features.Additionally,a dynamic convolution instead of normal convolution is employed to better adapt the network to input data with different distributions.The proposed MSAW-based semantic segmentation network(MSAW-Net)was evaluated on three neuronal soma images from mouse brain and one neuronal soma image from macaque brain,demonstrating the efficiency of the proposed method.It achieved an F1 score of 91.8%on Fezf2-2A-CreER dataset,97.1%on LSL-H2B-GFP dataset,82.8%on Thy1-EGFP-Mline dataset,and 86.9%on macaque dataset,achieving improvements over the 3D U-Net model by 3.1%,3.3%,3.9%,and 2.3%,respectively.
基金supported by the National Key Research and Development Program of China (2023YFB4302403)the Research and Practical Innovation Program of NUAA (xcxjh20230735)。
文摘Current aero-engine life prediction areas typically focus on single-scale degradation features,and the existing methods are not comprehensive enough to capture the relationship within time series data.To address this problem,we propose a novel remaining useful life(RUL)estimation method based on the attention mechanism.Our approach designs a two-layer multi-scale feature extraction module that integrates degradation features at different scales.These features are then processed in parallel by a self-attention module and a three-layer long short-term memory(LSTM)network,which together capture long-term dependencies and adaptively weigh important feature.The integration of degradation patterns from both components into the attention module enhances the model’s ability to capture long-term dependencies.Visualizing the attention module’s weight matrices further improves model interpretability.Experimental results on the C-MAPSS dataset demonstrate that our approach outperforms the existing state-of-the-art methods.