Satellite image segmentation plays a crucial role in remote sensing,supporting applications such as environmental monitoring,land use analysis,and disaster management.However,traditional segmentation methods often rel...Satellite image segmentation plays a crucial role in remote sensing,supporting applications such as environmental monitoring,land use analysis,and disaster management.However,traditional segmentation methods often rely on large amounts of labeled data,which are costly and time-consuming to obtain,especially in largescale or dynamic environments.To address this challenge,we propose the Semi-Supervised Multi-View Picture Fuzzy Clustering(SS-MPFC)algorithm,which improves segmentation accuracy and robustness,particularly in complex and uncertain remote sensing scenarios.SS-MPFC unifies three paradigms:semi-supervised learning,multi-view clustering,and picture fuzzy set theory.This integration allows the model to effectively utilize a small number of labeled samples,fuse complementary information from multiple data views,and handle the ambiguity and uncertainty inherent in satellite imagery.We design a novel objective function that jointly incorporates picture fuzzy membership functions across multiple views of the data,and embeds pairwise semi-supervised constraints(must-link and cannot-link)directly into the clustering process to enhance segmentation accuracy.Experiments conducted on several benchmark satellite datasets demonstrate that SS-MPFC significantly outperforms existing state-of-the-art methods in segmentation accuracy,noise robustness,and semantic interpretability.On the Augsburg dataset,SS-MPFC achieves a Purity of 0.8158 and an Accuracy of 0.6860,highlighting its outstanding robustness and efficiency.These results demonstrate that SSMPFC offers a scalable and effective solution for real-world satellite-based monitoring systems,particularly in scenarios where rapid annotation is infeasible,such as wildfire tracking,agricultural monitoring,and dynamic urban mapping.展开更多
Medical image segmentation is of critical importance in the domain of contemporary medical imaging.However,U-Net and its variants exhibit limitations in capturing complex nonlinear patterns and global contextual infor...Medical image segmentation is of critical importance in the domain of contemporary medical imaging.However,U-Net and its variants exhibit limitations in capturing complex nonlinear patterns and global contextual information.Although the subsequent U-KAN model enhances nonlinear representation capabilities,it still faces challenges such as gradient vanishing during deep network training and spatial detail loss during feature downsampling,resulting in insufficient segmentation accuracy for edge structures and minute lesions.To address these challenges,this paper proposes the RE-UKAN model,which innovatively improves upon U-KAN.Firstly,a residual network is introduced into the encoder to effectively mitigate gradient vanishing through cross-layer identity mappings,thus enhancing modelling capabilities for complex pathological structures.Secondly,Efficient Local Attention(ELA)is integrated to suppress spatial detail loss during downsampling,thereby improving the perception of edge structures and minute lesions.Experimental results on four public datasets demonstrate that RE-UKAN outperforms existing medical image segmentation methods across multiple evaluation metrics,with particularly outstanding performance on the TN-SCUI 2020 dataset,achieving IoU of 88.18%and Dice of 93.57%.Compared to the baseline model,it achieves improvements of 3.05%and 1.72%,respectively.These results fully demonstrate RE-UKAN’s superior detail retention capability and boundary recognition accuracy in complex medical image segmentation tasks,providing a reliable solution for clinical precision segmentation.展开更多
This article studies the problem of image segmentation-based semantic communication in autonomous driving.In real traffic scenes,the detecting of objects(e.g.,vehicles and pedestrians)is more important to guarantee dr...This article studies the problem of image segmentation-based semantic communication in autonomous driving.In real traffic scenes,the detecting of objects(e.g.,vehicles and pedestrians)is more important to guarantee driving safety,which is always ignored in existing works.Therefore,we propose a vehicular image segmentation-oriented semantic communication system,termed VIS-SemCom,focusing on transmitting and recovering image semantic features of high-important objects to reduce transmission redundancy.First,we develop a semantic codec based on Swin Transformer architecture,which expands the perceptual field thus improving the segmentation accuracy.To highlight the important objects'accuracy,we propose a multi-scale semantic extraction method by assigning the number of Swin Transformer blocks for diverse resolution semantic features.Also,an importance-aware loss incorporating important levels is devised,and an online hard example mining(OHEM)strategy is proposed to handle small sample issues in the dataset.Finally,experimental results demonstrate that the proposed VIS-SemCom can achieve a significant mean intersection over union(mIoU)performance in the SNR regions,a reduction of transmitted data volume by about 60%at 60%mIoU,and improve the segmentation accuracy of important objects,compared to baseline image communication.展开更多
Convolutional neural networks(CNNs)-based medical image segmentation technologies have been widely used in medical image segmentation because of their strong representation and generalization abilities.However,due to ...Convolutional neural networks(CNNs)-based medical image segmentation technologies have been widely used in medical image segmentation because of their strong representation and generalization abilities.However,due to the inability to effectively capture global information from images,CNNs can easily lead to loss of contours and textures in segmentation results.Notice that the transformer model can effectively capture the properties of long-range dependencies in the image,and furthermore,combining the CNN and the transformer can effectively extract local details and global contextual features of the image.Motivated by this,we propose a multi-branch and multi-scale attention network(M2ANet)for medical image segmentation,whose architecture consists of three components.Specifically,in the first component,we construct an adaptive multi-branch patch module for parallel extraction of image features to reduce information loss caused by downsampling.In the second component,we apply residual block to the well-known convolutional block attention module to enhance the network’s ability to recognize important features of images and alleviate the phenomenon of gradient vanishing.In the third component,we design a multi-scale feature fusion module,in which we adopt adaptive average pooling and position encoding to enhance contextual features,and then multi-head attention is introduced to further enrich feature representation.Finally,we validate the effectiveness and feasibility of the proposed M2ANet method through comparative experiments on four benchmark medical image segmentation datasets,particularly in the context of preserving contours and textures.展开更多
Convolutional neural network(CNN)with the encoder-decoder structure is popular in medical image segmentation due to its excellent local feature extraction ability but it faces limitations in capturing the global featu...Convolutional neural network(CNN)with the encoder-decoder structure is popular in medical image segmentation due to its excellent local feature extraction ability but it faces limitations in capturing the global feature.The transformer can extract the global information well but adapting it to small medical datasets is challenging and its computational complexity can be heavy.In this work,a serial and parallel network is proposed for the accurate 3D medical image segmentation by combining CNN and transformer and promoting feature interactions across various semantic levels.The core components of the proposed method include the cross window self-attention based transformer(CWST)and multi-scale local enhanced(MLE)modules.The CWST module enhances the global context understanding by partitioning 3D images into non-overlapping windows and calculating sparse global attention between windows.The MLE module selectively fuses features by computing the voxel attention between different branch features,and uses convolution to strengthen the dense local information.The experiments on the prostate,atrium,and pancreas MR/CT image datasets consistently demonstrate the advantage of the proposed method over six popular segmentation models in both qualitative evaluation and quantitative indexes such as dice similarity coefficient,Intersection over Union,95%Hausdorff distance and average symmetric surface distance.展开更多
Existing semi-supervisedmedical image segmentation algorithms use copy-paste data augmentation to correct the labeled-unlabeled data distribution mismatch.However,current copy-paste methods have three limitations:(1)t...Existing semi-supervisedmedical image segmentation algorithms use copy-paste data augmentation to correct the labeled-unlabeled data distribution mismatch.However,current copy-paste methods have three limitations:(1)training the model solely with copy-paste mixed pictures from labeled and unlabeled input loses a lot of labeled information;(2)low-quality pseudo-labels can cause confirmation bias in pseudo-supervised learning on unlabeled data;(3)the segmentation performance in low-contrast and local regions is less than optimal.We design a Stochastic Augmentation-Based Dual-Teaching Auxiliary Training Strategy(SADT),which enhances feature diversity and learns high-quality features to overcome these problems.To be more precise,SADT trains the Student Network by using pseudo-label-based training from Teacher Network 1 and supervised learning with labeled data,which prevents the loss of rare labeled data.We introduce a bi-directional copy-pastemask with progressive high-entropy filtering to reduce data distribution disparities and mitigate confirmation bias in pseudo-supervision.For the mixed images,Deep-Shallow Spatial Contrastive Learning(DSSCL)is proposed in the feature spaces of Teacher Network 2 and the Student Network to improve the segmentation capabilities in low-contrast and local areas.In this procedure,the features retrieved by the Student Network are subjected to a random feature perturbation technique.On two openly available datasets,extensive trials show that our proposed SADT performs much better than the state-ofthe-art semi-supervised medical segmentation techniques.Using only 10%of the labeled data for training,SADT was able to acquire a Dice score of 90.10%on the ACDC(Automatic Cardiac Diagnosis Challenge)dataset.展开更多
Accurate medical image segmentation plays a crucial role in improving the precision of computer-aided diagnosis.However,complex boundary shapes,low contrast and blurred anatomical structures make fine-grained segmenta...Accurate medical image segmentation plays a crucial role in improving the precision of computer-aided diagnosis.However,complex boundary shapes,low contrast and blurred anatomical structures make fine-grained segmentation a challenging task.Variational Bayesian inference quantifies uncertainty through probability distributions and can construct robust probabilistic models for the boundaries of ambiguous organs and tissues.In this paper,we apply variational Bayesian inference to medical image segmentation and propose variational attention to model the uncertainty of low-contrast and blurry tissue and organ boundaries.This enhances the model's ability to perceive segmentation boundaries,improving robustness and segmentation accuracy.Variational attention first estimates the parameters of the probability distribution of latent representations based on input features.Then,it samples latent representations from the learnt distribution to generate attention weights that optimise the interaction between global features and ambiguous boundaries.We integrate variational attention into the U-Net model by replacing its skip connections,constructing a multi-scale variational attention segmentation model(V-UNet).Experiments on the ISBI 2012 and MoNuSeg 2018 datasets show that our method achieves Dice scores of 95.89%and 82.18%,respectively.Moreover,we integrate V-UNet into the Mask R-CNN framework by replacing the FPN feature extraction head and propose a two-stage segmentation method.Compared to the original Mask R-CNN,our method improves the Dice score by 0.81%,mAP by 8.06%and F1 score by 0.51%.展开更多
Medical image segmentation has become a cornerstone for many healthcare applications,allowing for the automated extraction of critical information from images such as Computed Tomography(CT)scans,Magnetic Resonance Im...Medical image segmentation has become a cornerstone for many healthcare applications,allowing for the automated extraction of critical information from images such as Computed Tomography(CT)scans,Magnetic Resonance Imaging(MRIs),and X-rays.The introduction of U-Net in 2015 has significantly advanced segmentation capabilities,especially for small datasets commonly found in medical imaging.Since then,various modifications to the original U-Net architecture have been proposed to enhance segmentation accuracy and tackle challenges like class imbalance,data scarcity,and multi-modal image processing.This paper provides a detailed review and comparison of several U-Net-based architectures,focusing on their effectiveness in medical image segmentation tasks.We evaluate performance metrics such as Dice Similarity Coefficient(DSC)and Intersection over Union(IoU)across different U-Net variants including HmsU-Net,CrossU-Net,mResU-Net,and others.Our results indicate that architectural enhancements such as transformers,attention mechanisms,and residual connections improve segmentation performance across diverse medical imaging applications,including tumor detection,organ segmentation,and lesion identification.The study also identifies current challenges in the field,including data variability,limited dataset sizes,and issues with class imbalance.Based on these findings,the paper suggests potential future directions for improving the robustness and clinical applicability of U-Net-based models in medical image segmentation.展开更多
Medical image segmentation,i.e.,labeling structures of interest in medical images,is crucial for disease diagnosis and treatment in radiology.In reversible data hiding in medical images(RDHMI),segmentation consists of...Medical image segmentation,i.e.,labeling structures of interest in medical images,is crucial for disease diagnosis and treatment in radiology.In reversible data hiding in medical images(RDHMI),segmentation consists of only two regions:the focal and nonfocal regions.The focal region mainly contains information for diagnosis,while the nonfocal region serves as the monochrome background.The current traditional segmentation methods utilized in RDHMI are inaccurate for complex medical images,and manual segmentation is time-consuming,poorly reproducible,and operator-dependent.Implementing state-of-the-art deep learning(DL)models will facilitate key benefits,but the lack of domain-specific labels for existing medical datasets makes it impossible.To address this problem,this study provides labels of existing medical datasets based on a hybrid segmentation approach to facilitate the implementation of DL segmentation models in this domain.First,an initial segmentation based on a 33 kernel is performed to analyze×identified contour pixels before classifying pixels into focal and nonfocal regions.Then,several human expert raters evaluate and classify the generated labels into accurate and inaccurate labels.The inaccurate labels undergo manual segmentation by medical practitioners and are scored based on a hierarchical voting scheme before being assigned to the proposed dataset.To ensure reliability and integrity in the proposed dataset,we evaluate the accurate automated labels with manually segmented labels by medical practitioners using five assessment metrics:dice coefficient,Jaccard index,precision,recall,and accuracy.The experimental results show labels in the proposed dataset are consistent with the subjective judgment of human experts,with an average accuracy score of 94%and dice coefficient scores between 90%-99%.The study further proposes a ResNet-UNet with concatenated spatial and channel squeeze and excitation(scSE)architecture for semantic segmentation to validate and illustrate the usefulness of the proposed dataset.The results demonstrate the superior performance of the proposed architecture in accurately separating the focal and nonfocal regions compared to state-of-the-art architectures.Dataset information is released under the following URL:https://www.kaggle.com/lordamoah/datasets(accessed on 31 March 2025).展开更多
In the field of medical image processing,combining global and local relationship modeling constitutes an effective strategy for precise segmentation.Prior research has established the validity of Convolutional Neural ...In the field of medical image processing,combining global and local relationship modeling constitutes an effective strategy for precise segmentation.Prior research has established the validity of Convolutional Neural Networks(CNN)in modeling local relationships.Conversely,Transformers have demonstrated their capability to effectively capture global contextual information.However,when utilized to address CNNs’limitations in modeling global relationships,Transformers are hindered by substantial computational complexity.To address this issue,we introduce Mamba,a State-Space Model(SSM)that exhibits exceptional proficiency in modeling long-range dependencies in sequential data.Given Mamba’s demonstrated potential in 2D medical image segmentation in previous studies,we have designed a Dual-encoder Global-local Feature Extraction Network based on Mamba,termed DGFE-Mamba,to accurately capture and fuse long-range dependencies and local dependencies within multi-scale features.Compared to Transformer-based methods,the DGFE-Mamba model excels in comprehensive feature modeling and demonstrates significantly improved segmentation accuracy.To validate the effectiveness and practicality of DGFE-Mamba,we conducted tests on the Automatic Cardiac Diagnosis Challenge(ACDC)dataset,the Synapse multi-organ CT abdominal segmentation dataset,and the Colorectal Cancer Clinic(CVC-ClinicDB)dataset.The results showed that DGFE-Mamba achieved Dice coefficients of 92.20,83.67,and 94.13,respectively.These findings comprehensively validate the effectiveness and practicality of the proposed DGFE-Mamba architecture.展开更多
Deep learning(DL),derived from the domain of Artificial Neural Networks(ANN),forms one of the most essential components of modern deep learning algorithms.DL segmentation models rely on layer-by-layer convolution-base...Deep learning(DL),derived from the domain of Artificial Neural Networks(ANN),forms one of the most essential components of modern deep learning algorithms.DL segmentation models rely on layer-by-layer convolution-based feature representation,guided by forward and backward propagation.Acritical aspect of this process is the selection of an appropriate activation function(AF)to ensure robustmodel learning.However,existing activation functions often fail to effectively address the vanishing gradient problem or are complicated by the need for manual parameter tuning.Most current research on activation function design focuses on classification tasks using natural image datasets such asMNIST,CIFAR-10,and CIFAR-100.To address this gap,this study proposesMed-ReLU,a novel activation function specifically designed for medical image segmentation.Med-ReLU prevents deep learning models fromsuffering dead neurons or vanishing gradient issues.It is a hybrid activation function that combines the properties of ReLU and Softsign.For positive inputs,Med-ReLU adopts the linear behavior of ReLU to avoid vanishing gradients,while for negative inputs,it exhibits the Softsign’s polynomial convergence,ensuring robust training and avoiding inactive neurons across the training set.The training performance and segmentation accuracy ofMed-ReLU have been thoroughly evaluated,demonstrating stable learning behavior and resistance to overfitting.It consistently outperforms state-of-the-art activation functions inmedical image segmentation tasks.Designed as a parameter-free function,Med-ReLU is simple to implement in complex deep learning architectures,and its effectiveness spans various neural network models and anomaly detection scenarios.展开更多
The self-attention mechanism of Transformers,which captures long-range contextual information,has demonstrated significant potential in image segmentation.However,their ability to learn local,contextual relationships ...The self-attention mechanism of Transformers,which captures long-range contextual information,has demonstrated significant potential in image segmentation.However,their ability to learn local,contextual relationships between pixels requires further improvement.Previous methods face challenges in efficiently managing multi-scale fea-tures of different granularities from the encoder backbone,leaving room for improvement in their global representation and feature extraction capabilities.To address these challenges,we propose a novel Decoder with Multi-Head Feature Receptors(DMHFR),which receives multi-scale features from the encoder backbone and organizes them into three feature groups with different granularities:coarse,fine-grained,and full set.These groups are subsequently processed by Multi-Head Feature Receptors(MHFRs)after feature capture and modeling operations.MHFRs include two Three-Head Feature Receptors(THFRs)and one Four-Head Feature Receptor(FHFR).Each group of features is passed through these MHFRs and then fed into axial transformers,which help the model capture long-range dependencies within the features.The three MHFRs produce three distinct feature outputs.The output from the FHFR serves as auxiliary auxiliary features in the prediction head,and the prediction output and their losses will eventually be aggregated.Experimental results show that the Transformer using DMHFR outperforms 15 state of the arts(SOTA)methods on five public datasets.Specifically,it achieved significant improvements in mean DICE scores over the classic Parallel Reverse Attention Network(PraNet)method,with gains of 4.1%,2.2%,1.4%,8.9%,and 16.3%on the CVC-ClinicDB,Kvasir-SEG,CVC-T,CVC-ColonDB,and ETIS-LaribPolypDB datasets,respectively.展开更多
This study seeks to establish a novel,semi-automatic system that utilizes Industry 4.0 principles to effectively determine both acceptable and rejectable concrete cubes with regard to their failure modes,significantly...This study seeks to establish a novel,semi-automatic system that utilizes Industry 4.0 principles to effectively determine both acceptable and rejectable concrete cubes with regard to their failure modes,significantly contributing to the dependability of concrete quality evaluations.The study utilizes image processing and machine learning(ML)methods,namely object detectionmodels such as YOLOv8 and Convolutional Neural Networks(CNNs),to evaluate images of concrete cubes.These models are trained and validated on an extensive database of annotated images from real-world and laboratory conditions.Preliminary results indicate a good performance in the classification of concrete cube failure modes.The proposed system accurately identifies cracks,determines the severity of damage to structures,indicating the potential to minimize human errors and discrepancies that might occur through the current techniques to detect the failure mode of concrete cubes.Thedeveloped systemcould significantly improve the reliability of concrete cube assessments,reduce resource wastage,and contribute to more sustainable construction practices.By minimizing material costs and errors,this innovation supports the construction industry’s move towards sustainability.展开更多
Remote sensing image segmentation has a wide range of applications in land cover classification,urban building recognition,crop monitoring,and other fields.In recent years,with the booming development of deep learning...Remote sensing image segmentation has a wide range of applications in land cover classification,urban building recognition,crop monitoring,and other fields.In recent years,with the booming development of deep learning,remote sensing image segmentation models based on deep learning have gradually emerged and produced a large number of scientific research achievements.This article is based on deep learning and reviews the latest achievements in remote sensing image segmentation,exploring future development directions.Firstly,the basic concepts,characteristics,classification,tasks,and commonly used datasets of remote sensingimages are presented.Secondly,the segmentation models based on deep learning were classified and summarized,and the principles,characteristics,and applications of various models were presented.Then,the key technologies involved in deep learning remote sensing image segmentation were introduced.Finally,the future development direction and applicationprospects of remote sensing image segmentation were discussed.This article reviews the latest research achievements in remote sensing image segmentationfrom the perspective of deep learning,which can provide reference and inspiration for the research of remote sensing image segmentation.展开更多
Despite its remarkable performance on natural images,the segment anything model(SAM)lacks domain-specific information in medical imaging.and faces the challenge of losing local multi-scale information in the encoding ...Despite its remarkable performance on natural images,the segment anything model(SAM)lacks domain-specific information in medical imaging.and faces the challenge of losing local multi-scale information in the encoding phase.This paper presents a medical image segmentation model based on SAM with a local multi-scale feature encoder(LMSFE-SAM)to address the issues above.Firstly,based on the SAM,a local multi-scale feature encoder is introduced to improve the representation of features within local receptive field,thereby supplying the Vision Transformer(ViT)branch in SAM with enriched local multi-scale contextual information.At the same time,a multiaxial Hadamard product module(MHPM)is incorporated into the local multi-scale feature encoder in a lightweight manner to reduce the quadratic complexity and noise interference.Subsequently,a cross-branch balancing adapter is designed to balance the local and global information between the local multi-scale feature encoder and the ViT encoder in SAM.Finally,to obtain smaller input image size and to mitigate overlapping in patch embeddings,the size of the input image is reduced from 1024×1024 pixels to 256×256 pixels,and a multidimensional information adaptation component is developed,which includes feature adapters,position adapters,and channel-spatial adapters.This component effectively integrates the information from small-sized medical images into SAM,enhancing its suitability for clinical deployment.The proposed model demonstrates an average enhancement ranging from 0.0387 to 0.3191 across six objective evaluation metrics on BUSI,DDTI,and TN3K datasets compared to eight other representative image segmentation models.This significantly enhances the performance of the SAM on medical images,providing clinicians with a powerful tool in clinical diagnosis.展开更多
In complex industrial scenes,it is difficult to acquire high-precision non-cooperative target pose under monocular visual servo control.This paper presents a new method of target extraction and high-precision edge fit...In complex industrial scenes,it is difficult to acquire high-precision non-cooperative target pose under monocular visual servo control.This paper presents a new method of target extraction and high-precision edge fitting for the wheel of the sintering trolley in steel production,which fuses multiple target extraction algorithms adapting to the working environment of the target.Firstly,based on obvious difference between the pixels of the target image and the non-target image in the gray histogram,these pixels were classified and then segmented in intraclass,removing interference factors and remaining the target image.Then,multiple segmentation results were merged and a final target image was obtained after small connected regions were eliminated.In the edge fitting stage,the edge fitting method with best-circumscribed rectangle was proposed to accurately fit the circular target edge.Finally,PnP algorithm was adopted for pose measurement of the target.The experimental results showed that the average estimation error of pose angleγwith respect to the z-axis rotation was 0.2346°,the average measurement error of pose angleαwith respect to the x-axis rotation was 0.1703°,and the average measurement error of pose angle β with respect to the y-axis rotation was 0.2275°.The proposed method has practical application value.展开更多
In micro milling machining,tool wear directly affects workpiece quality and accuracy,making effective tool wear monitoring a key factor in ensuring product integrity.The use of machine vision-based methods can provide...In micro milling machining,tool wear directly affects workpiece quality and accuracy,making effective tool wear monitoring a key factor in ensuring product integrity.The use of machine vision-based methods can provide an intuitive and efficient representation of tool wear conditions.However,micro milling tools have non-flat flanks,thin coatings can peel off,and spindle orientation is uncertain during downtime.These factors result in low pixel values,uneven illumination,and arbitrary tool position.To address this,we propose an image-based tool wear monitoring method.It combines multiple algorithms to restore lost pixels due to uneven illumination during segmentation and accurately extract wear areas.Experimental results demonstrate that the proposed algorithm exhibits high robustness to such images,effectively addressing the effects of illumination and spindle orientation.Additionally,the algorithm has low complexity,fast execution time,and significantly reduces the detection time in situ.展开更多
Structural Health Monitoring(SHM)systems play a key role in managing buildings and infrastructure by delivering vital insights into their strength and structural integrity.There is a need for more efficient techniques...Structural Health Monitoring(SHM)systems play a key role in managing buildings and infrastructure by delivering vital insights into their strength and structural integrity.There is a need for more efficient techniques to detect defects,as traditional methods are often prone to human error,and this issue is also addressed through image processing(IP).In addition to IP,automated,accurate,and real-time detection of structural defects,such as cracks,corrosion,and material degradation that conventional inspection techniques may miss,is made possible by Artificial Intelligence(AI)technologies like Machine Learning(ML)and Deep Learning(DL).This review examines the integration of computer vision and AI techniques in Structural Health Monitoring(SHM),investigating their effectiveness in detecting various forms of structural deterioration.Also,it evaluates ML and DL models in SHM for their accuracy in identifying and assessing structural damage,ultimately enhancing safety,durability,and maintenance practices in the field.Key findings reveal that AI-powered approaches,especially those utilizing IP and DL models like CNNs,significantly improve detection efficiency and accuracy,with reported accuracies in various SHM tasks.However,significant research gaps remain,including challenges with the consistency,quality,and environmental resilience of image data,a notable lack of standardized models and datasets for training across diverse structures,and concerns regarding computational costs,model interpretability,and seamless integration with existing systems.Future work should focus on developing more robust models through data augmentation,transfer learning,and hybrid approaches,standardizing protocols,and fostering interdisciplinary collaboration to overcome these limitations and achieve more reliable,scalable,and affordable SHM systems.展开更多
Medical image segmentation is a crucial task in clinical applications.However,obtaining labeled data for medical images is often challenging.This has led to the appeal of semi-supervised learning(SSL),a technique adep...Medical image segmentation is a crucial task in clinical applications.However,obtaining labeled data for medical images is often challenging.This has led to the appeal of semi-supervised learning(SSL),a technique adept at leveraging a modest amount of labeled data.Nonetheless,most prevailing SSL segmentation methods for medical images either rely on the single consistency training method or directly fine-tune SSL methods designed for natural images.In this paper,we propose an innovative semi-supervised method called multi-consistency training(MCT)for medical image segmentation.Our approach transcends the constraints of prior methodologies by considering consistency from a dual perspective:output consistency across different up-sampling methods and output consistency of the same data within the same network under various perturbations to the intermediate features.We design distinct semi-supervised loss regression methods for these two types of consistencies.To enhance the application of our MCT model,we also develop a dedicated decoder as the core of our neural network.Thorough experiments were conducted on the polyp dataset and the dental dataset,rigorously compared against other SSL methods.Experimental results demonstrate the superiority of our approach,achieving higher segmentation accuracy.Moreover,comprehensive ablation studies and insightful discussion substantiate the efficacy of our approach in navigating the intricacies of medical image segmentation.展开更多
All-optical image processing has been viewed as a promising technique for its high computation speed and low power consumption.However,current methods are often restricted to few functionalities and low reconfigurabil...All-optical image processing has been viewed as a promising technique for its high computation speed and low power consumption.However,current methods are often restricted to few functionalities and low reconfigurabilities,which cannot meet the growing demand for device integration and scenario adaptation in next-generation vision regimes.Here,we propose and experimentally demonstrate a bilayer liquid crystal computing platform for reconfigurable image processing.Under different in-situ/ex-situ twisted/untwisted conditions of the layers,our approach allows for eight kinds of image processing functions,including one/two-channel bright field imaging,one/two-channel vortex filtering,horizontally/vertically one-dimensional edge detection,vertex detection,and photonic spin Hall effect-based resolution adjustable edge detection.A unified theoretical framework for this scheme is established on the transfer function theory,which coincides well with the experimental results.The proposed method offers an easily-switchable multi-functional solution to optical image processing by introducing mechanical degrees of freedom,which may enable emerging applications in computer vision,autonomous driving,and biomedical microscopy.展开更多
基金funded by the Research Project:THTETN.05/24-25,VietnamAcademy of Science and Technology.
文摘Satellite image segmentation plays a crucial role in remote sensing,supporting applications such as environmental monitoring,land use analysis,and disaster management.However,traditional segmentation methods often rely on large amounts of labeled data,which are costly and time-consuming to obtain,especially in largescale or dynamic environments.To address this challenge,we propose the Semi-Supervised Multi-View Picture Fuzzy Clustering(SS-MPFC)algorithm,which improves segmentation accuracy and robustness,particularly in complex and uncertain remote sensing scenarios.SS-MPFC unifies three paradigms:semi-supervised learning,multi-view clustering,and picture fuzzy set theory.This integration allows the model to effectively utilize a small number of labeled samples,fuse complementary information from multiple data views,and handle the ambiguity and uncertainty inherent in satellite imagery.We design a novel objective function that jointly incorporates picture fuzzy membership functions across multiple views of the data,and embeds pairwise semi-supervised constraints(must-link and cannot-link)directly into the clustering process to enhance segmentation accuracy.Experiments conducted on several benchmark satellite datasets demonstrate that SS-MPFC significantly outperforms existing state-of-the-art methods in segmentation accuracy,noise robustness,and semantic interpretability.On the Augsburg dataset,SS-MPFC achieves a Purity of 0.8158 and an Accuracy of 0.6860,highlighting its outstanding robustness and efficiency.These results demonstrate that SSMPFC offers a scalable and effective solution for real-world satellite-based monitoring systems,particularly in scenarios where rapid annotation is infeasible,such as wildfire tracking,agricultural monitoring,and dynamic urban mapping.
文摘Medical image segmentation is of critical importance in the domain of contemporary medical imaging.However,U-Net and its variants exhibit limitations in capturing complex nonlinear patterns and global contextual information.Although the subsequent U-KAN model enhances nonlinear representation capabilities,it still faces challenges such as gradient vanishing during deep network training and spatial detail loss during feature downsampling,resulting in insufficient segmentation accuracy for edge structures and minute lesions.To address these challenges,this paper proposes the RE-UKAN model,which innovatively improves upon U-KAN.Firstly,a residual network is introduced into the encoder to effectively mitigate gradient vanishing through cross-layer identity mappings,thus enhancing modelling capabilities for complex pathological structures.Secondly,Efficient Local Attention(ELA)is integrated to suppress spatial detail loss during downsampling,thereby improving the perception of edge structures and minute lesions.Experimental results on four public datasets demonstrate that RE-UKAN outperforms existing medical image segmentation methods across multiple evaluation metrics,with particularly outstanding performance on the TN-SCUI 2020 dataset,achieving IoU of 88.18%and Dice of 93.57%.Compared to the baseline model,it achieves improvements of 3.05%and 1.72%,respectively.These results fully demonstrate RE-UKAN’s superior detail retention capability and boundary recognition accuracy in complex medical image segmentation tasks,providing a reliable solution for clinical precision segmentation.
基金National Natural Science Foundation of China under Grants No.62171047,U22B2001,62271065,62001051Beijing Natural Science Foundation under Grant L223027BUPT Excellent Ph.D Students Foundation under Grants CX2021114。
文摘This article studies the problem of image segmentation-based semantic communication in autonomous driving.In real traffic scenes,the detecting of objects(e.g.,vehicles and pedestrians)is more important to guarantee driving safety,which is always ignored in existing works.Therefore,we propose a vehicular image segmentation-oriented semantic communication system,termed VIS-SemCom,focusing on transmitting and recovering image semantic features of high-important objects to reduce transmission redundancy.First,we develop a semantic codec based on Swin Transformer architecture,which expands the perceptual field thus improving the segmentation accuracy.To highlight the important objects'accuracy,we propose a multi-scale semantic extraction method by assigning the number of Swin Transformer blocks for diverse resolution semantic features.Also,an importance-aware loss incorporating important levels is devised,and an online hard example mining(OHEM)strategy is proposed to handle small sample issues in the dataset.Finally,experimental results demonstrate that the proposed VIS-SemCom can achieve a significant mean intersection over union(mIoU)performance in the SNR regions,a reduction of transmitted data volume by about 60%at 60%mIoU,and improve the segmentation accuracy of important objects,compared to baseline image communication.
基金supported by the Natural Science Foundation of the Anhui Higher Education Institutions of China(Grant Nos.2023AH040149 and 2024AH051915)the Anhui Provincial Natural Science Foundation(Grant No.2208085MF168)+1 种基金the Science and Technology Innovation Tackle Plan Project of Maanshan(Grant No.2024RGZN001)the Scientific Research Fund Project of Anhui Medical University(Grant No.2023xkj122).
文摘Convolutional neural networks(CNNs)-based medical image segmentation technologies have been widely used in medical image segmentation because of their strong representation and generalization abilities.However,due to the inability to effectively capture global information from images,CNNs can easily lead to loss of contours and textures in segmentation results.Notice that the transformer model can effectively capture the properties of long-range dependencies in the image,and furthermore,combining the CNN and the transformer can effectively extract local details and global contextual features of the image.Motivated by this,we propose a multi-branch and multi-scale attention network(M2ANet)for medical image segmentation,whose architecture consists of three components.Specifically,in the first component,we construct an adaptive multi-branch patch module for parallel extraction of image features to reduce information loss caused by downsampling.In the second component,we apply residual block to the well-known convolutional block attention module to enhance the network’s ability to recognize important features of images and alleviate the phenomenon of gradient vanishing.In the third component,we design a multi-scale feature fusion module,in which we adopt adaptive average pooling and position encoding to enhance contextual features,and then multi-head attention is introduced to further enrich feature representation.Finally,we validate the effectiveness and feasibility of the proposed M2ANet method through comparative experiments on four benchmark medical image segmentation datasets,particularly in the context of preserving contours and textures.
基金National Key Research and Development Program of China,Grant/Award Number:2018YFE0206900China Postdoctoral Science Foundation,Grant/Award Number:2023M731204+2 种基金The Open Project of Key Laboratory for Quality Evaluation of Ultrasound Surgical Equipment of National Medical Products Administration,Grant/Award Number:SMDTKL-2023-1-01The Hubei Province Key Research and Development Project,Grant/Award Number:2023BCB007CAAI-Huawei MindSpore Open Fund。
文摘Convolutional neural network(CNN)with the encoder-decoder structure is popular in medical image segmentation due to its excellent local feature extraction ability but it faces limitations in capturing the global feature.The transformer can extract the global information well but adapting it to small medical datasets is challenging and its computational complexity can be heavy.In this work,a serial and parallel network is proposed for the accurate 3D medical image segmentation by combining CNN and transformer and promoting feature interactions across various semantic levels.The core components of the proposed method include the cross window self-attention based transformer(CWST)and multi-scale local enhanced(MLE)modules.The CWST module enhances the global context understanding by partitioning 3D images into non-overlapping windows and calculating sparse global attention between windows.The MLE module selectively fuses features by computing the voxel attention between different branch features,and uses convolution to strengthen the dense local information.The experiments on the prostate,atrium,and pancreas MR/CT image datasets consistently demonstrate the advantage of the proposed method over six popular segmentation models in both qualitative evaluation and quantitative indexes such as dice similarity coefficient,Intersection over Union,95%Hausdorff distance and average symmetric surface distance.
基金supported by the Natural Science Foundation of China(No.41804112,author:Chengyun Song).
文摘Existing semi-supervisedmedical image segmentation algorithms use copy-paste data augmentation to correct the labeled-unlabeled data distribution mismatch.However,current copy-paste methods have three limitations:(1)training the model solely with copy-paste mixed pictures from labeled and unlabeled input loses a lot of labeled information;(2)low-quality pseudo-labels can cause confirmation bias in pseudo-supervised learning on unlabeled data;(3)the segmentation performance in low-contrast and local regions is less than optimal.We design a Stochastic Augmentation-Based Dual-Teaching Auxiliary Training Strategy(SADT),which enhances feature diversity and learns high-quality features to overcome these problems.To be more precise,SADT trains the Student Network by using pseudo-label-based training from Teacher Network 1 and supervised learning with labeled data,which prevents the loss of rare labeled data.We introduce a bi-directional copy-pastemask with progressive high-entropy filtering to reduce data distribution disparities and mitigate confirmation bias in pseudo-supervision.For the mixed images,Deep-Shallow Spatial Contrastive Learning(DSSCL)is proposed in the feature spaces of Teacher Network 2 and the Student Network to improve the segmentation capabilities in low-contrast and local areas.In this procedure,the features retrieved by the Student Network are subjected to a random feature perturbation technique.On two openly available datasets,extensive trials show that our proposed SADT performs much better than the state-ofthe-art semi-supervised medical segmentation techniques.Using only 10%of the labeled data for training,SADT was able to acquire a Dice score of 90.10%on the ACDC(Automatic Cardiac Diagnosis Challenge)dataset.
基金supported by the China Chongqing Municipal Education Commission(Grant KJZDM202500505)China Chongqing Municipal Science and Technology Bureau(Grants CSTB2024TIADCYKJCXX0009,CSTB2024NSCQ-LZX0043)+1 种基金Chongqing University of Technology graduate education high-quality development project(Grants gzlsz202304,gzlkc202401,gzltd202502)Chongqing University of Technology-Chongqing LINGLUE Technology Co.Ltd.Electronic Information(artificial intelligence)graduate joint training base.
文摘Accurate medical image segmentation plays a crucial role in improving the precision of computer-aided diagnosis.However,complex boundary shapes,low contrast and blurred anatomical structures make fine-grained segmentation a challenging task.Variational Bayesian inference quantifies uncertainty through probability distributions and can construct robust probabilistic models for the boundaries of ambiguous organs and tissues.In this paper,we apply variational Bayesian inference to medical image segmentation and propose variational attention to model the uncertainty of low-contrast and blurry tissue and organ boundaries.This enhances the model's ability to perceive segmentation boundaries,improving robustness and segmentation accuracy.Variational attention first estimates the parameters of the probability distribution of latent representations based on input features.Then,it samples latent representations from the learnt distribution to generate attention weights that optimise the interaction between global features and ambiguous boundaries.We integrate variational attention into the U-Net model by replacing its skip connections,constructing a multi-scale variational attention segmentation model(V-UNet).Experiments on the ISBI 2012 and MoNuSeg 2018 datasets show that our method achieves Dice scores of 95.89%and 82.18%,respectively.Moreover,we integrate V-UNet into the Mask R-CNN framework by replacing the FPN feature extraction head and propose a two-stage segmentation method.Compared to the original Mask R-CNN,our method improves the Dice score by 0.81%,mAP by 8.06%and F1 score by 0.51%.
文摘Medical image segmentation has become a cornerstone for many healthcare applications,allowing for the automated extraction of critical information from images such as Computed Tomography(CT)scans,Magnetic Resonance Imaging(MRIs),and X-rays.The introduction of U-Net in 2015 has significantly advanced segmentation capabilities,especially for small datasets commonly found in medical imaging.Since then,various modifications to the original U-Net architecture have been proposed to enhance segmentation accuracy and tackle challenges like class imbalance,data scarcity,and multi-modal image processing.This paper provides a detailed review and comparison of several U-Net-based architectures,focusing on their effectiveness in medical image segmentation tasks.We evaluate performance metrics such as Dice Similarity Coefficient(DSC)and Intersection over Union(IoU)across different U-Net variants including HmsU-Net,CrossU-Net,mResU-Net,and others.Our results indicate that architectural enhancements such as transformers,attention mechanisms,and residual connections improve segmentation performance across diverse medical imaging applications,including tumor detection,organ segmentation,and lesion identification.The study also identifies current challenges in the field,including data variability,limited dataset sizes,and issues with class imbalance.Based on these findings,the paper suggests potential future directions for improving the robustness and clinical applicability of U-Net-based models in medical image segmentation.
基金supported by the National Natural Science Foundation of China(Grant Nos.62072250,61772281,61702235,U1636117,U1804263,62172435,61872203 and 61802212)the Zhongyuan Science and Technology Innovation Leading Talent Project of China(Grant No.214200510019)+3 种基金the Suqian Municipal Science and Technology Plan Project in 2020(S202015)the Plan for Scientific Talent of Henan Province(Grant No.2018JR0018)the Opening Project of Guangdong Provincial Key Laboratory of Information Security Technology(Grant No.2020B1212060078)the Priority Academic Program Development of Jiangsu Higher Education Institutions(PAPD)Fund.
文摘Medical image segmentation,i.e.,labeling structures of interest in medical images,is crucial for disease diagnosis and treatment in radiology.In reversible data hiding in medical images(RDHMI),segmentation consists of only two regions:the focal and nonfocal regions.The focal region mainly contains information for diagnosis,while the nonfocal region serves as the monochrome background.The current traditional segmentation methods utilized in RDHMI are inaccurate for complex medical images,and manual segmentation is time-consuming,poorly reproducible,and operator-dependent.Implementing state-of-the-art deep learning(DL)models will facilitate key benefits,but the lack of domain-specific labels for existing medical datasets makes it impossible.To address this problem,this study provides labels of existing medical datasets based on a hybrid segmentation approach to facilitate the implementation of DL segmentation models in this domain.First,an initial segmentation based on a 33 kernel is performed to analyze×identified contour pixels before classifying pixels into focal and nonfocal regions.Then,several human expert raters evaluate and classify the generated labels into accurate and inaccurate labels.The inaccurate labels undergo manual segmentation by medical practitioners and are scored based on a hierarchical voting scheme before being assigned to the proposed dataset.To ensure reliability and integrity in the proposed dataset,we evaluate the accurate automated labels with manually segmented labels by medical practitioners using five assessment metrics:dice coefficient,Jaccard index,precision,recall,and accuracy.The experimental results show labels in the proposed dataset are consistent with the subjective judgment of human experts,with an average accuracy score of 94%and dice coefficient scores between 90%-99%.The study further proposes a ResNet-UNet with concatenated spatial and channel squeeze and excitation(scSE)architecture for semantic segmentation to validate and illustrate the usefulness of the proposed dataset.The results demonstrate the superior performance of the proposed architecture in accurately separating the focal and nonfocal regions compared to state-of-the-art architectures.Dataset information is released under the following URL:https://www.kaggle.com/lordamoah/datasets(accessed on 31 March 2025).
基金supported by the National Natural Science Foundation of China(62276092,62303167)the Postdoctoral Fellowship Program(Grade C)of China Postdoctoral Science Foundation(GZC20230707)+3 种基金the Key Science and Technology Program of Henan Province,China(242102211051)MRC(MC_PC_17171)Royal Society(RP202G0230)BHF(AA/18/3/34220)。
文摘In the field of medical image processing,combining global and local relationship modeling constitutes an effective strategy for precise segmentation.Prior research has established the validity of Convolutional Neural Networks(CNN)in modeling local relationships.Conversely,Transformers have demonstrated their capability to effectively capture global contextual information.However,when utilized to address CNNs’limitations in modeling global relationships,Transformers are hindered by substantial computational complexity.To address this issue,we introduce Mamba,a State-Space Model(SSM)that exhibits exceptional proficiency in modeling long-range dependencies in sequential data.Given Mamba’s demonstrated potential in 2D medical image segmentation in previous studies,we have designed a Dual-encoder Global-local Feature Extraction Network based on Mamba,termed DGFE-Mamba,to accurately capture and fuse long-range dependencies and local dependencies within multi-scale features.Compared to Transformer-based methods,the DGFE-Mamba model excels in comprehensive feature modeling and demonstrates significantly improved segmentation accuracy.To validate the effectiveness and practicality of DGFE-Mamba,we conducted tests on the Automatic Cardiac Diagnosis Challenge(ACDC)dataset,the Synapse multi-organ CT abdominal segmentation dataset,and the Colorectal Cancer Clinic(CVC-ClinicDB)dataset.The results showed that DGFE-Mamba achieved Dice coefficients of 92.20,83.67,and 94.13,respectively.These findings comprehensively validate the effectiveness and practicality of the proposed DGFE-Mamba architecture.
基金The researchers would like to thank the Deanship of Graduate Studies and Scientific Research at Qassim University for financial support(QU-APC-2025).
文摘Deep learning(DL),derived from the domain of Artificial Neural Networks(ANN),forms one of the most essential components of modern deep learning algorithms.DL segmentation models rely on layer-by-layer convolution-based feature representation,guided by forward and backward propagation.Acritical aspect of this process is the selection of an appropriate activation function(AF)to ensure robustmodel learning.However,existing activation functions often fail to effectively address the vanishing gradient problem or are complicated by the need for manual parameter tuning.Most current research on activation function design focuses on classification tasks using natural image datasets such asMNIST,CIFAR-10,and CIFAR-100.To address this gap,this study proposesMed-ReLU,a novel activation function specifically designed for medical image segmentation.Med-ReLU prevents deep learning models fromsuffering dead neurons or vanishing gradient issues.It is a hybrid activation function that combines the properties of ReLU and Softsign.For positive inputs,Med-ReLU adopts the linear behavior of ReLU to avoid vanishing gradients,while for negative inputs,it exhibits the Softsign’s polynomial convergence,ensuring robust training and avoiding inactive neurons across the training set.The training performance and segmentation accuracy ofMed-ReLU have been thoroughly evaluated,demonstrating stable learning behavior and resistance to overfitting.It consistently outperforms state-of-the-art activation functions inmedical image segmentation tasks.Designed as a parameter-free function,Med-ReLU is simple to implement in complex deep learning architectures,and its effectiveness spans various neural network models and anomaly detection scenarios.
基金supported by Xiamen Medical and Health Guidance Project in 2021(No.3502Z20214ZD1070)supported by a grant from Guangxi Key Laboratory of Machine Vision and Intelligent Control,China(No.2023B02).
文摘The self-attention mechanism of Transformers,which captures long-range contextual information,has demonstrated significant potential in image segmentation.However,their ability to learn local,contextual relationships between pixels requires further improvement.Previous methods face challenges in efficiently managing multi-scale fea-tures of different granularities from the encoder backbone,leaving room for improvement in their global representation and feature extraction capabilities.To address these challenges,we propose a novel Decoder with Multi-Head Feature Receptors(DMHFR),which receives multi-scale features from the encoder backbone and organizes them into three feature groups with different granularities:coarse,fine-grained,and full set.These groups are subsequently processed by Multi-Head Feature Receptors(MHFRs)after feature capture and modeling operations.MHFRs include two Three-Head Feature Receptors(THFRs)and one Four-Head Feature Receptor(FHFR).Each group of features is passed through these MHFRs and then fed into axial transformers,which help the model capture long-range dependencies within the features.The three MHFRs produce three distinct feature outputs.The output from the FHFR serves as auxiliary auxiliary features in the prediction head,and the prediction output and their losses will eventually be aggregated.Experimental results show that the Transformer using DMHFR outperforms 15 state of the arts(SOTA)methods on five public datasets.Specifically,it achieved significant improvements in mean DICE scores over the classic Parallel Reverse Attention Network(PraNet)method,with gains of 4.1%,2.2%,1.4%,8.9%,and 16.3%on the CVC-ClinicDB,Kvasir-SEG,CVC-T,CVC-ColonDB,and ETIS-LaribPolypDB datasets,respectively.
文摘This study seeks to establish a novel,semi-automatic system that utilizes Industry 4.0 principles to effectively determine both acceptable and rejectable concrete cubes with regard to their failure modes,significantly contributing to the dependability of concrete quality evaluations.The study utilizes image processing and machine learning(ML)methods,namely object detectionmodels such as YOLOv8 and Convolutional Neural Networks(CNNs),to evaluate images of concrete cubes.These models are trained and validated on an extensive database of annotated images from real-world and laboratory conditions.Preliminary results indicate a good performance in the classification of concrete cube failure modes.The proposed system accurately identifies cracks,determines the severity of damage to structures,indicating the potential to minimize human errors and discrepancies that might occur through the current techniques to detect the failure mode of concrete cubes.Thedeveloped systemcould significantly improve the reliability of concrete cube assessments,reduce resource wastage,and contribute to more sustainable construction practices.By minimizing material costs and errors,this innovation supports the construction industry’s move towards sustainability.
文摘Remote sensing image segmentation has a wide range of applications in land cover classification,urban building recognition,crop monitoring,and other fields.In recent years,with the booming development of deep learning,remote sensing image segmentation models based on deep learning have gradually emerged and produced a large number of scientific research achievements.This article is based on deep learning and reviews the latest achievements in remote sensing image segmentation,exploring future development directions.Firstly,the basic concepts,characteristics,classification,tasks,and commonly used datasets of remote sensingimages are presented.Secondly,the segmentation models based on deep learning were classified and summarized,and the principles,characteristics,and applications of various models were presented.Then,the key technologies involved in deep learning remote sensing image segmentation were introduced.Finally,the future development direction and applicationprospects of remote sensing image segmentation were discussed.This article reviews the latest research achievements in remote sensing image segmentationfrom the perspective of deep learning,which can provide reference and inspiration for the research of remote sensing image segmentation.
基金supported by Natural Science Foundation Programme of Gansu Province(No.24JRRA231)National Natural Science Foundation of China(No.62061023)Gansu Provincial Science and Technology Plan Key Research and Development Program Project(No.24YFFA024).
文摘Despite its remarkable performance on natural images,the segment anything model(SAM)lacks domain-specific information in medical imaging.and faces the challenge of losing local multi-scale information in the encoding phase.This paper presents a medical image segmentation model based on SAM with a local multi-scale feature encoder(LMSFE-SAM)to address the issues above.Firstly,based on the SAM,a local multi-scale feature encoder is introduced to improve the representation of features within local receptive field,thereby supplying the Vision Transformer(ViT)branch in SAM with enriched local multi-scale contextual information.At the same time,a multiaxial Hadamard product module(MHPM)is incorporated into the local multi-scale feature encoder in a lightweight manner to reduce the quadratic complexity and noise interference.Subsequently,a cross-branch balancing adapter is designed to balance the local and global information between the local multi-scale feature encoder and the ViT encoder in SAM.Finally,to obtain smaller input image size and to mitigate overlapping in patch embeddings,the size of the input image is reduced from 1024×1024 pixels to 256×256 pixels,and a multidimensional information adaptation component is developed,which includes feature adapters,position adapters,and channel-spatial adapters.This component effectively integrates the information from small-sized medical images into SAM,enhancing its suitability for clinical deployment.The proposed model demonstrates an average enhancement ranging from 0.0387 to 0.3191 across six objective evaluation metrics on BUSI,DDTI,and TN3K datasets compared to eight other representative image segmentation models.This significantly enhances the performance of the SAM on medical images,providing clinicians with a powerful tool in clinical diagnosis.
基金supported by Key Research and Development Projects in Shaanxi Province (No. 2021GY-265)Xi’an University Talent Service Enterprise Project (No.2020KJRC0049)。
文摘In complex industrial scenes,it is difficult to acquire high-precision non-cooperative target pose under monocular visual servo control.This paper presents a new method of target extraction and high-precision edge fitting for the wheel of the sintering trolley in steel production,which fuses multiple target extraction algorithms adapting to the working environment of the target.Firstly,based on obvious difference between the pixels of the target image and the non-target image in the gray histogram,these pixels were classified and then segmented in intraclass,removing interference factors and remaining the target image.Then,multiple segmentation results were merged and a final target image was obtained after small connected regions were eliminated.In the edge fitting stage,the edge fitting method with best-circumscribed rectangle was proposed to accurately fit the circular target edge.Finally,PnP algorithm was adopted for pose measurement of the target.The experimental results showed that the average estimation error of pose angleγwith respect to the z-axis rotation was 0.2346°,the average measurement error of pose angleαwith respect to the x-axis rotation was 0.1703°,and the average measurement error of pose angle β with respect to the y-axis rotation was 0.2275°.The proposed method has practical application value.
基金Supported by National Natural Science Foundation of China(Grant No.52175528)。
文摘In micro milling machining,tool wear directly affects workpiece quality and accuracy,making effective tool wear monitoring a key factor in ensuring product integrity.The use of machine vision-based methods can provide an intuitive and efficient representation of tool wear conditions.However,micro milling tools have non-flat flanks,thin coatings can peel off,and spindle orientation is uncertain during downtime.These factors result in low pixel values,uneven illumination,and arbitrary tool position.To address this,we propose an image-based tool wear monitoring method.It combines multiple algorithms to restore lost pixels due to uneven illumination during segmentation and accurately extract wear areas.Experimental results demonstrate that the proposed algorithm exhibits high robustness to such images,effectively addressing the effects of illumination and spindle orientation.Additionally,the algorithm has low complexity,fast execution time,and significantly reduces the detection time in situ.
文摘Structural Health Monitoring(SHM)systems play a key role in managing buildings and infrastructure by delivering vital insights into their strength and structural integrity.There is a need for more efficient techniques to detect defects,as traditional methods are often prone to human error,and this issue is also addressed through image processing(IP).In addition to IP,automated,accurate,and real-time detection of structural defects,such as cracks,corrosion,and material degradation that conventional inspection techniques may miss,is made possible by Artificial Intelligence(AI)technologies like Machine Learning(ML)and Deep Learning(DL).This review examines the integration of computer vision and AI techniques in Structural Health Monitoring(SHM),investigating their effectiveness in detecting various forms of structural deterioration.Also,it evaluates ML and DL models in SHM for their accuracy in identifying and assessing structural damage,ultimately enhancing safety,durability,and maintenance practices in the field.Key findings reveal that AI-powered approaches,especially those utilizing IP and DL models like CNNs,significantly improve detection efficiency and accuracy,with reported accuracies in various SHM tasks.However,significant research gaps remain,including challenges with the consistency,quality,and environmental resilience of image data,a notable lack of standardized models and datasets for training across diverse structures,and concerns regarding computational costs,model interpretability,and seamless integration with existing systems.Future work should focus on developing more robust models through data augmentation,transfer learning,and hybrid approaches,standardizing protocols,and fostering interdisciplinary collaboration to overcome these limitations and achieve more reliable,scalable,and affordable SHM systems.
基金the Innovation Program of Shanghai Industrial Synergy(No.XTCX-KJ-2023-2-12)。
文摘Medical image segmentation is a crucial task in clinical applications.However,obtaining labeled data for medical images is often challenging.This has led to the appeal of semi-supervised learning(SSL),a technique adept at leveraging a modest amount of labeled data.Nonetheless,most prevailing SSL segmentation methods for medical images either rely on the single consistency training method or directly fine-tune SSL methods designed for natural images.In this paper,we propose an innovative semi-supervised method called multi-consistency training(MCT)for medical image segmentation.Our approach transcends the constraints of prior methodologies by considering consistency from a dual perspective:output consistency across different up-sampling methods and output consistency of the same data within the same network under various perturbations to the intermediate features.We design distinct semi-supervised loss regression methods for these two types of consistencies.To enhance the application of our MCT model,we also develop a dedicated decoder as the core of our neural network.Thorough experiments were conducted on the polyp dataset and the dental dataset,rigorously compared against other SSL methods.Experimental results demonstrate the superiority of our approach,achieving higher segmentation accuracy.Moreover,comprehensive ablation studies and insightful discussion substantiate the efficacy of our approach in navigating the intricacies of medical image segmentation.
基金supported in part by the National Natural Science Foundation of China(12421005,12374273,and 61805077)in part by the Natural Science Foundation of Hunan Province(2025JJ50046)in part by the Hunan Provincial Major Sci-Tech Program(2023ZJ1010)。
文摘All-optical image processing has been viewed as a promising technique for its high computation speed and low power consumption.However,current methods are often restricted to few functionalities and low reconfigurabilities,which cannot meet the growing demand for device integration and scenario adaptation in next-generation vision regimes.Here,we propose and experimentally demonstrate a bilayer liquid crystal computing platform for reconfigurable image processing.Under different in-situ/ex-situ twisted/untwisted conditions of the layers,our approach allows for eight kinds of image processing functions,including one/two-channel bright field imaging,one/two-channel vortex filtering,horizontally/vertically one-dimensional edge detection,vertex detection,and photonic spin Hall effect-based resolution adjustable edge detection.A unified theoretical framework for this scheme is established on the transfer function theory,which coincides well with the experimental results.The proposed method offers an easily-switchable multi-functional solution to optical image processing by introducing mechanical degrees of freedom,which may enable emerging applications in computer vision,autonomous driving,and biomedical microscopy.