In modern industrial production,foreign object detection in complex environments is crucial to ensure product quality and production safety.Detection systems based on deep-learning image processing algorithms often fa...In modern industrial production,foreign object detection in complex environments is crucial to ensure product quality and production safety.Detection systems based on deep-learning image processing algorithms often face challenges with handling high-resolution images and achieving accurate detection against complex backgrounds.To address these issues,this study employs the PatchCore unsupervised anomaly detection algorithm combined with data augmentation techniques to enhance the system’s generalization capability across varying lighting conditions,viewing angles,and object scales.The proposed method is evaluated in a complex industrial detection scenario involving the bogie of an electric multiple unit(EMU).A dataset consisting of complex backgrounds,diverse lighting conditions,and multiple viewing angles is constructed to validate the performance of the detection system in real industrial environments.Experimental results show that the proposed model achieves an average area under the receiver operating characteristic curve(AUROC)of 0.92 and an average F1 score of 0.85.Combined with data augmentation,the proposed model exhibits improvements in AUROC by 0.06 and F1 score by 0.03,demonstrating enhanced accuracy and robustness for foreign object detection in complex industrial settings.In addition,the effects of key factors on detection performance are systematically analyzed,providing practical guidance for parameter selection in real industrial applications.展开更多
Existing elevator fault diagnosis algorithms have limited engineering applicability due to variations in working conditions and differences in equipment structures.To address this limitation,this study proposes an uns...Existing elevator fault diagnosis algorithms have limited engineering applicability due to variations in working conditions and differences in equipment structures.To address this limitation,this study proposes an unsupervised subdomain adaptation method based on a time-frequency feature attention mechanism,LMMD-based subdomain alignment,and contrastive local alignment.This enables the application of the diagnosis model across different working conditions and equipment types.First,a novel time-frequency feature attention mechanism assigns weights to vibration signals of varying dimensions.Second,the time series is transformed to obtain a three-channel time-frequency diagram.This diagram is input into the proposed dimension-segmentation cross-channel multihead self-attention framework to extract high-dimensional frequencydomain fault features.These features are concatenated with the time-domain features to obtain a global feature representation.Then,the extracted high-dimensional features are sent to the classification module to obtain the predicted labels for the source and target domains.Finally,after confidence filtering,the true labels from the source domain and the prediction labels from the target domain are fed into a dynamically weighted multilevel feature alignment module to promote proximity between similar fault features across domains while enhancing separation among different fault types.The validity and superiority of the proposed method were demonstrated through simulation experiments conducted on two types of manned escalator systems under multiple working conditions.For the most challenging transfer task,the proposed method achieved higher accuracy on the target domain test set than DANN,ADDA,C-CLCN,TFA-CCN,and TFA-LCN by 26.87%,24.72%,11.44%,28.94%,and 16.85%,respectively.展开更多
Single-pixel imaging(SPI)is a prominent scattering media imaging technique that allows image transmission via one-dimensional detection under structured illumination,with applications spanning from long-range imaging ...Single-pixel imaging(SPI)is a prominent scattering media imaging technique that allows image transmission via one-dimensional detection under structured illumination,with applications spanning from long-range imaging to microscopy.Recent advancements leveraging deep learning(DL)have significantly improved SPI performance,especially at low compression ratios.However,most DL-based SPI methods proposed so far rely heavily on extensive labeled datasets for supervised training,which are often impractical in real-world scenarios.Here,we propose an unsupervised learningenabled label-free SPI method for resilient information transmission through unknown dynamic scattering media.Additionally,we introduce a physics-informed autoencoder framework to optimize encoding schemes,further enhancing image quality at low compression ratios.Simulation and experimental results demonstrate that high-efficiency data transmission with structural similarity exceeding 0.9 is achieved through challenging turbulent channels.Moreover,experiments demonstrate that in a 5 m underwater dynamic turbulent channel,USAF target imaging quality surpasses traditional methods by over 13 dB.The compressive encoded transmission of 720×720 resolution video exceeding 30 seconds with great fidelity is also successfully demonstrated.These preliminary results suggest that our proposed method opens up a new paradigm for resilient information transmission through unknown dynamic scattering media and holds potential for broader applications within many other scattering media imaging technologies.展开更多
To avoid the laborious annotation process for dense prediction tasks like semantic segmentation,unsupervised domain adaptation(UDA)methods have been proposed to leverage the abundant annotations from a source domain,s...To avoid the laborious annotation process for dense prediction tasks like semantic segmentation,unsupervised domain adaptation(UDA)methods have been proposed to leverage the abundant annotations from a source domain,such as virtual world(e.g.,3D games),and adapt models to the target domain(the real world)by narrowing the domain discrepancies.However,because of the large domain gap,directly aligning two distinct domains without considering the intermediates leads to inefficient alignment and inferior adaptation.To address this issue,we propose a novel learnable evolutionary Category Intermediates(CIs)guided UDA model named Leci,which enables the information transfer between the two domains via two processes,i.e.,Distilling and Blending.Starting from a random initialization,the CIs learn shared category-wise semantics automatically from two domains in the Distilling process.Then,the learned semantics in the CIs are sent back to blend the domain features through a residual attentive fusion(RAF)module,such that the categorywise features of both domains shift towards each other.As the CIs progressively and consistently learn from the varying feature distributions during training,they are evolutionary to guide the model to achieve category-wise feature alignment.Experiments on both GTA5 and SYNTHIA datasets demonstrate Leci's superiority over prior representative methods.展开更多
Detecting sitting posture abnormalities in wheelchair users enables early identification of changes in their functional status.To date,this detection has relied on in-person observation by medical specialists.However,...Detecting sitting posture abnormalities in wheelchair users enables early identification of changes in their functional status.To date,this detection has relied on in-person observation by medical specialists.However,given the challenges faced by health specialists to carry out continuous monitoring,the development of an intelligent anomaly detection system is proposed.Unlike other authors,where they use supervised techniques,this work proposes using unsupervised techniques due to the advantages they offer.These advantages include the lack of prior labeling of data,and the detection of anomalies previously not contemplated,among others.In the present work,an individualized methodology consisting of two phases is developed:characterizing the normal sitting pattern and determining abnormal samples.An analysis has been carried out between different unsupervised techniques to study which ones are more suitable for postural diagnosis.It can be concluded,among other aspects,that the utilization of dimensionality reduction techniques leads to improved results.Moreover,the normality characterization phase is deemed necessary for enhancing the system’s learning capabilities.Additionally,employing an individualized approach to the model aids in capturing the particularities of the various pathologies present among subjects.展开更多
Climate downscaling is used to transform large-scale meteorological data into small-scale data with enhanced detail,which finds wide applications in climate modeling,numerical weather forecasting,and renewable energy....Climate downscaling is used to transform large-scale meteorological data into small-scale data with enhanced detail,which finds wide applications in climate modeling,numerical weather forecasting,and renewable energy.Although deeplearning-based downscaling methods effectively capture the complex nonlinear mapping between meteorological data of varying scales,the supervised deep-learning-based downscaling methods suffer from insufficient high-resolution data in practice,and unsupervised methods struggle with accurately inferring small-scale specifics from limited large-scale inputs due to small-scale uncertainty.This article presents DualDS,a dual-learning framework utilizing a Generative Adversarial Network–based neural network and subgrid-scale auxiliary information for climate downscaling.Such a learning method is unified in a two-stream framework through up-and downsamplers,where the downsampler is used to simulate the information loss process during the upscaling,and the upsampler is used to reconstruct lost details and correct errors incurred during the upscaling.This dual learning strategy can eliminate the dependence on high-resolution ground truth data in the training process and refine the downscaling results by constraining the mapping process.Experimental findings demonstrate that DualDS is comparable to several state-of-the-art deep learning downscaling approaches,both qualitatively and quantitatively.Specifically,for a single surface-temperature data downscaling task,our method is comparable with other unsupervised algorithms with the same dataset,and we can achieve a 0.469 dB higher peak signal-to-noise ratio,0.017 higher structural similarity,0.08 lower RMSE,and the best correlation coefficient.In summary,this paper presents a novel approach to addressing small-scale uncertainty issues in unsupervised downscaling processes.展开更多
The original monitoring data from aero-engines possess characteristics such as high dimen-sionality,strong noise,and imbalance,which present substantial challenges to traditional anomalydetection methods.In response,t...The original monitoring data from aero-engines possess characteristics such as high dimen-sionality,strong noise,and imbalance,which present substantial challenges to traditional anomalydetection methods.In response,this paper proposes a method based on Fuzzy Fusion of variablesand Discriminant mapping of features for Clustering(FFD-Clustering)to detect anomalies in originalmonitoring data from Aircraft Communication Addressing and Reporting System(ACARS).Firstly,associated variables are fuzzily grouped to extract the underlying distribution characteristics and trendsfrom the data.Secondly,a multi-layer contrastive denoising-based feature Fusion Encoding Network(FEN)is designed for each variable group,which can construct representative features for each variablegroup through eliminating strong noise and complex interrelations between variables.Thirdly,a featureDiscriminative Mapping Network(DMN)based on reconstruction difference re-clustering is designed,which can distinguish dissimilar feature vectors when mapping representative features to a unified fea-ture space.Finally,the K-means clustering is used to detect the abnormal feature vectors in the unifiedfeature space.Additionally,the algorithm is capable of reconstructing identified abnormal vectors,thereby locating the abnormal variable groups.The performance of this algorithm was tested ontwo public datasets and real original monitoring data from four aero-engines'ACARS,demonstratingits superiority and application potential in aero-engine anomaly detection.展开更多
Time series anomaly detection is crucial in finance,healthcare,and industrial monitoring.However,traditional methods often face challenges when handling time series data,such as limited feature extraction capability,p...Time series anomaly detection is crucial in finance,healthcare,and industrial monitoring.However,traditional methods often face challenges when handling time series data,such as limited feature extraction capability,poor temporal dependency handling,and suboptimal real-time performance,sometimes even neglecting the temporal relationships between data.To address these issues and improve anomaly detection performance by better capturing temporal dependencies,we propose an unsupervised time series anomaly detection method,VLT-Anomaly.First,we enhance the Variational Autoencoder(VAE)module by redesigning its network structure to better suit anomaly detection through data reconstruction.We introduce hyperparameters to control the weight of the Kullback-Leibler(KL)divergence term in the Evidence Lower Bound(ELBO),thereby improving the encoder module’s decoupling and expressive power in the latent space,which yields more effective latent representations of the data.Next,we incorporate transformer and Long Short-Term Memory(LSTM)modules to estimate the long-term dependencies of the latent representations,capturing both forward and backward temporal relationships and performing time series forecasting.Finally,we compute the reconstruction error by averaging the predicted results and decoder reconstruction and detect anomalies through grid search for optimal threshold values.Experimental results demonstrate that the proposed method performs superior anomaly detection on multiple public time series datasets,effectively extracting complex time-related features and enabling efficient computation and real-time anomaly detection.It improves detection accuracy and robustness while reducing false positives and false negatives.展开更多
Reliable electricity infrastructure is critical for modern society,highlighting the importance of securing the stability of fundamental power electronic systems.However,as such systems frequently involve high-current ...Reliable electricity infrastructure is critical for modern society,highlighting the importance of securing the stability of fundamental power electronic systems.However,as such systems frequently involve high-current and high-voltage conditions,there is a greater likelihood of failures.Consequently,anomaly detection of power electronic systems holds great significance,which is a task that properly-designed neural networks can well undertake,as proven in various scenarios.Transformer-like networks are promising for such application,yet with its structure initially designed for different tasks,features extracted by beginning layers are often lost,decreasing detection performance.Also,such data-driven methods typically require sufficient anomalous data for training,which could be difficult to obtain in practice.Therefore,to improve feature utilization while achieving efficient unsupervised learning,a novel model,Densely-connected Decoder Transformer(DDformer),is proposed for unsupervised anomaly detection of power electronic systems in this paper.First,efficient labelfree training is achieved based on the concept of autoencoder with recursive-free output.An encoder-decoder structure with densely-connected decoder is then adopted,merging features from all encoder layers to avoid possible loss of mined features while reducing training difficulty.Both simulation and real-world experiments are conducted to validate the capabilities of DDformer,and the average FDR has surpassed baseline models,reaching 89.39%,93.91%,95.98%in different experiment setups respectively.展开更多
Background Zonal application maps are designed to represent field variability using key variables that can be translated into tailored management practices.For cotton,zonal maps for crop growth regulator(CGR)applicati...Background Zonal application maps are designed to represent field variability using key variables that can be translated into tailored management practices.For cotton,zonal maps for crop growth regulator(CGR)applications under variable-rate(VR)strategies are commonly based exclusively on vegetation indices(VIs)variability.However,VIs often saturate in dense crop vegetation areas,limiting their effectiveness in distinguishing variability in crop growth.This study aimed to compare unsupervised framework(UF)and supervised framework(SUF)approaches for generat-ing zonal application maps for CGR under VR conditions.During 2022-2023 agricultural seasons,an UF was employed to generate zonal maps based on locally collected field data on plant height of cotton,satellite imagery,soil texture,and phenology data.Subsequently,a SUF(based on historical data between 2020-2021 to 2022-2023 agricultural seasons)was developed to predict plant height using remote sensing and phenology data,aiming to replicate same zonal maps but without relying on direct field measurements of plant height.Both approaches were tested in three fields and on two different dates per field.Results The predictive model for plant height of SUF performed well,as indicated by the model metrics.However,when comparing zonal application maps for specific field-date combinations,the predicted plant height exhibited lower variability compared with field measurements.This led to variable compatibility between SUF maps,which utilized the model predictions,and the UF maps,which were based on the real field data.Fields characterized by much pronounced soil texture variability yielded the highest compatibility between the zonal application maps produced by both SUF and UF approaches.This was predominantly due to the greater consistency in estimating plant development patterns within these heterogeneous field environments.While VR application approach can facilitate product savings during the application operation,other key factors must be considered.These include the availability of specialized machinery required for this type of applications,as well as the inherent operational costs associated with applying a single CGR product which differs from the typical uniform rate applications that often integrate multi-ple inputs.Conclusion Predictive modeling shows promise for assisting in the creation of zonal application maps for VR of CGR applications.However,the degree of agreement with the actual variability in crop growth found in the field should be evaluated on a field-by-field basis.The SUF approach,which is based on plant heigh prediction,demonstrated potential for supporting the development of zonal application maps for VR of CGR applications.However,the degree to which this approach aligns itself with the actual variability in crop growth observed in the field may vary,necessi-tating field-by-field evaluation.展开更多
Side-channel analysis(SCA)has emerged as a research hotspot in the field of cryptanalysis.Among various approaches,unsupervised deep learning-based methods demonstrate powerful information extraction capabilities with...Side-channel analysis(SCA)has emerged as a research hotspot in the field of cryptanalysis.Among various approaches,unsupervised deep learning-based methods demonstrate powerful information extraction capabilities without requiring labeled data.However,existing unsupervised methods,particularly those represented by differential deep learning analysis(DDLA)and its improved variants,while overcoming the dependency on labeled data inherent in template analysis,still suffer from high time complexity and training costs when handling key byte difference comparisons.To address this issue,this paper introduces invariant information clustering(IIC)into SCA for the first time,and thus proposes a novel unsupervised learning-based SCA method,named IIC-SCA.By leveraging mutual information maximization techniques for automatic feature extraction of power leakage data,our approach achieves key recovery through a single training session,eliminating the prohibitive computational overhead of traditional methods that require separate training for all possible key bytes.Experimental results on the ASCAD dataset demonstrate successful key extraction using only 50000 training traces and 2000 attack traces.Furthermore,compared with DDLA,the proposed method reduces training time by approximately 93.40%and memory consumption by about 6.15%,significantly decreasing the temporal and resource costs of unsupervised SCA.This breakthrough provides new insights for developing low-cost,high-efficiency cryptographic attack methodologies.展开更多
Skin cancer,a severe health threat,can spread rapidly if undetected.Therefore,early detection can lead to an advanced and efficient diagnosis,thus reducing mortality.Unsupervised classification techniques analyse exte...Skin cancer,a severe health threat,can spread rapidly if undetected.Therefore,early detection can lead to an advanced and efficient diagnosis,thus reducing mortality.Unsupervised classification techniques analyse extensive skin image datasets,identifying patterns and anomalies without prior labelling,facilitating early detection and effective diagnosis and potentially saving lives.In this study,the authors aim to explore the potential of unsupervised learning methods in classifying different types of skin lesions in dermatoscopic images.The authors aim to bridge the gap in dermatological research by introducing innovative techniques that enhance image quality and improve feature extraction.To achieve this,enhanced super-resolution generative adversarial networks(ESRGAN)was fine-tuned to strengthen the resolution of skin lesion images,making critical features more visible.The authors extracted histogram features to capture essential colour characteristics and used the Davies-Bouldin index and silhouette score to determine optimal clusters.Fine-tuned k-means clustering with Euclidean distance in the histogram feature space achieved 87.77% and 90.5% test accuracies on the ISIC2019 and HAM10000 datasets,respectively.The unsupervised approach effectively categorises skin lesions,indicating that unsupervised learning can significantly advance dermatology by enabling early detection and classification without extensive manual annotation.展开更多
In the dynamic scene of autonomous vehicles,the depth estimation of monocular cameras often faces the problem of inaccurate edge depth estimation.To solve this problem,we propose an unsupervised monocular depth estima...In the dynamic scene of autonomous vehicles,the depth estimation of monocular cameras often faces the problem of inaccurate edge depth estimation.To solve this problem,we propose an unsupervised monocular depth estimation model based on edge enhancement,which is specifically aimed at the depth perception challenge in dynamic scenes.The model consists of two core networks:a deep prediction network and a motion estimation network,both of which adopt an encoder-decoder architecture.The depth prediction network is based on the U-Net structure of ResNet18,which is responsible for generating the depth map of the scene.The motion estimation network is based on the U-Net structure of Flow-Net,focusing on the motion estimation of dynamic targets.In the decoding stage of the motion estimation network,we innovatively introduce an edge-enhanced decoder,which integrates a convolutional block attention module(CBAM)in the decoding process to enhance the recognition ability of the edge features of moving objects.In addition,we also designed a strip convolution module to improve the model’s capture efficiency of discrete moving targets.To further improve the performance of the model,we propose a novel edge regularization method based on the Laplace operator,which effectively accelerates the convergence process of themodel.Experimental results on the KITTI and Cityscapes datasets show that compared with the current advanced dynamic unsupervised monocular model,the proposed model has a significant improvement in depth estimation accuracy and convergence speed.Specifically,the rootmean square error(RMSE)is reduced by 4.8%compared with the DepthMotion algorithm,while the training convergence speed is increased by 36%,which shows the superior performance of the model in the depth estimation task in dynamic scenes.展开更多
Defect detection based on computer vision is a critical component in ensuring the quality of industrial products.However,existing detection methods encounter several challenges in practical applications,including the ...Defect detection based on computer vision is a critical component in ensuring the quality of industrial products.However,existing detection methods encounter several challenges in practical applications,including the scarcity of labeled samples,limited adaptability of pre-trained models,and the data heterogeneity in distributed environments.To address these issues,this research proposes an unsupervised defect detection method,FLAME(Federated Learning with Adaptive Multi-Model Embeddings).The method comprises three stages:(1)Feature learning stage:this work proposes FADE(Feature-Adaptive Domain-Specific Embeddings),a framework employs Gaussian noise injection to simulate defective patterns and implements a feature discriminator for defect detection,thereby enhancing the pre-trained model’s industrial imagery representation capabilities.(2)Knowledge distillation co-training stage:a multi-model feature knowledge distillation mechanism is introduced.Through feature-level knowledge transfer between the global model and historical local models,the current local model is guided to learn better feature representations from the global model.The approach prevents local models from converging to local optima and mitigates performance degradation caused by data heterogeneity.(3)Model parameter aggregation stage:participating clients utilize weighted averaging aggregation to synthesize an updated global model,facilitating efficient knowledge consolidation.Experimental results demonstrate that FADE improves the average image-level Area under the Receiver Operating Characteristic Curve(AUROC)by 7.34%compared to methods directly utilizing pre-trained models.In federated learning environments,FLAME’s multi-model feature knowledge distillation mechanism outperforms the classic FedAvg algorithm by 2.34%in average image-level AUROC,while exhibiting superior convergence properties.展开更多
Web data extraction has become a key technology for extracting valuable data from websites.At present,most extraction methods based on rule learning,visual pattern or tree matching have limited performance on complex ...Web data extraction has become a key technology for extracting valuable data from websites.At present,most extraction methods based on rule learning,visual pattern or tree matching have limited performance on complex web pages.Through ana-lyzing various statistical characteristics of HTML el-ements in web documents,this paper proposes,based on statistical features,an unsupervised web data ex-traction method—traversing the HTML DOM parse tree at first,calculating and generating the statistical matrix of the elements,and then locating data records by clustering method and heuristic rules that reveal in-herent links between the visual characteristics of the data recording areas and the statistical characteristics of the HTML nodes—which is both suitable for data records extraction of single-page and multi-pages,and it has strong generality and needs no training.The ex-periments show that the accuracy and efficiency of this method are equally better than the current data extrac-tion method.展开更多
The unsupervised vehicle re-identification task aims at identifying specific vehicles in surveillance videos without utilizing annotation information.Due to the higher similarity in appearance between vehicles compare...The unsupervised vehicle re-identification task aims at identifying specific vehicles in surveillance videos without utilizing annotation information.Due to the higher similarity in appearance between vehicles compared to pedestrians,pseudo-labels generated through clustering are ineffective in mitigating the impact of noise,and the feature distance between inter-class and intra-class has not been adequately improved.To address the aforementioned issues,we design a dual contrastive learning method based on knowledge distillation.During each iteration,we utilize a teacher model to randomly partition the entire dataset into two sub-domains based on clustering pseudo-label categories.By conducting contrastive learning between the two student models,we extract more discernible vehicle identity cues to improve the problem of imbalanced data distribution.Subsequently,we propose a context-aware pseudo label refinement strategy that leverages contextual features by progressively associating granularity information from different bottleneck blocks.To produce more trustworthy pseudo-labels and lessen noise interference during the clustering process,the context-aware scores are obtained by calculating the similarity between global features and contextual ones,which are subsequently added to the pseudo-label encoding process.The proposed method has achieved excellent performance in overcoming label noise and optimizing data distribution through extensive experimental results on publicly available datasets.展开更多
Representation learning from unlabeled skeleton data is a challenging task.Prior unsupervised learning algorithms mainly rely on the modeling ability of recurrent neural networks to extract the action representations....Representation learning from unlabeled skeleton data is a challenging task.Prior unsupervised learning algorithms mainly rely on the modeling ability of recurrent neural networks to extract the action representations.However,the structural information of the skeleton data,which also plays a critical role in action recognition,is rarely explored in existing unsupervised methods.To deal with this limitation,we propose a novel twostream autoencoder network to combine the topological information with temporal information of skeleton data.Specifically,we encode the graph structure by graph convolutional network(GCN)and integrate the extracted GCN-based representations into the gate recurrent unit stream.Then we design a transfer module to merge the representations of the two streams adaptively.According to the characteristics of the two-stream autoencoder,a unified loss function composed of multiple tasks is proposed to update the learnable parameters of our model.Comprehensive experiments on NW-UCLA,UWA3D,and NTU-RGBD 60 datasets demonstrate that our proposed method can achieve an excellent performance among the unsupervised skeleton-based methods and even perform a similar or superior performance over numerous supervised skeleton-based methods.展开更多
Unsupervised vehicle re-identification(Re-ID)methods have garnered widespread attention due to their potential in real-world traffic monitoring.However,existing unsupervised domain adaptation techniques often rely on ...Unsupervised vehicle re-identification(Re-ID)methods have garnered widespread attention due to their potential in real-world traffic monitoring.However,existing unsupervised domain adaptation techniques often rely on pseudo-labels generated from the source domain,which struggle to effectively address the diversity and dynamic nature of real-world scenarios.Given the limited variety of common vehicle types,enhancing the model’s generalization capability across these types is crucial.To this end,an innovative approach called meta-type generalization(MTG)is proposed.By dividing the training data into meta-train and meta-test sets based on vehicle type information,a novel gradient interaction computation strategy is designed to enhance the model’s ability to learn typeinvariant features.Integrated into the ResNet50 backbone,the MTG model achieves improvements of 4.50%and 12.04%on the Veri-776 and VRAI datasets,respectively,compared with traditional unsupervised algorithms,and surpasses current state-of-the-art methods.This achievement holds promise for application in intelligent traffic systems,enabling more efficient urban traffic solutions.展开更多
Feature selection (FS) is a process to select features which are more informative. It is one of the important steps in knowledge discovery. The problem is that not all features are important. Some of the features ma...Feature selection (FS) is a process to select features which are more informative. It is one of the important steps in knowledge discovery. The problem is that not all features are important. Some of the features may be redundant, and others may be irrelevant and noisy. The conventional supervised FS methods evaluate various feature subsets using an evaluation function or metric to select only those features which are related to the decision classes of the data under consideration. However, for many data mining applications, decision class labels are often unknown or incomplete, thus indicating the significance of unsupervised feature selection. However, in unsupervised learning, decision class labels are not provided. In this paper, we propose a new unsupervised quick reduct (QR) algorithm using rough set theory. The quality of the reduced data is measured by the classification performance and it is evaluated using WEKA classifier tool. The method is compared with existing supervised methods and the result demonstrates the efficiency of the proposed algorithm.展开更多
This paper presents a fuzzy logic approach to efficiently perform unsupervised character classification for improvement in robustness, correctness and speed of a character recognition system. The characters are first ...This paper presents a fuzzy logic approach to efficiently perform unsupervised character classification for improvement in robustness, correctness and speed of a character recognition system. The characters are first split into eight typographical categories. The classification scheme uses pattern matching to classify the characters in each category into a set of fuzzy prototypes based on a nonlinear weighted similarity function. The fuzzy unsupervised character classification, which is natural in the repre...展开更多
文摘In modern industrial production,foreign object detection in complex environments is crucial to ensure product quality and production safety.Detection systems based on deep-learning image processing algorithms often face challenges with handling high-resolution images and achieving accurate detection against complex backgrounds.To address these issues,this study employs the PatchCore unsupervised anomaly detection algorithm combined with data augmentation techniques to enhance the system’s generalization capability across varying lighting conditions,viewing angles,and object scales.The proposed method is evaluated in a complex industrial detection scenario involving the bogie of an electric multiple unit(EMU).A dataset consisting of complex backgrounds,diverse lighting conditions,and multiple viewing angles is constructed to validate the performance of the detection system in real industrial environments.Experimental results show that the proposed model achieves an average area under the receiver operating characteristic curve(AUROC)of 0.92 and an average F1 score of 0.85.Combined with data augmentation,the proposed model exhibits improvements in AUROC by 0.06 and F1 score by 0.03,demonstrating enhanced accuracy and robustness for foreign object detection in complex industrial settings.In addition,the effects of key factors on detection performance are systematically analyzed,providing practical guidance for parameter selection in real industrial applications.
基金supported by the National Natural Science Foundation of China(Grant Nos.52375255,51935007)the Shanghai Rising-Star Program(Grant No.24QB2705000)。
文摘Existing elevator fault diagnosis algorithms have limited engineering applicability due to variations in working conditions and differences in equipment structures.To address this limitation,this study proposes an unsupervised subdomain adaptation method based on a time-frequency feature attention mechanism,LMMD-based subdomain alignment,and contrastive local alignment.This enables the application of the diagnosis model across different working conditions and equipment types.First,a novel time-frequency feature attention mechanism assigns weights to vibration signals of varying dimensions.Second,the time series is transformed to obtain a three-channel time-frequency diagram.This diagram is input into the proposed dimension-segmentation cross-channel multihead self-attention framework to extract high-dimensional frequencydomain fault features.These features are concatenated with the time-domain features to obtain a global feature representation.Then,the extracted high-dimensional features are sent to the classification module to obtain the predicted labels for the source and target domains.Finally,after confidence filtering,the true labels from the source domain and the prediction labels from the target domain are fed into a dynamically weighted multilevel feature alignment module to promote proximity between similar fault features across domains while enhancing separation among different fault types.The validity and superiority of the proposed method were demonstrated through simulation experiments conducted on two types of manned escalator systems under multiple working conditions.For the most challenging transfer task,the proposed method achieved higher accuracy on the target domain test set than DANN,ADDA,C-CLCN,TFA-CCN,and TFA-LCN by 26.87%,24.72%,11.44%,28.94%,and 16.85%,respectively.
基金supported by the Natural Science Foundation of China Project(No.62525102).
文摘Single-pixel imaging(SPI)is a prominent scattering media imaging technique that allows image transmission via one-dimensional detection under structured illumination,with applications spanning from long-range imaging to microscopy.Recent advancements leveraging deep learning(DL)have significantly improved SPI performance,especially at low compression ratios.However,most DL-based SPI methods proposed so far rely heavily on extensive labeled datasets for supervised training,which are often impractical in real-world scenarios.Here,we propose an unsupervised learningenabled label-free SPI method for resilient information transmission through unknown dynamic scattering media.Additionally,we introduce a physics-informed autoencoder framework to optimize encoding schemes,further enhancing image quality at low compression ratios.Simulation and experimental results demonstrate that high-efficiency data transmission with structural similarity exceeding 0.9 is achieved through challenging turbulent channels.Moreover,experiments demonstrate that in a 5 m underwater dynamic turbulent channel,USAF target imaging quality surpasses traditional methods by over 13 dB.The compressive encoded transmission of 720×720 resolution video exceeding 30 seconds with great fidelity is also successfully demonstrated.These preliminary results suggest that our proposed method opens up a new paradigm for resilient information transmission through unknown dynamic scattering media and holds potential for broader applications within many other scattering media imaging technologies.
基金Australian Research Council Project(FL-170100117).
文摘To avoid the laborious annotation process for dense prediction tasks like semantic segmentation,unsupervised domain adaptation(UDA)methods have been proposed to leverage the abundant annotations from a source domain,such as virtual world(e.g.,3D games),and adapt models to the target domain(the real world)by narrowing the domain discrepancies.However,because of the large domain gap,directly aligning two distinct domains without considering the intermediates leads to inefficient alignment and inferior adaptation.To address this issue,we propose a novel learnable evolutionary Category Intermediates(CIs)guided UDA model named Leci,which enables the information transfer between the two domains via two processes,i.e.,Distilling and Blending.Starting from a random initialization,the CIs learn shared category-wise semantics automatically from two domains in the Distilling process.Then,the learned semantics in the CIs are sent back to blend the domain features through a residual attentive fusion(RAF)module,such that the categorywise features of both domains shift towards each other.As the CIs progressively and consistently learn from the varying feature distributions during training,they are evolutionary to guide the model to achieve category-wise feature alignment.Experiments on both GTA5 and SYNTHIA datasets demonstrate Leci's superiority over prior representative methods.
基金FEDER/Ministry of Science and Innovation-State Research Agency/Project PID2020-112667RB-I00 funded by MCIN/AEI/10.13039/501100011033the Basque Government,IT1726-22+2 种基金by the predoctoral contracts PRE_2022_2_0022 and EP_2023_1_0015 of the Basque Governmentpartially supported by the Italian MIUR,PRIN 2020 Project“COMMON-WEARS”,N.2020HCWWLP,CUP:H23C22000230005co-funding from Next Generation EU,in the context of the National Recovery and Resilience Plan,through the Italian MUR,PRIN 2022 Project”COCOWEARS”(A framework for COntinuum COmputing WEARable Systems),N.2022T2XNJE,CUP:H53D23003640006.
文摘Detecting sitting posture abnormalities in wheelchair users enables early identification of changes in their functional status.To date,this detection has relied on in-person observation by medical specialists.However,given the challenges faced by health specialists to carry out continuous monitoring,the development of an intelligent anomaly detection system is proposed.Unlike other authors,where they use supervised techniques,this work proposes using unsupervised techniques due to the advantages they offer.These advantages include the lack of prior labeling of data,and the detection of anomalies previously not contemplated,among others.In the present work,an individualized methodology consisting of two phases is developed:characterizing the normal sitting pattern and determining abnormal samples.An analysis has been carried out between different unsupervised techniques to study which ones are more suitable for postural diagnosis.It can be concluded,among other aspects,that the utilization of dimensionality reduction techniques leads to improved results.Moreover,the normality characterization phase is deemed necessary for enhancing the system’s learning capabilities.Additionally,employing an individualized approach to the model aids in capturing the particularities of the various pathologies present among subjects.
基金supported by the following funding bodies:the National Key Research and Development Program of China(Grant No.2020YFA0608000)National Science Foundation of China(Grant Nos.42075142,42375148,42125503+2 种基金42130608)FY-APP-2022.0609,Sichuan Province Key Tech nology Research and Development project(Grant Nos.2024ZHCG0168,2024ZHCG0176,2023YFG0305,2023YFG-0124,and 23ZDYF0091)the CUIT Science and Technology Innovation Capacity Enhancement Program project(Grant No.KYQN202305)。
文摘Climate downscaling is used to transform large-scale meteorological data into small-scale data with enhanced detail,which finds wide applications in climate modeling,numerical weather forecasting,and renewable energy.Although deeplearning-based downscaling methods effectively capture the complex nonlinear mapping between meteorological data of varying scales,the supervised deep-learning-based downscaling methods suffer from insufficient high-resolution data in practice,and unsupervised methods struggle with accurately inferring small-scale specifics from limited large-scale inputs due to small-scale uncertainty.This article presents DualDS,a dual-learning framework utilizing a Generative Adversarial Network–based neural network and subgrid-scale auxiliary information for climate downscaling.Such a learning method is unified in a two-stream framework through up-and downsamplers,where the downsampler is used to simulate the information loss process during the upscaling,and the upsampler is used to reconstruct lost details and correct errors incurred during the upscaling.This dual learning strategy can eliminate the dependence on high-resolution ground truth data in the training process and refine the downscaling results by constraining the mapping process.Experimental findings demonstrate that DualDS is comparable to several state-of-the-art deep learning downscaling approaches,both qualitatively and quantitatively.Specifically,for a single surface-temperature data downscaling task,our method is comparable with other unsupervised algorithms with the same dataset,and we can achieve a 0.469 dB higher peak signal-to-noise ratio,0.017 higher structural similarity,0.08 lower RMSE,and the best correlation coefficient.In summary,this paper presents a novel approach to addressing small-scale uncertainty issues in unsupervised downscaling processes.
基金co-supported by the National Science and Technology Major Project,China(No.J2019-I-0001-0001)the National Natural Science Foundation of China(No.52105545)。
文摘The original monitoring data from aero-engines possess characteristics such as high dimen-sionality,strong noise,and imbalance,which present substantial challenges to traditional anomalydetection methods.In response,this paper proposes a method based on Fuzzy Fusion of variablesand Discriminant mapping of features for Clustering(FFD-Clustering)to detect anomalies in originalmonitoring data from Aircraft Communication Addressing and Reporting System(ACARS).Firstly,associated variables are fuzzily grouped to extract the underlying distribution characteristics and trendsfrom the data.Secondly,a multi-layer contrastive denoising-based feature Fusion Encoding Network(FEN)is designed for each variable group,which can construct representative features for each variablegroup through eliminating strong noise and complex interrelations between variables.Thirdly,a featureDiscriminative Mapping Network(DMN)based on reconstruction difference re-clustering is designed,which can distinguish dissimilar feature vectors when mapping representative features to a unified fea-ture space.Finally,the K-means clustering is used to detect the abnormal feature vectors in the unifiedfeature space.Additionally,the algorithm is capable of reconstructing identified abnormal vectors,thereby locating the abnormal variable groups.The performance of this algorithm was tested ontwo public datasets and real original monitoring data from four aero-engines'ACARS,demonstratingits superiority and application potential in aero-engine anomaly detection.
基金support from the Fundamental Research Funds for Central Public Welfare Research Institutes(SK202324)the Central Guidance on Local Science and Technology Development Fund of Hebei Province(236Z0104G)+1 种基金the National Natural Science Foundation of China(62476078)the Geological Survey Project of China Geological Survey(G202304-2).
文摘Time series anomaly detection is crucial in finance,healthcare,and industrial monitoring.However,traditional methods often face challenges when handling time series data,such as limited feature extraction capability,poor temporal dependency handling,and suboptimal real-time performance,sometimes even neglecting the temporal relationships between data.To address these issues and improve anomaly detection performance by better capturing temporal dependencies,we propose an unsupervised time series anomaly detection method,VLT-Anomaly.First,we enhance the Variational Autoencoder(VAE)module by redesigning its network structure to better suit anomaly detection through data reconstruction.We introduce hyperparameters to control the weight of the Kullback-Leibler(KL)divergence term in the Evidence Lower Bound(ELBO),thereby improving the encoder module’s decoupling and expressive power in the latent space,which yields more effective latent representations of the data.Next,we incorporate transformer and Long Short-Term Memory(LSTM)modules to estimate the long-term dependencies of the latent representations,capturing both forward and backward temporal relationships and performing time series forecasting.Finally,we compute the reconstruction error by averaging the predicted results and decoder reconstruction and detect anomalies through grid search for optimal threshold values.Experimental results demonstrate that the proposed method performs superior anomaly detection on multiple public time series datasets,effectively extracting complex time-related features and enabling efficient computation and real-time anomaly detection.It improves detection accuracy and robustness while reducing false positives and false negatives.
基金supported in part by the National Natural Science Foundation of China under Grant 62303090,U2330206in part by the Postdoctoral Science Foundation of China under Grant 2023M740516+1 种基金in part by the Natural Science Foundation of Sichuan Province under Grant 2024NSFSC1480in part by the New Cornerstone Science Foundation through the XPLORER PRIZE.
文摘Reliable electricity infrastructure is critical for modern society,highlighting the importance of securing the stability of fundamental power electronic systems.However,as such systems frequently involve high-current and high-voltage conditions,there is a greater likelihood of failures.Consequently,anomaly detection of power electronic systems holds great significance,which is a task that properly-designed neural networks can well undertake,as proven in various scenarios.Transformer-like networks are promising for such application,yet with its structure initially designed for different tasks,features extracted by beginning layers are often lost,decreasing detection performance.Also,such data-driven methods typically require sufficient anomalous data for training,which could be difficult to obtain in practice.Therefore,to improve feature utilization while achieving efficient unsupervised learning,a novel model,Densely-connected Decoder Transformer(DDformer),is proposed for unsupervised anomaly detection of power electronic systems in this paper.First,efficient labelfree training is achieved based on the concept of autoencoder with recursive-free output.An encoder-decoder structure with densely-connected decoder is then adopted,merging features from all encoder layers to avoid possible loss of mined features while reducing training difficulty.Both simulation and real-world experiments are conducted to validate the capabilities of DDformer,and the average FDR has surpassed baseline models,reaching 89.39%,93.91%,95.98%in different experiment setups respectively.
文摘Background Zonal application maps are designed to represent field variability using key variables that can be translated into tailored management practices.For cotton,zonal maps for crop growth regulator(CGR)applications under variable-rate(VR)strategies are commonly based exclusively on vegetation indices(VIs)variability.However,VIs often saturate in dense crop vegetation areas,limiting their effectiveness in distinguishing variability in crop growth.This study aimed to compare unsupervised framework(UF)and supervised framework(SUF)approaches for generat-ing zonal application maps for CGR under VR conditions.During 2022-2023 agricultural seasons,an UF was employed to generate zonal maps based on locally collected field data on plant height of cotton,satellite imagery,soil texture,and phenology data.Subsequently,a SUF(based on historical data between 2020-2021 to 2022-2023 agricultural seasons)was developed to predict plant height using remote sensing and phenology data,aiming to replicate same zonal maps but without relying on direct field measurements of plant height.Both approaches were tested in three fields and on two different dates per field.Results The predictive model for plant height of SUF performed well,as indicated by the model metrics.However,when comparing zonal application maps for specific field-date combinations,the predicted plant height exhibited lower variability compared with field measurements.This led to variable compatibility between SUF maps,which utilized the model predictions,and the UF maps,which were based on the real field data.Fields characterized by much pronounced soil texture variability yielded the highest compatibility between the zonal application maps produced by both SUF and UF approaches.This was predominantly due to the greater consistency in estimating plant development patterns within these heterogeneous field environments.While VR application approach can facilitate product savings during the application operation,other key factors must be considered.These include the availability of specialized machinery required for this type of applications,as well as the inherent operational costs associated with applying a single CGR product which differs from the typical uniform rate applications that often integrate multi-ple inputs.Conclusion Predictive modeling shows promise for assisting in the creation of zonal application maps for VR of CGR applications.However,the degree of agreement with the actual variability in crop growth found in the field should be evaluated on a field-by-field basis.The SUF approach,which is based on plant heigh prediction,demonstrated potential for supporting the development of zonal application maps for VR of CGR applications.However,the degree to which this approach aligns itself with the actual variability in crop growth observed in the field may vary,necessi-tating field-by-field evaluation.
文摘Side-channel analysis(SCA)has emerged as a research hotspot in the field of cryptanalysis.Among various approaches,unsupervised deep learning-based methods demonstrate powerful information extraction capabilities without requiring labeled data.However,existing unsupervised methods,particularly those represented by differential deep learning analysis(DDLA)and its improved variants,while overcoming the dependency on labeled data inherent in template analysis,still suffer from high time complexity and training costs when handling key byte difference comparisons.To address this issue,this paper introduces invariant information clustering(IIC)into SCA for the first time,and thus proposes a novel unsupervised learning-based SCA method,named IIC-SCA.By leveraging mutual information maximization techniques for automatic feature extraction of power leakage data,our approach achieves key recovery through a single training session,eliminating the prohibitive computational overhead of traditional methods that require separate training for all possible key bytes.Experimental results on the ASCAD dataset demonstrate successful key extraction using only 50000 training traces and 2000 attack traces.Furthermore,compared with DDLA,the proposed method reduces training time by approximately 93.40%and memory consumption by about 6.15%,significantly decreasing the temporal and resource costs of unsupervised SCA.This breakthrough provides new insights for developing low-cost,high-efficiency cryptographic attack methodologies.
文摘Skin cancer,a severe health threat,can spread rapidly if undetected.Therefore,early detection can lead to an advanced and efficient diagnosis,thus reducing mortality.Unsupervised classification techniques analyse extensive skin image datasets,identifying patterns and anomalies without prior labelling,facilitating early detection and effective diagnosis and potentially saving lives.In this study,the authors aim to explore the potential of unsupervised learning methods in classifying different types of skin lesions in dermatoscopic images.The authors aim to bridge the gap in dermatological research by introducing innovative techniques that enhance image quality and improve feature extraction.To achieve this,enhanced super-resolution generative adversarial networks(ESRGAN)was fine-tuned to strengthen the resolution of skin lesion images,making critical features more visible.The authors extracted histogram features to capture essential colour characteristics and used the Davies-Bouldin index and silhouette score to determine optimal clusters.Fine-tuned k-means clustering with Euclidean distance in the histogram feature space achieved 87.77% and 90.5% test accuracies on the ISIC2019 and HAM10000 datasets,respectively.The unsupervised approach effectively categorises skin lesions,indicating that unsupervised learning can significantly advance dermatology by enabling early detection and classification without extensive manual annotation.
基金funded by the Yangtze River Delta Science and Technology Innovation Community Joint Research Project(2023CSJGG1600)the Natural Science Foundation of Anhui Province(2208085MF173)Wuhu“ChiZhu Light”Major Science and Technology Project(2023ZD01,2023ZD03).
文摘In the dynamic scene of autonomous vehicles,the depth estimation of monocular cameras often faces the problem of inaccurate edge depth estimation.To solve this problem,we propose an unsupervised monocular depth estimation model based on edge enhancement,which is specifically aimed at the depth perception challenge in dynamic scenes.The model consists of two core networks:a deep prediction network and a motion estimation network,both of which adopt an encoder-decoder architecture.The depth prediction network is based on the U-Net structure of ResNet18,which is responsible for generating the depth map of the scene.The motion estimation network is based on the U-Net structure of Flow-Net,focusing on the motion estimation of dynamic targets.In the decoding stage of the motion estimation network,we innovatively introduce an edge-enhanced decoder,which integrates a convolutional block attention module(CBAM)in the decoding process to enhance the recognition ability of the edge features of moving objects.In addition,we also designed a strip convolution module to improve the model’s capture efficiency of discrete moving targets.To further improve the performance of the model,we propose a novel edge regularization method based on the Laplace operator,which effectively accelerates the convergence process of themodel.Experimental results on the KITTI and Cityscapes datasets show that compared with the current advanced dynamic unsupervised monocular model,the proposed model has a significant improvement in depth estimation accuracy and convergence speed.Specifically,the rootmean square error(RMSE)is reduced by 4.8%compared with the DepthMotion algorithm,while the training convergence speed is increased by 36%,which shows the superior performance of the model in the depth estimation task in dynamic scenes.
基金supported in part by the National Natural Science Foundation of China under Grants 32171909,52205254,32301704the Guangdong Basic and Applied Basic Research Foundation under Grants 2023A1515011255,2024A1515010199+1 种基金the Scientific Research Projects of Universities in Guangdong Province under Grants 2024ZDZX1042,2024ZDZX3057the Ji-Hua Laboratory Open Project under Grant X220931UZ230.
文摘Defect detection based on computer vision is a critical component in ensuring the quality of industrial products.However,existing detection methods encounter several challenges in practical applications,including the scarcity of labeled samples,limited adaptability of pre-trained models,and the data heterogeneity in distributed environments.To address these issues,this research proposes an unsupervised defect detection method,FLAME(Federated Learning with Adaptive Multi-Model Embeddings).The method comprises three stages:(1)Feature learning stage:this work proposes FADE(Feature-Adaptive Domain-Specific Embeddings),a framework employs Gaussian noise injection to simulate defective patterns and implements a feature discriminator for defect detection,thereby enhancing the pre-trained model’s industrial imagery representation capabilities.(2)Knowledge distillation co-training stage:a multi-model feature knowledge distillation mechanism is introduced.Through feature-level knowledge transfer between the global model and historical local models,the current local model is guided to learn better feature representations from the global model.The approach prevents local models from converging to local optima and mitigates performance degradation caused by data heterogeneity.(3)Model parameter aggregation stage:participating clients utilize weighted averaging aggregation to synthesize an updated global model,facilitating efficient knowledge consolidation.Experimental results demonstrate that FADE improves the average image-level Area under the Receiver Operating Characteristic Curve(AUROC)by 7.34%compared to methods directly utilizing pre-trained models.In federated learning environments,FLAME’s multi-model feature knowledge distillation mechanism outperforms the classic FedAvg algorithm by 2.34%in average image-level AUROC,while exhibiting superior convergence properties.
文摘Web data extraction has become a key technology for extracting valuable data from websites.At present,most extraction methods based on rule learning,visual pattern or tree matching have limited performance on complex web pages.Through ana-lyzing various statistical characteristics of HTML el-ements in web documents,this paper proposes,based on statistical features,an unsupervised web data ex-traction method—traversing the HTML DOM parse tree at first,calculating and generating the statistical matrix of the elements,and then locating data records by clustering method and heuristic rules that reveal in-herent links between the visual characteristics of the data recording areas and the statistical characteristics of the HTML nodes—which is both suitable for data records extraction of single-page and multi-pages,and it has strong generality and needs no training.The ex-periments show that the accuracy and efficiency of this method are equally better than the current data extrac-tion method.
基金supported by the National Natural Science Foundation of China under Grant Nos.62461037,62076117 and 62166026the Jiangxi Provincial Natural Science Foundation under Grant Nos.20224BAB212011,20232BAB202051,20232BAB212008 and 20242BAB25078the Jiangxi Provincial Key Laboratory of Virtual Reality under Grant No.2024SSY03151.
文摘The unsupervised vehicle re-identification task aims at identifying specific vehicles in surveillance videos without utilizing annotation information.Due to the higher similarity in appearance between vehicles compared to pedestrians,pseudo-labels generated through clustering are ineffective in mitigating the impact of noise,and the feature distance between inter-class and intra-class has not been adequately improved.To address the aforementioned issues,we design a dual contrastive learning method based on knowledge distillation.During each iteration,we utilize a teacher model to randomly partition the entire dataset into two sub-domains based on clustering pseudo-label categories.By conducting contrastive learning between the two student models,we extract more discernible vehicle identity cues to improve the problem of imbalanced data distribution.Subsequently,we propose a context-aware pseudo label refinement strategy that leverages contextual features by progressively associating granularity information from different bottleneck blocks.To produce more trustworthy pseudo-labels and lessen noise interference during the clustering process,the context-aware scores are obtained by calculating the similarity between global features and contextual ones,which are subsequently added to the pseudo-label encoding process.The proposed method has achieved excellent performance in overcoming label noise and optimizing data distribution through extensive experimental results on publicly available datasets.
文摘Representation learning from unlabeled skeleton data is a challenging task.Prior unsupervised learning algorithms mainly rely on the modeling ability of recurrent neural networks to extract the action representations.However,the structural information of the skeleton data,which also plays a critical role in action recognition,is rarely explored in existing unsupervised methods.To deal with this limitation,we propose a novel twostream autoencoder network to combine the topological information with temporal information of skeleton data.Specifically,we encode the graph structure by graph convolutional network(GCN)and integrate the extracted GCN-based representations into the gate recurrent unit stream.Then we design a transfer module to merge the representations of the two streams adaptively.According to the characteristics of the two-stream autoencoder,a unified loss function composed of multiple tasks is proposed to update the learnable parameters of our model.Comprehensive experiments on NW-UCLA,UWA3D,and NTU-RGBD 60 datasets demonstrate that our proposed method can achieve an excellent performance among the unsupervised skeleton-based methods and even perform a similar or superior performance over numerous supervised skeleton-based methods.
基金Supported by the National Natural Science Foundation of China(No.61976098)the Natural Science Foundation for Outstanding Young Scholars of Fujian Province(No.2022J06023).
文摘Unsupervised vehicle re-identification(Re-ID)methods have garnered widespread attention due to their potential in real-world traffic monitoring.However,existing unsupervised domain adaptation techniques often rely on pseudo-labels generated from the source domain,which struggle to effectively address the diversity and dynamic nature of real-world scenarios.Given the limited variety of common vehicle types,enhancing the model’s generalization capability across these types is crucial.To this end,an innovative approach called meta-type generalization(MTG)is proposed.By dividing the training data into meta-train and meta-test sets based on vehicle type information,a novel gradient interaction computation strategy is designed to enhance the model’s ability to learn typeinvariant features.Integrated into the ResNet50 backbone,the MTG model achieves improvements of 4.50%and 12.04%on the Veri-776 and VRAI datasets,respectively,compared with traditional unsupervised algorithms,and surpasses current state-of-the-art methods.This achievement holds promise for application in intelligent traffic systems,enabling more efficient urban traffic solutions.
基金supported by the UGC, SERO, Hyderabad under FDP during XI plan periodthe UGC, New Delhi for financial assistance under major research project Grant No. F-34-105/2008
文摘Feature selection (FS) is a process to select features which are more informative. It is one of the important steps in knowledge discovery. The problem is that not all features are important. Some of the features may be redundant, and others may be irrelevant and noisy. The conventional supervised FS methods evaluate various feature subsets using an evaluation function or metric to select only those features which are related to the decision classes of the data under consideration. However, for many data mining applications, decision class labels are often unknown or incomplete, thus indicating the significance of unsupervised feature selection. However, in unsupervised learning, decision class labels are not provided. In this paper, we propose a new unsupervised quick reduct (QR) algorithm using rough set theory. The quality of the reduced data is measured by the classification performance and it is evaluated using WEKA classifier tool. The method is compared with existing supervised methods and the result demonstrates the efficiency of the proposed algorithm.
文摘This paper presents a fuzzy logic approach to efficiently perform unsupervised character classification for improvement in robustness, correctness and speed of a character recognition system. The characters are first split into eight typographical categories. The classification scheme uses pattern matching to classify the characters in each category into a set of fuzzy prototypes based on a nonlinear weighted similarity function. The fuzzy unsupervised character classification, which is natural in the repre...