Anomaly detection is an important method for intrusion detection.In recent years,unsupervised methods have been widely researched because they do not require labeling.For example,a nonlinear autoencoder can use recons...Anomaly detection is an important method for intrusion detection.In recent years,unsupervised methods have been widely researched because they do not require labeling.For example,a nonlinear autoencoder can use reconstruction errors to attain the discrimination threshold.This method is not effective when the model complexity is high or the data contains noise.The method for detecting the density of compressed features in a hidden layer can be used to reduce the influence of noise on the selection of the threshold because the density of abnormal data in hidden layers is smaller than normal data.However,compressed features may lose some of the high-dimensional distribution information of the original data.In this paper,we present an efficient anomaly detection framework for unsupervised anomaly detection,which includes network data capturing,processing,feature extraction,and anomaly detection.We employ a deep autoencoder to obtain compressed features and multi-layer reconstruction errors,and feeds them the same to the Gaussian mixture model to estimate the density.The proposed approach is trained and tested on multiple current intrusion detection datasets and real network scenes,and performance indicators,namely accuracy,recall,and F1-score,are better than other autoencoder models.展开更多
The initial noise present in the depth images obtained with RGB-D sensors is a combination of hardware limitations in addition to the environmental factors,due to the limited capabilities of sensors,which also produce...The initial noise present in the depth images obtained with RGB-D sensors is a combination of hardware limitations in addition to the environmental factors,due to the limited capabilities of sensors,which also produce poor computer vision results.The common image denoising techniques tend to remove significant image details and also remove noise,provided they are based on space and frequency filtering.The updated framework presented in this paper is a novel denoising model that makes use of Boruta-driven feature selection using a Long Short-Term Memory Autoencoder(LSTMAE).The Boruta algorithm identifies the most useful depth features that are used to maximize the spatial structure integrity and reduce redundancy.An LSTMAE is then used to process these selected features and model depth pixel sequences to generate robust,noise-resistant representations.The system uses the encoder to encode the input data into a latent space that has been compressed before it is decoded to retrieve the clean image.Experiments on a benchmark data set show that the suggested technique attains a PSNR of 45 dB and an SSIM of 0.90,which is 10 dB higher than the performance of conventional convolutional autoencoders and 15 times higher than that of the wavelet-based models.Moreover,the feature selection step will decrease the input dimensionality by 40%,resulting in a 37.5%reduction in training time and a real-time inference rate of 200 FPS.Boruta-LSTMAE framework,therefore,offers a highly efficient and scalable system for depth image denoising,with a high potential to be applied to close-range 3D systems,such as robotic manipulation and gesture-based interfaces.展开更多
Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel a...Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel autoencoder-based imputation framework that integrates a composite loss function to enhance robustness and precision.The proposed loss combines(i)a guided,masked mean squared error focusing on missing entries;(ii)a noise-aware regularization term to improve resilience against data corruption;and(iii)a variance penalty to encourage expressive yet stable reconstructions.We evaluate the proposed model across four missingness mechanisms,such as Missing Completely at Random,Missing at Random,Missing Not at Random,and Missing Not at Random with quantile censorship,under systematically varied feature counts,sample sizes,and missingness ratios ranging from 5%to 60%.Four publicly available real-world datasets(Stroke Prediction,Pima Indians Diabetes,Cardiovascular Disease,and Framingham Heart Study)were used,and the obtained results show that our proposed model consistently outperforms baseline methods,including traditional and deep learning-based techniques.An ablation study reveals the additive value of each component in the loss function.Additionally,we assessed the downstream utility of imputed data through classification tasks,where datasets imputed by the proposed method yielded the highest receiver operating characteristic area under the curve scores across all scenarios.The model demonstrates strong scalability and robustness,improving performance with larger datasets and higher feature counts.These results underscore the capacity of the proposed method to produce not only numerically accurate but also semantically useful imputations,making it a promising solution for robust data recovery in clinical applications.展开更多
The rapid growth of biomedical data,particularly multi-omics data including genomes,transcriptomics,proteomics,metabolomics,and epigenomics,medical research and clinical decision-making confront both new opportunities...The rapid growth of biomedical data,particularly multi-omics data including genomes,transcriptomics,proteomics,metabolomics,and epigenomics,medical research and clinical decision-making confront both new opportunities and obstacles.The huge and diversified nature of these datasets cannot always be managed using traditional data analysis methods.As a consequence,deep learning has emerged as a strong tool for analysing numerous omics data due to its ability to handle complex and non-linear relationships.This paper explores the fundamental concepts of deep learning and how they are used in multi-omics medical data mining.We demonstrate how autoencoders,variational autoencoders,multimodal models,attention mechanisms,transformers,and graph neural networks enable pattern analysis and recognition across all omics data.Deep learning has been found to be effective in illness classification,biomarker identification,gene network learning,and therapeutic efficacy prediction.We also consider critical problems like as data quality,model explainability,whether findings can be repeated,and computational power requirements.We now consider future elements of combining omics with clinical and imaging data,explainable AI,federated learning,and real-time diagnostics.Overall,this study emphasises the need of collaborating across disciplines to advance deep learning-based multi-omics research for precision medicine and comprehending complicated disorders.展开更多
Inverse design of advanced materials represents a pivotal challenge in materials science.Leveraging the latent space of Variational Autoencoders(VAEs)for material optimization has emerged as a significant advancement ...Inverse design of advanced materials represents a pivotal challenge in materials science.Leveraging the latent space of Variational Autoencoders(VAEs)for material optimization has emerged as a significant advancement in the field of material inverse design.However,VAEs are inherently prone to generating blurred images,posing challenges for precise inverse design and microstructure manufacturing.While increasing the dimensionality of the VAE latent space can mitigate reconstruction blurriness to some extent,it simultaneously imposes a substantial burden on target optimization due to an excessively high search space.To address these limitations,this study adopts a Variational Autoencoder guided Conditional Diffusion Generative Model(VAE-CDGM)framework integrated with Bayesian optimization to achieve the inverse design of composite materials with targeted mechanical properties.The VAE-CDGM model synergizes the strengths of VAEs and Denoising Diffusion Probabilistic Models(DDPM),enabling the generation of high-quality,sharp images while preserving a manipulable latent space.To accommodate varying dimensional requirements of the latent space,two optimization strategies are proposed.When the latent space dimensionality is excessively high,SHapley Additive exPlanations(SHAP)sensitivity analysis is employed to identify critical latent features for optimization within a reduced subspace.Conversely,direct optimization is performed in the low-dimensional latent space of VAE-CDGM when dimensionality is modest.The results demonstrate that both strategies accurately achieve the targeted design of composite materials while circumventing the blurred reconstruction flaws of VAEs,which offers a novel pathway for the precise design of advanced materials.展开更多
Automated Program Repair(APR)techniques have shown significant potential in mitigating the cost and complexity associated with debugging by automatically generating corrective patches for software defects.Despite cons...Automated Program Repair(APR)techniques have shown significant potential in mitigating the cost and complexity associated with debugging by automatically generating corrective patches for software defects.Despite considerable progress in APR methodologies,existing approaches frequently lack contextual awareness of runtime behaviors and structural intricacies inherent in buggy source code.In this paper,we propose a novel APR approach that integrates attention mechanisms within an autoencoder-based framework,explicitly utilizing structural code affinity and execution context correlation derived from stack trace analysis.Our approach begins with an innovative preprocessing pipeline,where code segments and stack traces are transformed into tokenized representations.Subsequently,the BM25 ranking algorithm is employed to quantitatively measure structural code affinity and execution context correlation,identifying syntactically and semantically analogous buggy code snippets and relevant runtime error contexts from extensive repositories.These extracted features are then encoded via an attention-enhanced autoencoder model,specifically designed to capture significant patterns and correlations essential for effective patch generation.To assess the efficacy and generalizability of our proposed method,we conducted rigorous experimental comparisons against DeepFix,a state-of-the-art APR system,using a substantial dataset comprising 53,478 studentdeveloped C programs.Experimental outcomes indicate that our model achieves a notable bug repair success rate of approximately 62.36%,representing a statistically significant performance improvement of over 6%compared to the baseline.Furthermore,a thorough K-fold cross-validation reinforced the consistency,robustness,and reliability of our method across diverse subsets of the dataset.Our findings present the critical advantage of integrating attentionbased learning with code structural and execution context features in APR tasks,leading to improved accuracy and practical applicability.Future work aims to extend the model’s applicability across different programming languages,systematically optimize hyperparameters,and explore alternative feature representation methods to further enhance debugging efficiency and effectiveness.展开更多
In the field of intelligent air combat,real-time and accurate recognition of within-visual-range(WVR)maneuver actions serves as the foundational cornerstone for constructing autonomous decision-making systems.However,...In the field of intelligent air combat,real-time and accurate recognition of within-visual-range(WVR)maneuver actions serves as the foundational cornerstone for constructing autonomous decision-making systems.However,existing methods face two major challenges:traditional feature engineering suffers from insufficient effective dimensionality in the feature space due to kinematic coupling,making it difficult to distinguish essential differences between maneuvers,while end-to-end deep learning models lack controllability in implicit feature learning and fail to model high-order long-range temporal dependencies.This paper proposes a trajectory feature pre-extraction method based on a Long-range Masked Autoencoder(LMAE),incorporating three key innovations:(1)Random Fragment High-ratio Masking(RFH-Mask),which enforces the model to learn long-range temporal correlations by masking 80%of trajectory data while retaining continuous fragments;(2)Kalman Filter-Guided Objective Function(KFG-OF),integrating trajectory continuity constraints to align the feature space with kinematic principles;and(3)Two-stage Decoupled Architecture,enabling efficient and controllable feature learning through unsupervised pre-training and frozen-feature transfer.Experimental results demonstrate that LMAE significantly improves the average recognition accuracy for 20-class maneuvers compared to traditional end-to-end models,while significantly accelerating convergence speed.The contributions of this work lie in:introducing high-masking-rate autoencoders into low-informationdensity trajectory analysis,proposing a feature engineering framework with enhanced controllability and efficiency,and providing a novel technical pathway for intelligent air combat decision-making systems.展开更多
With the increasing complexity of industrial automation,planetary gearboxes play a vital role in largescale equipment transmission systems,directly impacting operational efficiency and safety.Traditional maintenance s...With the increasing complexity of industrial automation,planetary gearboxes play a vital role in largescale equipment transmission systems,directly impacting operational efficiency and safety.Traditional maintenance strategies often struggle to accurately predict the degradation process of equipment,leading to excessive maintenance costs or potential failure risks.However,existing prediction methods based on statistical models are difficult to adapt to nonlinear degradation processes.To address these challenges,this study proposes a novel condition-based maintenance framework for planetary gearboxes.A comprehensive full-lifecycle degradation experiment was conducted to collect raw vibration signals,which were then processed using a temporal convolutional network autoencoder with multi-scale perception capability to extract deep temporal degradation features,enabling the collaborative extraction of longperiod meshing frequencies and short-term impact features from the vibration signals.Kernel principal component analysis was employed to fuse and normalize these features,enhancing the characterization of degradation progression.A nonlinear Wiener process was used to model the degradation trajectory,with a threshold decay function introduced to dynamically adjust maintenance strategies,and model parameters optimized through maximum likelihood estimation.Meanwhile,the maintenance strategy was optimized to minimize costs per unit time,determining the optimal maintenance timing and preventive maintenance threshold.The comprehensive indicator of degradation trends extracted by this method reaches 0.756,which is 41.2%higher than that of traditional time-domain features;the dynamic threshold strategy reduces the maintenance cost per unit time to 55.56,which is 8.9%better than that of the static threshold optimization.Experimental results demonstrate significant reductions in maintenance costs while enhancing system reliability and safety.This study realizes the organic integration of deep learning and reliability theory in the maintenance of planetary gearboxes,provides an interpretable solution for the predictive maintenance of complex mechanical systems,and promotes the development of condition-based maintenance strategies for planetary gearboxes.展开更多
As oil and gas exploration continues to progress into deeper and unconventional reservoirs,the likelihood of kick risk increases,making kick warning a critical factor in ensuring drilling safety and efficiency.Due to ...As oil and gas exploration continues to progress into deeper and unconventional reservoirs,the likelihood of kick risk increases,making kick warning a critical factor in ensuring drilling safety and efficiency.Due to the scarcity of kick samples,traditional supervised models perform poorly,and significant fluctuations in field data lead to high false alarm rates.This study proposes an unsupervised graph autoencoder(GAE)-based kick warning method,which effectively reduces false alarms by eliminating the influence of field engineer operations and incorporating real-time model updates.The method utilizes the GAE model to process time-series data during drilling,accurately identifying kick risk while overcoming challenges related to small sample sizes and missing features.To further reduce false alarms,the weighted dynamic time warping(WDTW)algorithm is introduced to identify fluctuations in logging data caused by field engineer operations during drilling,with real-time updates applied to prevent normal conditions from being misclassified as kick risk.Experimental results show that the GAE-based kick warning method achieves an accuracy of 92.7%and significantly reduces the false alarm rate.The GAE model continues to operate effectively even under conditions of missing features and issues kick warnings 4 min earlier than field engineers,demonstrating its high sensitivity and robustness.After integrating the WDTW algorithm and real-time updates,the false alarm rate is reduced from 17.3%to 5.6%,further improving the accuracy of kick warnings.The proposed method provides an efficient and reliable approach for kick warning in drilling operations,offering strong practical value and technical support for the intelligent management of future drilling operations.展开更多
Credit risk assessment is a crucial task in bank risk management.By making lending decisions based on credit risk assessment results,banks can reduce the probability of non-performing loans.However,class imbalance in ...Credit risk assessment is a crucial task in bank risk management.By making lending decisions based on credit risk assessment results,banks can reduce the probability of non-performing loans.However,class imbalance in bank credit default datasets limits the predictive performance of traditional machine learning and deep learning models.To address this issue,this study employs the conditional variational autoencoder-Wasserstein generative adversarial network with gradient penalty(CVAE-WGAN-gp)model for oversampling,generating samples similar to the original default customer data to enhance model prediction performance.To evaluate the quality of the data generated by the CVAE-WGAN-gp model,we selected several bank loan datasets for experimentation.The experimental results demonstrate that using the CVAE-WGAN-gp model for oversampling can significantly improve the predictive performance in credit risk assessment problems.展开更多
Task-oriented point cloud sampling aims to select a representative subset from the input,tailored to specific application scenarios and task requirements.However,existing approaches rarely tackle the problem of redund...Task-oriented point cloud sampling aims to select a representative subset from the input,tailored to specific application scenarios and task requirements.However,existing approaches rarely tackle the problem of redundancy caused by local structural similarities in 3D objects,which limits the performance of sampling.To address this issue,this paper introduces a novel task-oriented point cloud masked autoencoder-based sampling network(Point-MASNet),inspired by the masked autoencoder mechanism.Point-MASNet employs a voxel-based random non-overlapping masking strategy,which allows the model to selectively learn and capture distinctive local structural features from the input data.This approach effectively mitigates redundancy and enhances the representativeness of the sampled subset.In addition,we propose a lightweight,symmetrically structured keypoint reconstruction network,designed as an autoencoder.This network is optimized to efficiently extract latent features while enabling refined reconstructions.Extensive experiments demonstrate that Point-MASNet achieves competitive sampling performance across classification,registration,and reconstruction tasks.展开更多
Electronic nose and thermal images are effective ways to diagnose the presence of gases in real-time realtime.Multimodal fusion of these modalities can result in the development of highly accurate diagnostic systems.T...Electronic nose and thermal images are effective ways to diagnose the presence of gases in real-time realtime.Multimodal fusion of these modalities can result in the development of highly accurate diagnostic systems.The low-cost thermal imaging software produces low-resolution thermal images in grayscale format,hence necessitating methods for improving the resolution and colorizing the images.The objective of this paper is to develop and train a super-resolution generative adversarial network for improving the resolution of the thermal images,followed by a sparse autoencoder for colorization of thermal images and amultimodal convolutional neural network for gas detection using electronic nose and thermal images.The dataset used comprises 6400 thermal images and electronic nose measurements for four classes.A multimodal Convolutional Neural Network(CNN)comprising an EfficientNetB2 pre-trainedmodel was developed using both early and late feature fusion.The Super Resolution Generative Adversarial Network(SRGAN)model was developed and trained on low and high-resolution thermal images.Asparse autoencoder was trained on the grayscale and colorized thermal images.The SRGAN was trained on lowand high-resolution thermal images,achieving a Structural Similarity Index(SSIM)of 90.28,a Peak Signal-to-Noise Ratio(PSNR)of 68.74,and a Mean Absolute Error(MAE)of 0.066.The autoencoder model produced an MAE of 0.035,a Mean Squared Error(MSE)of 0.006,and a Root Mean Squared Error(RMSE)of 0.0705.The multimodal CNN,trained on these images and electronic nose measurements using both early and late fusion techniques,achieved accuracies of 97.89% and 98.55%,respectively.Hence,the proposed framework can be of great aid for the integration with low-cost software to generate high quality thermal camera images and highly accurate detection of gases in real-time.展开更多
High-speed railway bridges are essential components of any railway transportation system that should keep adequate levels of serviceability and safety.In this context,drive-by methodologies have emerged as a feasible ...High-speed railway bridges are essential components of any railway transportation system that should keep adequate levels of serviceability and safety.In this context,drive-by methodologies have emerged as a feasible and cost-effective monitor-ing solution for detecting damage on railway bridges while minimizing train operation interruptions.Moreover,integrating advanced sensor technologies and machine learning algorithms has significantly enhanced structural health monitoring(SHM)for bridges.Despite being increasingly used in traditional SHM applications,studies using autoencoders within drive-by methodologies are rare,especially in the railway field.This study presents a novel approach for drive-by damage detection in HSR bridges.The methodology relies on acceleration records collected from multiple bridge crossings by an operational train equipped with onboard sensors.Log-Mel spectrogram features derived from the acceleration records are used together with sparse autoencoders for computing statistical distribution-based damage indexes.Numerical simulations were performed on a 3D vehicle-track-bridge interaction system model implemented in Matlab to evaluate the robustness and effectiveness of the proposed approach,considering several damage scenarios,vehicle speeds,and environmental and operational variations,such as multiple track irregularities and varying measurement noise.The results show that the pro-posed approach can successfully detect damages,as well as characterize their severity,especially for very early-stage dam-ages.This demonstrates the high potential of applying Mel-frequency damage-sensitive features associated with machine learning algorithms in the drive-by condition assessment of high-speed railway bridges.展开更多
Traditional inverse neural network(INN)approaches for inverse design typically require auxiliary feedforward networks,leading to increased computational complexity and architectural dependencies.This study introduces ...Traditional inverse neural network(INN)approaches for inverse design typically require auxiliary feedforward networks,leading to increased computational complexity and architectural dependencies.This study introduces a standalone INN methodology that eliminates the need for feedforward networks while maintaining high reconstruction accuracy.The approach integrates Principal Component Analysis(PCA)and Partial Least Squares(PLS)for optimized feature space learning,enabling the standalone INN to effectively capture bidirectionalmappings between geometric parameters and mechanical properties.Validation using established numerical datasets demonstrates that the standalone INN architecture achieves reconstruction accuracy equal or better than traditional tandem approaches while completely eliminating the workload and training time required for Feedforward Neural Networks(FNN).These findings contribute to AI methodology development by proving that standalone invertible architectures can achieve comparable performance to complex hybrid systems with significantly improved computational efficiency.展开更多
Time series anomaly detection is crucial in finance,healthcare,and industrial monitoring.However,traditional methods often face challenges when handling time series data,such as limited feature extraction capability,p...Time series anomaly detection is crucial in finance,healthcare,and industrial monitoring.However,traditional methods often face challenges when handling time series data,such as limited feature extraction capability,poor temporal dependency handling,and suboptimal real-time performance,sometimes even neglecting the temporal relationships between data.To address these issues and improve anomaly detection performance by better capturing temporal dependencies,we propose an unsupervised time series anomaly detection method,VLT-Anomaly.First,we enhance the Variational Autoencoder(VAE)module by redesigning its network structure to better suit anomaly detection through data reconstruction.We introduce hyperparameters to control the weight of the Kullback-Leibler(KL)divergence term in the Evidence Lower Bound(ELBO),thereby improving the encoder module’s decoupling and expressive power in the latent space,which yields more effective latent representations of the data.Next,we incorporate transformer and Long Short-Term Memory(LSTM)modules to estimate the long-term dependencies of the latent representations,capturing both forward and backward temporal relationships and performing time series forecasting.Finally,we compute the reconstruction error by averaging the predicted results and decoder reconstruction and detect anomalies through grid search for optimal threshold values.Experimental results demonstrate that the proposed method performs superior anomaly detection on multiple public time series datasets,effectively extracting complex time-related features and enabling efficient computation and real-time anomaly detection.It improves detection accuracy and robustness while reducing false positives and false negatives.展开更多
Osteoporosis is a major cause of bone fracture and can be characterised by both mass loss and microstructure deterioration of the bone.The modern way of osteoporosis assessment is through the measurement of bone miner...Osteoporosis is a major cause of bone fracture and can be characterised by both mass loss and microstructure deterioration of the bone.The modern way of osteoporosis assessment is through the measurement of bone mineral density,which is not able to unveil the pathological condition from the mesoscale aspect.To obtain mesoscale information from computed tomography(CT),the super-resolution(SR)approach for volumetric imaging data is required.A deep learning model AESR3D is proposed to recover high-resolution(HR)Micro-CT from low-resolution Micro-CT and implement an unsupervised segmentation for better trabecular observation and measurement.A new regularisation overcomplete autoencoder framework for the SR task is proposed and theoretically analysed.The best performance is achieved on structural similarity measure of trabecular CT SR task compared with the state-of-the-art models in both natural and medical image SR tasks.The HR and SR images show a high correlation(r=0.996,intraclass correlation coefficients=0.917)on trabecular bone morphological indicators.The results also prove the effectiveness of our regularisation framework when training a large capacity model.展开更多
The emergence of Generative Adversarial Network(GAN)techniques has garnered significant attention from the research community for the development of Intrusion Detection Systems(IDS).However,conventional GAN-based IDS ...The emergence of Generative Adversarial Network(GAN)techniques has garnered significant attention from the research community for the development of Intrusion Detection Systems(IDS).However,conventional GAN-based IDS models face several challenges,including training instability,high computational costs,and system failures.To address these limitations,we propose a Hybrid Wasserstein GAN and Autoencoder Model(WGAN-AE)for intrusion detection.The proposed framework leverages the stability of WGAN and the feature extraction capabilities of the Autoencoder Model.The model was trained and evaluated using two recent benchmark datasets,5GNIDD and IDSIoT2024.When trained on the 5GNIDD dataset,the model achieved an average area under the precisionrecall curve is 99.8%using five-fold cross-validation and demonstrated a high detection accuracy of 97.35%when tested on independent test data.Additionally,the model is well-suited for deployment on resource-limited Internetof-Things(IoT)devices due to its ability to detect attacks within microseconds and its small memory footprint of 60.24 kB.Similarly,when trained on the IDSIoT2024 dataset,the model achieved an average PR-AUC of 94.09%and an attack detection accuracy of 97.35%on independent test data,with a memory requirement of 61.84 kB.Extensive simulation results demonstrate that the proposed hybrid model effectively addresses the shortcomings of traditional GAN-based IDS approaches in terms of detection accuracy,computational efficiency,and applicability to real-world IoT environments.展开更多
It is crucial to predict future mechanical behaviors for the prevention of structural disasters.Especially for underground construction,the structural mechanical behaviors are affected by multiple internal and externa...It is crucial to predict future mechanical behaviors for the prevention of structural disasters.Especially for underground construction,the structural mechanical behaviors are affected by multiple internal and external factors due to the complex conditions.Given that the existing models fail to take into account all the factors and accurate prediction of the multiple time series simultaneously is difficult using these models,this study proposed an improved prediction model through the autoencoder fused long-and short-term time-series network driven by the mass number of monitoring data.Then,the proposed model was formalized on multiple time series of strain monitoring data.Also,the discussion analysis with a classical baseline and an ablation experiment was conducted to verify the effectiveness of the prediction model.As the results indicate,the proposed model shows obvious superiority in predicting the future mechanical behaviors of structures.As a case study,the presented model was applied to the Nanjing Dinghuaimen tunnel to predict the stain variation on a different time scale in the future.展开更多
The aerial deployment method enables Unmanned Aerial Vehicles(UAVs)to be directly positioned at the required altitude for their mission.This method typically employs folding technology to improve loading efficiency,wi...The aerial deployment method enables Unmanned Aerial Vehicles(UAVs)to be directly positioned at the required altitude for their mission.This method typically employs folding technology to improve loading efficiency,with applications such as the gravity-only aerial deployment of high-aspect-ratio solar-powered UAVs,and aerial takeoff of fixed-wing drones in Mars research.However,the significant morphological changes during deployment are accompanied by strong nonlinear dynamic aerodynamic forces,which result in multiple degrees of freedom and an unstable character.This hinders the description and analysis of unknown dynamic behaviors,further leading to difficulties in the design of deployment strategies and flight control.To address this issue,this paper proposes an analysis method for dynamic behaviors during aerial deployment based on the Variational Autoencoder(VAE).Focusing on the gravity-only deployment problem of highaspect-ratio foldable-wing UAVs,the method encodes the multi-degree-of-freedom unstable motion signals into a low-dimensional feature space through a data-driven approach.By clustering in the feature space,this paper identifies and studies several dynamic behaviors during aerial deployment.The research presented in this paper offers a new method and perspective for feature extraction and analysis of complex and difficult-to-describe extreme flight dynamics,guiding the research on aerial deployment drones design and control strategies.展开更多
基金This work is supported by the Introducing Program of Dongguan for Leading Talents in Innovation and Entrepreneur(Dongren Han[2018],No.738).
文摘Anomaly detection is an important method for intrusion detection.In recent years,unsupervised methods have been widely researched because they do not require labeling.For example,a nonlinear autoencoder can use reconstruction errors to attain the discrimination threshold.This method is not effective when the model complexity is high or the data contains noise.The method for detecting the density of compressed features in a hidden layer can be used to reduce the influence of noise on the selection of the threshold because the density of abnormal data in hidden layers is smaller than normal data.However,compressed features may lose some of the high-dimensional distribution information of the original data.In this paper,we present an efficient anomaly detection framework for unsupervised anomaly detection,which includes network data capturing,processing,feature extraction,and anomaly detection.We employ a deep autoencoder to obtain compressed features and multi-layer reconstruction errors,and feeds them the same to the Gaussian mixture model to estimate the density.The proposed approach is trained and tested on multiple current intrusion detection datasets and real network scenes,and performance indicators,namely accuracy,recall,and F1-score,are better than other autoencoder models.
文摘The initial noise present in the depth images obtained with RGB-D sensors is a combination of hardware limitations in addition to the environmental factors,due to the limited capabilities of sensors,which also produce poor computer vision results.The common image denoising techniques tend to remove significant image details and also remove noise,provided they are based on space and frequency filtering.The updated framework presented in this paper is a novel denoising model that makes use of Boruta-driven feature selection using a Long Short-Term Memory Autoencoder(LSTMAE).The Boruta algorithm identifies the most useful depth features that are used to maximize the spatial structure integrity and reduce redundancy.An LSTMAE is then used to process these selected features and model depth pixel sequences to generate robust,noise-resistant representations.The system uses the encoder to encode the input data into a latent space that has been compressed before it is decoded to retrieve the clean image.Experiments on a benchmark data set show that the suggested technique attains a PSNR of 45 dB and an SSIM of 0.90,which is 10 dB higher than the performance of conventional convolutional autoencoders and 15 times higher than that of the wavelet-based models.Moreover,the feature selection step will decrease the input dimensionality by 40%,resulting in a 37.5%reduction in training time and a real-time inference rate of 200 FPS.Boruta-LSTMAE framework,therefore,offers a highly efficient and scalable system for depth image denoising,with a high potential to be applied to close-range 3D systems,such as robotic manipulation and gesture-based interfaces.
文摘Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel autoencoder-based imputation framework that integrates a composite loss function to enhance robustness and precision.The proposed loss combines(i)a guided,masked mean squared error focusing on missing entries;(ii)a noise-aware regularization term to improve resilience against data corruption;and(iii)a variance penalty to encourage expressive yet stable reconstructions.We evaluate the proposed model across four missingness mechanisms,such as Missing Completely at Random,Missing at Random,Missing Not at Random,and Missing Not at Random with quantile censorship,under systematically varied feature counts,sample sizes,and missingness ratios ranging from 5%to 60%.Four publicly available real-world datasets(Stroke Prediction,Pima Indians Diabetes,Cardiovascular Disease,and Framingham Heart Study)were used,and the obtained results show that our proposed model consistently outperforms baseline methods,including traditional and deep learning-based techniques.An ablation study reveals the additive value of each component in the loss function.Additionally,we assessed the downstream utility of imputed data through classification tasks,where datasets imputed by the proposed method yielded the highest receiver operating characteristic area under the curve scores across all scenarios.The model demonstrates strong scalability and robustness,improving performance with larger datasets and higher feature counts.These results underscore the capacity of the proposed method to produce not only numerically accurate but also semantically useful imputations,making it a promising solution for robust data recovery in clinical applications.
文摘The rapid growth of biomedical data,particularly multi-omics data including genomes,transcriptomics,proteomics,metabolomics,and epigenomics,medical research and clinical decision-making confront both new opportunities and obstacles.The huge and diversified nature of these datasets cannot always be managed using traditional data analysis methods.As a consequence,deep learning has emerged as a strong tool for analysing numerous omics data due to its ability to handle complex and non-linear relationships.This paper explores the fundamental concepts of deep learning and how they are used in multi-omics medical data mining.We demonstrate how autoencoders,variational autoencoders,multimodal models,attention mechanisms,transformers,and graph neural networks enable pattern analysis and recognition across all omics data.Deep learning has been found to be effective in illness classification,biomarker identification,gene network learning,and therapeutic efficacy prediction.We also consider critical problems like as data quality,model explainability,whether findings can be repeated,and computational power requirements.We now consider future elements of combining omics with clinical and imaging data,explainable AI,federated learning,and real-time diagnostics.Overall,this study emphasises the need of collaborating across disciplines to advance deep learning-based multi-omics research for precision medicine and comprehending complicated disorders.
文摘Inverse design of advanced materials represents a pivotal challenge in materials science.Leveraging the latent space of Variational Autoencoders(VAEs)for material optimization has emerged as a significant advancement in the field of material inverse design.However,VAEs are inherently prone to generating blurred images,posing challenges for precise inverse design and microstructure manufacturing.While increasing the dimensionality of the VAE latent space can mitigate reconstruction blurriness to some extent,it simultaneously imposes a substantial burden on target optimization due to an excessively high search space.To address these limitations,this study adopts a Variational Autoencoder guided Conditional Diffusion Generative Model(VAE-CDGM)framework integrated with Bayesian optimization to achieve the inverse design of composite materials with targeted mechanical properties.The VAE-CDGM model synergizes the strengths of VAEs and Denoising Diffusion Probabilistic Models(DDPM),enabling the generation of high-quality,sharp images while preserving a manipulable latent space.To accommodate varying dimensional requirements of the latent space,two optimization strategies are proposed.When the latent space dimensionality is excessively high,SHapley Additive exPlanations(SHAP)sensitivity analysis is employed to identify critical latent features for optimization within a reduced subspace.Conversely,direct optimization is performed in the low-dimensional latent space of VAE-CDGM when dimensionality is modest.The results demonstrate that both strategies accurately achieve the targeted design of composite materials while circumventing the blurred reconstruction flaws of VAEs,which offers a novel pathway for the precise design of advanced materials.
文摘Automated Program Repair(APR)techniques have shown significant potential in mitigating the cost and complexity associated with debugging by automatically generating corrective patches for software defects.Despite considerable progress in APR methodologies,existing approaches frequently lack contextual awareness of runtime behaviors and structural intricacies inherent in buggy source code.In this paper,we propose a novel APR approach that integrates attention mechanisms within an autoencoder-based framework,explicitly utilizing structural code affinity and execution context correlation derived from stack trace analysis.Our approach begins with an innovative preprocessing pipeline,where code segments and stack traces are transformed into tokenized representations.Subsequently,the BM25 ranking algorithm is employed to quantitatively measure structural code affinity and execution context correlation,identifying syntactically and semantically analogous buggy code snippets and relevant runtime error contexts from extensive repositories.These extracted features are then encoded via an attention-enhanced autoencoder model,specifically designed to capture significant patterns and correlations essential for effective patch generation.To assess the efficacy and generalizability of our proposed method,we conducted rigorous experimental comparisons against DeepFix,a state-of-the-art APR system,using a substantial dataset comprising 53,478 studentdeveloped C programs.Experimental outcomes indicate that our model achieves a notable bug repair success rate of approximately 62.36%,representing a statistically significant performance improvement of over 6%compared to the baseline.Furthermore,a thorough K-fold cross-validation reinforced the consistency,robustness,and reliability of our method across diverse subsets of the dataset.Our findings present the critical advantage of integrating attentionbased learning with code structural and execution context features in APR tasks,leading to improved accuracy and practical applicability.Future work aims to extend the model’s applicability across different programming languages,systematically optimize hyperparameters,and explore alternative feature representation methods to further enhance debugging efficiency and effectiveness.
文摘In the field of intelligent air combat,real-time and accurate recognition of within-visual-range(WVR)maneuver actions serves as the foundational cornerstone for constructing autonomous decision-making systems.However,existing methods face two major challenges:traditional feature engineering suffers from insufficient effective dimensionality in the feature space due to kinematic coupling,making it difficult to distinguish essential differences between maneuvers,while end-to-end deep learning models lack controllability in implicit feature learning and fail to model high-order long-range temporal dependencies.This paper proposes a trajectory feature pre-extraction method based on a Long-range Masked Autoencoder(LMAE),incorporating three key innovations:(1)Random Fragment High-ratio Masking(RFH-Mask),which enforces the model to learn long-range temporal correlations by masking 80%of trajectory data while retaining continuous fragments;(2)Kalman Filter-Guided Objective Function(KFG-OF),integrating trajectory continuity constraints to align the feature space with kinematic principles;and(3)Two-stage Decoupled Architecture,enabling efficient and controllable feature learning through unsupervised pre-training and frozen-feature transfer.Experimental results demonstrate that LMAE significantly improves the average recognition accuracy for 20-class maneuvers compared to traditional end-to-end models,while significantly accelerating convergence speed.The contributions of this work lie in:introducing high-masking-rate autoencoders into low-informationdensity trajectory analysis,proposing a feature engineering framework with enhanced controllability and efficiency,and providing a novel technical pathway for intelligent air combat decision-making systems.
基金funded by scientific research projects under Grant JY2024B011.
文摘With the increasing complexity of industrial automation,planetary gearboxes play a vital role in largescale equipment transmission systems,directly impacting operational efficiency and safety.Traditional maintenance strategies often struggle to accurately predict the degradation process of equipment,leading to excessive maintenance costs or potential failure risks.However,existing prediction methods based on statistical models are difficult to adapt to nonlinear degradation processes.To address these challenges,this study proposes a novel condition-based maintenance framework for planetary gearboxes.A comprehensive full-lifecycle degradation experiment was conducted to collect raw vibration signals,which were then processed using a temporal convolutional network autoencoder with multi-scale perception capability to extract deep temporal degradation features,enabling the collaborative extraction of longperiod meshing frequencies and short-term impact features from the vibration signals.Kernel principal component analysis was employed to fuse and normalize these features,enhancing the characterization of degradation progression.A nonlinear Wiener process was used to model the degradation trajectory,with a threshold decay function introduced to dynamically adjust maintenance strategies,and model parameters optimized through maximum likelihood estimation.Meanwhile,the maintenance strategy was optimized to minimize costs per unit time,determining the optimal maintenance timing and preventive maintenance threshold.The comprehensive indicator of degradation trends extracted by this method reaches 0.756,which is 41.2%higher than that of traditional time-domain features;the dynamic threshold strategy reduces the maintenance cost per unit time to 55.56,which is 8.9%better than that of the static threshold optimization.Experimental results demonstrate significant reductions in maintenance costs while enhancing system reliability and safety.This study realizes the organic integration of deep learning and reliability theory in the maintenance of planetary gearboxes,provides an interpretable solution for the predictive maintenance of complex mechanical systems,and promotes the development of condition-based maintenance strategies for planetary gearboxes.
基金Youth Foundation of National Natural Science Foundation of China (No. 52204020)Distinguished Young Foundation of National Natural Science Foundation of China (No. 52125401).
文摘As oil and gas exploration continues to progress into deeper and unconventional reservoirs,the likelihood of kick risk increases,making kick warning a critical factor in ensuring drilling safety and efficiency.Due to the scarcity of kick samples,traditional supervised models perform poorly,and significant fluctuations in field data lead to high false alarm rates.This study proposes an unsupervised graph autoencoder(GAE)-based kick warning method,which effectively reduces false alarms by eliminating the influence of field engineer operations and incorporating real-time model updates.The method utilizes the GAE model to process time-series data during drilling,accurately identifying kick risk while overcoming challenges related to small sample sizes and missing features.To further reduce false alarms,the weighted dynamic time warping(WDTW)algorithm is introduced to identify fluctuations in logging data caused by field engineer operations during drilling,with real-time updates applied to prevent normal conditions from being misclassified as kick risk.Experimental results show that the GAE-based kick warning method achieves an accuracy of 92.7%and significantly reduces the false alarm rate.The GAE model continues to operate effectively even under conditions of missing features and issues kick warnings 4 min earlier than field engineers,demonstrating its high sensitivity and robustness.After integrating the WDTW algorithm and real-time updates,the false alarm rate is reduced from 17.3%to 5.6%,further improving the accuracy of kick warnings.The proposed method provides an efficient and reliable approach for kick warning in drilling operations,offering strong practical value and technical support for the intelligent management of future drilling operations.
基金supported by National Key R&D Program of China(2022YFA1008000)the National Natural Science Foundation of China(12571297,12101585)+1 种基金the CAS Talent Introduction Program(Category B)the Young Elite Scientist Sponsorship Program by CAST(YESS20220125).
文摘Credit risk assessment is a crucial task in bank risk management.By making lending decisions based on credit risk assessment results,banks can reduce the probability of non-performing loans.However,class imbalance in bank credit default datasets limits the predictive performance of traditional machine learning and deep learning models.To address this issue,this study employs the conditional variational autoencoder-Wasserstein generative adversarial network with gradient penalty(CVAE-WGAN-gp)model for oversampling,generating samples similar to the original default customer data to enhance model prediction performance.To evaluate the quality of the data generated by the CVAE-WGAN-gp model,we selected several bank loan datasets for experimentation.The experimental results demonstrate that using the CVAE-WGAN-gp model for oversampling can significantly improve the predictive performance in credit risk assessment problems.
基金supported by the National Key Research and Development Program of China(2022YFB3103500)the National Natural Science Foundation of China(62473033,62571027)+1 种基金in part by the Beijing Natural Science Foundation(L231012)the State Scholarship Fund from the China Scholarship Council.
文摘Task-oriented point cloud sampling aims to select a representative subset from the input,tailored to specific application scenarios and task requirements.However,existing approaches rarely tackle the problem of redundancy caused by local structural similarities in 3D objects,which limits the performance of sampling.To address this issue,this paper introduces a novel task-oriented point cloud masked autoencoder-based sampling network(Point-MASNet),inspired by the masked autoencoder mechanism.Point-MASNet employs a voxel-based random non-overlapping masking strategy,which allows the model to selectively learn and capture distinctive local structural features from the input data.This approach effectively mitigates redundancy and enhances the representativeness of the sampled subset.In addition,we propose a lightweight,symmetrically structured keypoint reconstruction network,designed as an autoencoder.This network is optimized to efficiently extract latent features while enabling refined reconstructions.Extensive experiments demonstrate that Point-MASNet achieves competitive sampling performance across classification,registration,and reconstruction tasks.
基金funded by the Centre for Advanced Modelling and Geospatial Information Systems(CAMGIS),Faculty of Engineering and IT,University of Technology Sydneysupported by the Researchers Supporting Project,King Saud University,Riyadh,Saudi Arabia,under Project RSP2025 R14.
文摘Electronic nose and thermal images are effective ways to diagnose the presence of gases in real-time realtime.Multimodal fusion of these modalities can result in the development of highly accurate diagnostic systems.The low-cost thermal imaging software produces low-resolution thermal images in grayscale format,hence necessitating methods for improving the resolution and colorizing the images.The objective of this paper is to develop and train a super-resolution generative adversarial network for improving the resolution of the thermal images,followed by a sparse autoencoder for colorization of thermal images and amultimodal convolutional neural network for gas detection using electronic nose and thermal images.The dataset used comprises 6400 thermal images and electronic nose measurements for four classes.A multimodal Convolutional Neural Network(CNN)comprising an EfficientNetB2 pre-trainedmodel was developed using both early and late feature fusion.The Super Resolution Generative Adversarial Network(SRGAN)model was developed and trained on low and high-resolution thermal images.Asparse autoencoder was trained on the grayscale and colorized thermal images.The SRGAN was trained on lowand high-resolution thermal images,achieving a Structural Similarity Index(SSIM)of 90.28,a Peak Signal-to-Noise Ratio(PSNR)of 68.74,and a Mean Absolute Error(MAE)of 0.066.The autoencoder model produced an MAE of 0.035,a Mean Squared Error(MSE)of 0.006,and a Root Mean Squared Error(RMSE)of 0.0705.The multimodal CNN,trained on these images and electronic nose measurements using both early and late fusion techniques,achieved accuracies of 97.89% and 98.55%,respectively.Hence,the proposed framework can be of great aid for the integration with low-cost software to generate high quality thermal camera images and highly accurate detection of gases in real-time.
基金support of CNPq(Brazilian Ministry of Science and Technology Agency),of CAPES(Higher Education Improvement Agency),of FAPESP(São Paulo Research Foundation)under grant#2022/13045-1,of VALE Catedra Under Rail and of Base Funding-UIDB/04708/2020Programmatic Funding-UIDP/04708/2020 of the CONSTRUCT-“Instituto de I&D em Estruturas e Construções”.
文摘High-speed railway bridges are essential components of any railway transportation system that should keep adequate levels of serviceability and safety.In this context,drive-by methodologies have emerged as a feasible and cost-effective monitor-ing solution for detecting damage on railway bridges while minimizing train operation interruptions.Moreover,integrating advanced sensor technologies and machine learning algorithms has significantly enhanced structural health monitoring(SHM)for bridges.Despite being increasingly used in traditional SHM applications,studies using autoencoders within drive-by methodologies are rare,especially in the railway field.This study presents a novel approach for drive-by damage detection in HSR bridges.The methodology relies on acceleration records collected from multiple bridge crossings by an operational train equipped with onboard sensors.Log-Mel spectrogram features derived from the acceleration records are used together with sparse autoencoders for computing statistical distribution-based damage indexes.Numerical simulations were performed on a 3D vehicle-track-bridge interaction system model implemented in Matlab to evaluate the robustness and effectiveness of the proposed approach,considering several damage scenarios,vehicle speeds,and environmental and operational variations,such as multiple track irregularities and varying measurement noise.The results show that the pro-posed approach can successfully detect damages,as well as characterize their severity,especially for very early-stage dam-ages.This demonstrates the high potential of applying Mel-frequency damage-sensitive features associated with machine learning algorithms in the drive-by condition assessment of high-speed railway bridges.
基金funding by the Deutsche Forschungsgemeinschaft under Germany’s Excellence Strategy within the Cluster of Excellence PhoenixD(EXC 2122,Project ID 390833453).
文摘Traditional inverse neural network(INN)approaches for inverse design typically require auxiliary feedforward networks,leading to increased computational complexity and architectural dependencies.This study introduces a standalone INN methodology that eliminates the need for feedforward networks while maintaining high reconstruction accuracy.The approach integrates Principal Component Analysis(PCA)and Partial Least Squares(PLS)for optimized feature space learning,enabling the standalone INN to effectively capture bidirectionalmappings between geometric parameters and mechanical properties.Validation using established numerical datasets demonstrates that the standalone INN architecture achieves reconstruction accuracy equal or better than traditional tandem approaches while completely eliminating the workload and training time required for Feedforward Neural Networks(FNN).These findings contribute to AI methodology development by proving that standalone invertible architectures can achieve comparable performance to complex hybrid systems with significantly improved computational efficiency.
基金support from the Fundamental Research Funds for Central Public Welfare Research Institutes(SK202324)the Central Guidance on Local Science and Technology Development Fund of Hebei Province(236Z0104G)+1 种基金the National Natural Science Foundation of China(62476078)the Geological Survey Project of China Geological Survey(G202304-2).
文摘Time series anomaly detection is crucial in finance,healthcare,and industrial monitoring.However,traditional methods often face challenges when handling time series data,such as limited feature extraction capability,poor temporal dependency handling,and suboptimal real-time performance,sometimes even neglecting the temporal relationships between data.To address these issues and improve anomaly detection performance by better capturing temporal dependencies,we propose an unsupervised time series anomaly detection method,VLT-Anomaly.First,we enhance the Variational Autoencoder(VAE)module by redesigning its network structure to better suit anomaly detection through data reconstruction.We introduce hyperparameters to control the weight of the Kullback-Leibler(KL)divergence term in the Evidence Lower Bound(ELBO),thereby improving the encoder module’s decoupling and expressive power in the latent space,which yields more effective latent representations of the data.Next,we incorporate transformer and Long Short-Term Memory(LSTM)modules to estimate the long-term dependencies of the latent representations,capturing both forward and backward temporal relationships and performing time series forecasting.Finally,we compute the reconstruction error by averaging the predicted results and decoder reconstruction and detect anomalies through grid search for optimal threshold values.Experimental results demonstrate that the proposed method performs superior anomaly detection on multiple public time series datasets,effectively extracting complex time-related features and enabling efficient computation and real-time anomaly detection.It improves detection accuracy and robustness while reducing false positives and false negatives.
基金Beijing Natural Science Foundation-Haidian original Innovation Joint Foundation,Grant/Award Number:L192016Joint Funds of the National Natural Science Foundation of China,Grant/Award Number:U21A20489+3 种基金National Natural Science Foundation of China,Grant/Award Number:62003330Shenzhen Fundamental Research Funds,Grant/Award Numbers:JCYJ20220818101608019,JCYJ20190807170407391,JCYJ20180507182415428Natural Science Foundation of Guangdong Province,Grant/Award Number:2019A1515011699Guangdong-Hong Kong-Macao Joint Laboratory of Human-Machine Intelligence-Synergy Systems,Shenzhen Institute of Advanced Technology。
文摘Osteoporosis is a major cause of bone fracture and can be characterised by both mass loss and microstructure deterioration of the bone.The modern way of osteoporosis assessment is through the measurement of bone mineral density,which is not able to unveil the pathological condition from the mesoscale aspect.To obtain mesoscale information from computed tomography(CT),the super-resolution(SR)approach for volumetric imaging data is required.A deep learning model AESR3D is proposed to recover high-resolution(HR)Micro-CT from low-resolution Micro-CT and implement an unsupervised segmentation for better trabecular observation and measurement.A new regularisation overcomplete autoencoder framework for the SR task is proposed and theoretically analysed.The best performance is achieved on structural similarity measure of trabecular CT SR task compared with the state-of-the-art models in both natural and medical image SR tasks.The HR and SR images show a high correlation(r=0.996,intraclass correlation coefficients=0.917)on trabecular bone morphological indicators.The results also prove the effectiveness of our regularisation framework when training a large capacity model.
基金the Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Group Project under grant number(RGP.2/245/46)funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R760)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.The research team thanks the Deanship of Graduate Studies and Scientific Research at Najran University for supporting the research project through the Nama’a program,with the project code NU/GP/SERC/13/352-1。
文摘The emergence of Generative Adversarial Network(GAN)techniques has garnered significant attention from the research community for the development of Intrusion Detection Systems(IDS).However,conventional GAN-based IDS models face several challenges,including training instability,high computational costs,and system failures.To address these limitations,we propose a Hybrid Wasserstein GAN and Autoencoder Model(WGAN-AE)for intrusion detection.The proposed framework leverages the stability of WGAN and the feature extraction capabilities of the Autoencoder Model.The model was trained and evaluated using two recent benchmark datasets,5GNIDD and IDSIoT2024.When trained on the 5GNIDD dataset,the model achieved an average area under the precisionrecall curve is 99.8%using five-fold cross-validation and demonstrated a high detection accuracy of 97.35%when tested on independent test data.Additionally,the model is well-suited for deployment on resource-limited Internetof-Things(IoT)devices due to its ability to detect attacks within microseconds and its small memory footprint of 60.24 kB.Similarly,when trained on the IDSIoT2024 dataset,the model achieved an average PR-AUC of 94.09%and an attack detection accuracy of 97.35%on independent test data,with a memory requirement of 61.84 kB.Extensive simulation results demonstrate that the proposed hybrid model effectively addresses the shortcomings of traditional GAN-based IDS approaches in terms of detection accuracy,computational efficiency,and applicability to real-world IoT environments.
基金National Key Research and Development Program of China,Grant/Award Number:2018YFB2101003National Natural Science Foundation of China,Grant/Award Numbers:51991395,U1806226,51778033,51822802,71901011,U1811463,51991391Science and Technology Major Project of Beijing,Grant/Award Number:Z191100002519012。
文摘It is crucial to predict future mechanical behaviors for the prevention of structural disasters.Especially for underground construction,the structural mechanical behaviors are affected by multiple internal and external factors due to the complex conditions.Given that the existing models fail to take into account all the factors and accurate prediction of the multiple time series simultaneously is difficult using these models,this study proposed an improved prediction model through the autoencoder fused long-and short-term time-series network driven by the mass number of monitoring data.Then,the proposed model was formalized on multiple time series of strain monitoring data.Also,the discussion analysis with a classical baseline and an ablation experiment was conducted to verify the effectiveness of the prediction model.As the results indicate,the proposed model shows obvious superiority in predicting the future mechanical behaviors of structures.As a case study,the presented model was applied to the Nanjing Dinghuaimen tunnel to predict the stain variation on a different time scale in the future.
基金co-supported by the Natural Science Basic Research Program of Shaanxi,China(No.2023-JC-QN-0043)the ND Basic Research Funds,China(No.G2022WD).
文摘The aerial deployment method enables Unmanned Aerial Vehicles(UAVs)to be directly positioned at the required altitude for their mission.This method typically employs folding technology to improve loading efficiency,with applications such as the gravity-only aerial deployment of high-aspect-ratio solar-powered UAVs,and aerial takeoff of fixed-wing drones in Mars research.However,the significant morphological changes during deployment are accompanied by strong nonlinear dynamic aerodynamic forces,which result in multiple degrees of freedom and an unstable character.This hinders the description and analysis of unknown dynamic behaviors,further leading to difficulties in the design of deployment strategies and flight control.To address this issue,this paper proposes an analysis method for dynamic behaviors during aerial deployment based on the Variational Autoencoder(VAE).Focusing on the gravity-only deployment problem of highaspect-ratio foldable-wing UAVs,the method encodes the multi-degree-of-freedom unstable motion signals into a low-dimensional feature space through a data-driven approach.By clustering in the feature space,this paper identifies and studies several dynamic behaviors during aerial deployment.The research presented in this paper offers a new method and perspective for feature extraction and analysis of complex and difficult-to-describe extreme flight dynamics,guiding the research on aerial deployment drones design and control strategies.