The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and hist...The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.展开更多
High-quality data is essential for the success of data-driven learning tasks.The characteristics,precision,and completeness of the datasets critically determine the reliability,interpretability,and effectiveness of su...High-quality data is essential for the success of data-driven learning tasks.The characteristics,precision,and completeness of the datasets critically determine the reliability,interpretability,and effectiveness of subsequent analyzes and applications,such as fault detection,predictive maintenance,and process optimization.However,for many industrial processes,obtaining sufficient high-quality data remains a significant challenge due to high costs,safety concerns,and practical constraints.To overcome these challenges,data augmentation has emerged as a rapidly growing research area,attracting considerable attention across both academia and industry.By expanding datasets,data augmentation techniques improve greater generalization and more robust performance in actual applications.This paper provides a comprehensive,multi-perspective review of data augmentation methods for industrial processes.For clarity and organization,existing studies are systematically grouped into four categories:small sample with low dimension,small sample with high dimension,large sample with low dimension,and large sample with high dimension.Within this framework,the review examines current research from both methodological and application-oriented perspectives,highlighting main methods,advantages,and limitations.By synthesizing these findings,this review offers a structured overview for scholars and practitioners,serving as a valuable reference for newcomers and experienced researchers seeking to explore and advance data augmentation techniques in industrial processes.展开更多
An improved cycle-consistent generative adversarial network(CycleGAN) method for defect data augmentation based on feature fusion and self attention residual module is proposed to address the insufficiency of defect s...An improved cycle-consistent generative adversarial network(CycleGAN) method for defect data augmentation based on feature fusion and self attention residual module is proposed to address the insufficiency of defect sample data for light guide plate(LGP) in production,as well as the problem of minor defects.Two optimizations are made to the generator of CycleGAN:fusion of low resolution features obtained from partial up-sampling and down-sampling with high-resolution features,combination of self attention mechanism with residual network structure to replace the original residual module.Qualitative and quantitative experiments were conducted to compare different data augmentation methods,and the results show that the defect images of the LGP generated by the improved network were more realistic,and the accuracy of the you only look once version 5(YOLOv5) detection network for the LGP was improved by 5.6%,proving the effectiveness and accuracy of the proposed method.展开更多
Seismic data denoising is a critical process usually applied at various stages of the seismic processing workflow,as our ability to mitigate noise in seismic data affects the quality of our subsequent analyses.However...Seismic data denoising is a critical process usually applied at various stages of the seismic processing workflow,as our ability to mitigate noise in seismic data affects the quality of our subsequent analyses.However,finding an optimal balance between preserving seismic signals and effectively reducing seismic noise presents a substantial challenge.In this study,we introduce a multi-stage deep learning model,trained in a self-supervised manner,designed specifically to suppress seismic noise while minimizing signal leakage.This model operates as a patch-based approach,extracting overlapping patches from the noisy data and converting them into 1D vectors for input.It consists of two identical sub-networks,each configured differently.Inspired by the transformer architecture,each sub-network features an embedded block that comprises two fully connected layers,which are utilized for feature extraction from the input patches.After reshaping,a multi-head attention module enhances the model’s focus on significant features by assigning higher attention weights to them.The key difference between the two sub-networks lies in the number of neurons within their fully connected layers.The first sub-network serves as a strong denoiser with a small number of neurons,effectively attenuating seismic noise;in contrast,the second sub-network functions as a signal-add-back model,using a larger number of neurons to retrieve some of the signal that was not preserved in the output of the first sub-network.The proposed model produces two outputs,each corresponding to one of the sub-networks,and both sub-networks are optimized simultaneously using the noisy data as the label for both outputs.Evaluations conducted on both synthetic and field data demonstrate the model’s effectiveness in suppressing seismic noise with minimal signal leakage,outperforming some benchmark methods.展开更多
In order to address the widespread data shortage problem in battery research,this paper proposes a generative adversarial network model that combines it with deep convolutional networks,the Wasserstein distance,and th...In order to address the widespread data shortage problem in battery research,this paper proposes a generative adversarial network model that combines it with deep convolutional networks,the Wasserstein distance,and the gradient penalty to achieve data augmentation.To lower the threshold for implementing the proposed method,transfer learning is further introduced.The W-DC-GAN-GP-TL framework is thereby formed.This framework is evaluated on 3 different publicly available datasets to judge the quality of generated data.Through visual comparisons and the examination of two visualization methods(probability density function(PDF)and principal component analysis(PCA)),it is demonstrated that the generated data is hard to distinguish from the real data.The application of generated data for training a battery state model using transfer learning is further evaluated.Specifically,Bi-GRU-based and Transformer-based methods are implemented on 2 separate datasets for estimating state of health(SOH)and state of charge(SOC),respectively.The results indicate that the proposed framework demonstrates satisfactory performance in different scenarios:for the data replacement scenario,where real data are removed and replaced with generated data,the state estimator accuracy decreases only slightly;for the data enhancement scenario,the estimator accuracy is further improved.The estimation accuracy of SOH and SOC is as low as 0.69%and 0.58%root mean square error(RMSE)after applying the proposed framework.This framework provides a reliable method for enriching battery measurement data.It is a generalized framework capable of generating a variety of time series data.展开更多
Bird species classification is not only a challenging topic in artificial intelligence but also a domain closely related to environmental protection and ecological research.Additionally,performing edge computing on lo...Bird species classification is not only a challenging topic in artificial intelligence but also a domain closely related to environmental protection and ecological research.Additionally,performing edge computing on low-level devices using small neural networks can be an important research direction.In this paper,we use the EfficientNetV2B0 model for bird species classification,applying transfer learning on a dataset of 525 bird species.We also employ the BiRefNet model to remove backgrounds from images in the training set.The generated background-removed images are mixed with the original training set as a form of data augmentation.We aim for these background-removed images to help the model focus on key features,and by combining data augmentation with transfer learning,we trained a highly accurate and efficient bird species classification model.The training process is divided into a transfer learning stage and a fine-tuning stage.In the transfer learning stage,only the newly added custom layers are trained;while in the fine-tuning stage,all pre-trained layers except for the batch normalization layers are fine-tuned.According to the experimental results,the proposed model not only has an advantage in size compared to other models but also outperforms them in various metrics.The training results show that the proposed model achieved an accuracy of 99.54%and a precision of 99.62%,demonstrating that it achieves both lightweight design and high accuracy.To confirm the credibility of the results,we use heatmaps to interpret the model.The heatmaps show that our model can clearly highlight the image feature area.In addition,we also perform the 10-fold cross-validation on the model to verify its credibility.Finally,this paper proposes a model with low training cost and high accuracy,making it suitable for deployment on edge computing devices to provide lighter and more convenient services.展开更多
Rainfall-induced shallow landslides pose one of significant geological hazards,necessitating precise monitoring and prediction for effective disaster mitigation.Most studies on landslide prediction have focused on opt...Rainfall-induced shallow landslides pose one of significant geological hazards,necessitating precise monitoring and prediction for effective disaster mitigation.Most studies on landslide prediction have focused on optimizing machine learning(ML)algorithms,very limited attention has been paid to enhancing data quality for improved predictive performance.This study employs strategic data augmentation(DA)techniques to enhance the accuracy of shallow landslide prediction.Using five DA methods including singular spectrum analysis(SSA),moving averages(MA),wavelet denoising(WD),variational mode decomposition(VMD),and linear interpolation(LI),we utilize strategies such as smoothing,denoising,trend decomposition,and synthetic data generation to improve the training dataset.Four machine learning algorithms,i.e.artificial neural network(ANN),recurrent neural network(RNN),one-dimensional convolutional neural network(CNN1D),and long short-term memory(LSTM),are used to forecast landslide displacement.The case study of a landslide in southwest China shows the effectiveness of our approach in predicting landslide displacements,despite the inherent limitations of the monitoring dataset.VMD proves the most effective for smoothing and denoising,improving R^(2),RMSE,and MAPE by 172.16%,71.82%,and 98.9%,respectively.SSA addresses missing data,while LI is effective with limited data samples,improving metrics by 21.6%,52.59%,and 47.87%,respectively.This study demonstrates the potential of DA techniques to mitigate the impact of data defects on landslide prediction accuracy,with implications for similar cases.展开更多
Large-scale point cloud datasets form the basis for training various deep learning networks and achieving high-quality network processing tasks.Due to the diversity and robustness constraints of the data,data augmenta...Large-scale point cloud datasets form the basis for training various deep learning networks and achieving high-quality network processing tasks.Due to the diversity and robustness constraints of the data,data augmentation(DA)methods are utilised to expand dataset diversity and scale.However,due to the complex and distinct characteristics of LiDAR point cloud data from different platforms(such as missile-borne and vehicular LiDAR data),directly applying traditional 2D visual domain DA methods to 3D data can lead to networks trained using this approach not robustly achieving the corresponding tasks.To address this issue,the present study explores DA for missile-borne LiDAR point cloud using a Monte Carlo(MC)simulation method that closely resembles practical application.Firstly,the model of multi-sensor imaging system is established,taking into account the joint errors arising from the platform itself and the relative motion during the imaging process.A distortion simulation method based on MC simulation for augmenting missile-borne LiDAR point cloud data is proposed,underpinned by an analysis of combined errors between different modal sensors,achieving high-quality augmentation of point cloud data.The effectiveness of the proposed method in addressing imaging system errors and distortion simulation is validated using the imaging scene dataset constructed in this paper.Comparative experiments between the proposed point cloud DA algorithm and the current state-of-the-art algorithms in point cloud detection and single object tracking tasks demonstrate that the proposed method can improve the network performance obtained from unaugmented datasets by over 17.3%and 17.9%,surpassing SOTA performance of current point cloud DA algorithms.展开更多
Data augmentation plays an important role in boosting the performance of 3D models,while very few studies handle the 3D point cloud data with this technique.Global augmentation and cut-paste are commonly used augmenta...Data augmentation plays an important role in boosting the performance of 3D models,while very few studies handle the 3D point cloud data with this technique.Global augmentation and cut-paste are commonly used augmentation techniques for point clouds,where global augmentation is applied to the entire point cloud of the scene,and cut-paste samples objects from other frames into the current frame.Both types of data augmentation can improve performance,but the cut-paste technique cannot effectively deal with the occlusion relationship between the foreground object and the background scene and the rationality of object sampling,which may be counterproductive and may hurt the overall performance.In addition,LiDAR is susceptible to signal loss,external occlusion,extreme weather and other factors,which can easily cause object shape changes,while global augmentation and cut-paste cannot effectively enhance the robustness of the model.To this end,we propose Syn-Aug,a synchronous data augmentation framework for LiDAR-based 3D object detection.Specifically,we first propose a novel rendering-based object augmentation technique(Ren-Aug)to enrich training data while enhancing scene realism.Second,we propose a local augmentation technique(Local-Aug)to generate local noise by rotating and scaling objects in the scene while avoiding collisions,which can improve generalisation performance.Finally,we make full use of the structural information of 3D labels to make the model more robust by randomly changing the geometry of objects in the training frames.We verify the proposed framework with four different types of 3D object detectors.Experimental results show that our proposed Syn-Aug significantly improves the performance of various 3D object detectors in the KITTI and nuScenes datasets,proving the effectiveness and generality of Syn-Aug.On KITTI,four different types of baseline models using Syn-Aug improved mAP by 0.89%,1.35%,1.61%and 1.14%respectively.On nuScenes,four different types of baseline models using Syn-Aug improved mAP by 14.93%,10.42%,8.47%and 6.81%respectively.The code is available at https://github.com/liuhuaijjin/Syn-Aug.展开更多
Data augmentation plays an important role in training deep neural model by expanding the size and diversity of the dataset.Initially,data augmentation mainly involved some simple transformations of images.Later,in ord...Data augmentation plays an important role in training deep neural model by expanding the size and diversity of the dataset.Initially,data augmentation mainly involved some simple transformations of images.Later,in order to increase the diversity and complexity of data,more advanced methods appeared and evolved to sophisticated generative models.However,these methods required a mass of computation of training or searching.In this paper,a novel training-free method that utilises the Pre-Trained Segment Anything Model(SAM)model as a data augmentation tool(PTSAM-DA)is proposed to generate the augmented annotations for images.Without the need for training,it obtains prompt boxes from the original annotations and then feeds the boxes to the pre-trained SAM to generate diverse and improved annotations.In this way,annotations are augmented more ingenious than simple manipulations without incurring huge computation for training a data augmentation model.Multiple comparative experiments on three datasets are conducted,including an in-house dataset,ADE20K and COCO2017.On this in-house dataset,namely Agricultural Plot Segmentation Dataset,maximum improvements of 3.77%and 8.92%are gained in two mainstream metrics,mIoU and mAcc,respectively.Consequently,large vision models like SAM are proven to be promising not only in image segmentation but also in data augmentation.展开更多
The widespread adoption of tunnel boring machines(TBMs)has led to an increased focus on disc cutter wear,including both normal and abnormal types,for efficient and safe TBM excavation.However,abnormal wear has yet to ...The widespread adoption of tunnel boring machines(TBMs)has led to an increased focus on disc cutter wear,including both normal and abnormal types,for efficient and safe TBM excavation.However,abnormal wear has yet to be thoroughly investigated,primarily due to the complexity of considering mixed ground conditions and the imbalance in the number of instances between the two types of wear.This study developed a prediction model for abnormal TBM disc cutter wear,considering mixed ground conditions,by employing interpretable machine learning with data augmentation.An equivalent elastic modulus was used to consider the characteristics of mixed ground conditions,and wear data was obtained from 65 cutterhead intervention(CHI)reports covering both mixed ground and hard rock sections.With a balanced training dataset obtained by data augmentation,an extreme gradient boosting(XGB)model delivered acceptable results with an accuracy of 0.94,an F1-score of 0.808,and a recall of 0.8.In addition,the accuracy for each individual disc cutter exhibited low variability.When employing data augmentation,a significant improvement in recall was observed compared to when it was not used,although the difference in accuracy and F1-score was marginal.The subsequent model interpretation revealed the chamber pressure,cutter installation radius,and torque as significant contributors.Specifically,a threshold in chamber pressure was observed,which could induce abnormal wear.The study also explored how elevated values of these influential contributors correlate with abnormal wear.The proposed model offers a valuable tool for planning the replacement of abnormally worn disc cutters,enhancing the safety and efficiency of TBM operations.展开更多
With the emergence of new attack techniques,traffic classifiers usually fail to maintain the expected performance in real-world network environments.In order to have sufficient generalizability to deal with unknown ma...With the emergence of new attack techniques,traffic classifiers usually fail to maintain the expected performance in real-world network environments.In order to have sufficient generalizability to deal with unknown malicious samples,they require a large number of new samples for retraining.Considering the cost of data collection and labeling,data augmentation is an ideal solution.We propose an optimized noise-based traffic data augmentation system,ONTDAS.The system uses a gradient-based searching algorithm and an improved Bayesian optimizer to obtain optimized noise.The noise is injected into the original samples for data augmentation.Then,an improved bagging algorithm is used to integrate all the base traffic classifiers trained on noised datasets.The experiments verify ONTDAS on 6 types of base classifiers and 4 publicly available datasets respectively.The results show that ONTDAS can effectively enhance the traffic classifiers’performance and significantly improve their generalizability on unknown malicious samples.The system can also alleviate dataset imbalance.Moreover,the performance of ONTDAS is significantly superior to the existing data augmentation methods mentioned.展开更多
Supervised learning classification has arisen as a powerful tool to perform data-driven fault diagnosis in dynamical systems,achieving astonishing results.This approach assumes the availability of extensive,diverse an...Supervised learning classification has arisen as a powerful tool to perform data-driven fault diagnosis in dynamical systems,achieving astonishing results.This approach assumes the availability of extensive,diverse and labeled data corpora for train-ing.However,in some applications it may be difficult or not feasible to obtain a large and balanced dataset including enough representative instances of the fault behaviors of interest.This fact leads to the issues of data scarcity and class imbalance,greatly affecting the performance of supervised learning classifiers.Datasets from railway systems are usually both,scarce and imbalanced,turning supervised learning-based fault diagnosis into a highly challenging task.This article addresses time-series data augmentation for fault diagnosis purposes and presents two application cases in the context of railway track.The case studies employ generative adversarial networks(GAN)schemes to produce realistic synthetic samples of geometrical and structural track defects.The goal is to generate samples that enhance fault diagnosis performance;therefore,major attention was paid not only in the generation process,but also in the synthesis quality assessment,to guarantee the suitability of the samples for training of supervised learning classification models.In the first application,a convolutional classifier achieved a test accuracy of 87.5%for the train on synthetic,test on real(TSTR)scenario,while,in the second application,a fully-connected classifier achieved 96.18%in test accuracy for TSTR.The results indicate that the proposed augmentation approach produces samples having equivalent statistical characteristics and leading to a similar classification behavior as real data.展开更多
Medical image classification is crucial in disease diagnosis,treatment planning,and clinical decisionmaking.We introduced a novel medical image classification approach that integrates Bayesian Random Semantic Data Aug...Medical image classification is crucial in disease diagnosis,treatment planning,and clinical decisionmaking.We introduced a novel medical image classification approach that integrates Bayesian Random Semantic Data Augmentation(BSDA)with a Vision Mamba-based model for medical image classification(MedMamba),enhanced by residual connection blocks,we named the model BSDA-Mamba.BSDA augments medical image data semantically,enhancing the model’s generalization ability and classification performance.MedMamba,a deep learning-based state space model,excels in capturing long-range dependencies in medical images.By incorporating residual connections,BSDA-Mamba further improves feature extraction capabilities.Through comprehensive experiments on eight medical image datasets,we demonstrate that BSDA-Mamba outperforms existing models in accuracy,area under the curve,and F1-score.Our results highlight BSDA-Mamba’s potential as a reliable tool for medical image analysis,particularly in handling diverse imaging modalities from X-rays to MRI.The open-sourcing of our model’s code and datasets,will facilitate the reproduction and extension of our work.展开更多
Deep learning has been fully verified and accepted in the field of electromagnetic signal classification. However, in many specific scenarios, such as radio resource management for aircraft communications, labeled dat...Deep learning has been fully verified and accepted in the field of electromagnetic signal classification. However, in many specific scenarios, such as radio resource management for aircraft communications, labeled data are difficult to obtain, which makes the best deep learning methods at present seem almost powerless, because these methods need a large amount of labeled data for training. When the training dataset is small, it is highly possible to fall into overfitting, which causes performance degradation of the deep neural network. For few-shot electromagnetic signal classification, data augmentation is one of the most intuitive countermeasures. In this work, a generative adversarial network based on the data augmentation method is proposed to achieve better classification performance for electromagnetic signals. Based on the similarity principle, a screening mechanism is established to obtain high-quality generated signals. Then, a data union augmentation algorithm is designed by introducing spatiotemporally flipped shapes of the signal. To verify the effectiveness of the proposed data augmentation algorithm, experiments are conducted on the RADIOML 2016.04C dataset and real-world ACARS dataset. The experimental results show that the proposed method significantly improves the performance of few-shot electromagnetic signal classification.展开更多
With the advent of deep learning,self-driving schemes based on deep learning are becoming more and more popular.Robust perception-action models should learn from data with different scenarios and real behaviors,while ...With the advent of deep learning,self-driving schemes based on deep learning are becoming more and more popular.Robust perception-action models should learn from data with different scenarios and real behaviors,while current end-to-end model learning is generally limited to training of massive data,innovation of deep network architecture,and learning in-situ model in a simulation environment.Therefore,we introduce a new image style transfer method into data augmentation,and improve the diversity of limited data by changing the texture,contrast ratio and color of the image,and then it is extended to the scenarios that the model has been unobserved before.Inspired by rapid style transfer and artistic style neural algorithms,we propose an arbitrary style generation network architecture,including style transfer network,style learning network,style loss network and multivariate Gaussian distribution function.The style embedding vector is randomly sampled from the multivariate Gaussian distribution and linearly interpolated with the embedded vector predicted by the input image on the style learning network,which provides a set of normalization constants for the style transfer network,and finally realizes the diversity of the image style.In order to verify the effectiveness of the method,image classification and simulation experiments were performed separately.Finally,we built a small-sized smart car experiment platform,and apply the data augmentation technology based on image style transfer drive to the experiment of automatic driving for the first time.The experimental results show that:(1)The proposed scheme can improve the prediction accuracy of the end-to-end model and reduce the model’s error accumulation;(2)the method based on image style transfer provides a new scheme for data augmentation technology,and also provides a solution for the high cost that many deep models rely heavily on a large number of label data.展开更多
A data augmentation technique is employed in the current work on a training dataset of 610 bulk metallic glasses(BMGs),which are randomly selected from 762 collected data.An ensemble machine learning(ML)model is devel...A data augmentation technique is employed in the current work on a training dataset of 610 bulk metallic glasses(BMGs),which are randomly selected from 762 collected data.An ensemble machine learning(ML)model is developed on augmented training dataset and tested by the rest 152 data.The result shows that ML model has the ability to predict the maximal diameter Dmaxof BMGs more accurate than all reported ML models.In addition,the novel ML model gives the glass forming ability(GFA)rules:average atomic radius ranging from 140 pm to 165 pm,the value of TT/(T-T)(T-T)being higher than 2.5,the entropy of mixing being higher than 10 J/K/mol,and the enthalpy of mixing ranging from-32 k J/mol to-26 k J/mol.ML model is interpretative,thereby deepening the understanding of GFA.展开更多
This paper investigates the problem of data scarcity in spectrum prediction.A cognitive radio equipment may frequently switch the target frequency as the electromagnetic environment changes.The previously trained mode...This paper investigates the problem of data scarcity in spectrum prediction.A cognitive radio equipment may frequently switch the target frequency as the electromagnetic environment changes.The previously trained model for prediction often cannot maintain a good performance when facing small amount of historical data of the new target frequency.Moreover,the cognitive radio equipment usually implements the dynamic spectrum access in real time which means the time to recollect the data of the new task frequency band and retrain the model is very limited.To address the above issues,we develop a crossband data augmentation framework for spectrum prediction by leveraging the recent advances of generative adversarial network(GAN)and deep transfer learning.Firstly,through the similarity measurement,we pre-train a GAN model using the historical data of the frequency band that is the most similar to the target frequency band.Then,through the data augmentation by feeding the small amount of the target data into the pre-trained GAN,temporal-spectral residual network is further trained using deep transfer learning and the generated data with high similarity from GAN.Finally,experiment results demonstrate the effectiveness of the proposed framework.展开更多
Inspired by the iterative procedure of computing mean fields with known Reynolds stresses(Guo et al.,Theor Appl Mech Lett,2021),we proposed a way to achieve data augmentation by utilizing the intermediate mean fields ...Inspired by the iterative procedure of computing mean fields with known Reynolds stresses(Guo et al.,Theor Appl Mech Lett,2021),we proposed a way to achieve data augmentation by utilizing the intermediate mean fields after proper selections.We also proposed modifications to the Tensor Basis Neural Network(Ling et al.,J Fluid Mech,2016)model.With the modification of the learning targets and the inclusions of wall distance and logarithm of normalized eddy viscosity in the model inputs,the modified version of the model with augmented training datasets shows better performance on Reynolds stress predictions for two dimensional incompressible flow over periodic hills under different geometries.Furthermore,better propagated mean velocity fields can be achieved,showing better agreements with the direct numerical simulations(DNS)results.展开更多
A brain tumor is a lethal neurological disease that affects the average performance of the brain and can be fatal.In India,around 15 million cases are diagnosed yearly.To mitigate the seriousness of the tumor it is es...A brain tumor is a lethal neurological disease that affects the average performance of the brain and can be fatal.In India,around 15 million cases are diagnosed yearly.To mitigate the seriousness of the tumor it is essential to diagnose at the beginning.Notwithstanding,the manual evaluation process utilizing Magnetic Resonance Imaging(MRI)causes a few worries,remarkably inefficient and inaccurate brain tumor diagnoses.Similarly,the examination process of brain tumors is intricate as they display high unbalance in nature like shape,size,appearance,and location.Therefore,a precise and expeditious prognosis of brain tumors is essential for implementing the of an implicit treatment.Several computer models adapted to diagnose the tumor,but the accuracy of the model needs to be tested.Considering all the above mentioned things,this work aims to identify the best classification system by considering the prediction accuracy out of Alex-Net,ResNet 50,and Inception V3.Data augmentation is performed on the database and fed into the three convolutions neural network(CNN)models.A comparison line is drawn between the three models based on accuracy and performance.An accuracy of 96.2%is obtained for AlexNet with augmentation and performed better than ResNet 50 and Inception V3 for the 120th epoch.With the suggested model with higher accuracy,it is highly reliable if brain tumors are diagnosed with available datasets.展开更多
文摘The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.
基金supported by the Postdoctoral Fellowship Program(Grade B)of China(GZB20250435)the National Natural Science Foundation of China(62403270).
文摘High-quality data is essential for the success of data-driven learning tasks.The characteristics,precision,and completeness of the datasets critically determine the reliability,interpretability,and effectiveness of subsequent analyzes and applications,such as fault detection,predictive maintenance,and process optimization.However,for many industrial processes,obtaining sufficient high-quality data remains a significant challenge due to high costs,safety concerns,and practical constraints.To overcome these challenges,data augmentation has emerged as a rapidly growing research area,attracting considerable attention across both academia and industry.By expanding datasets,data augmentation techniques improve greater generalization and more robust performance in actual applications.This paper provides a comprehensive,multi-perspective review of data augmentation methods for industrial processes.For clarity and organization,existing studies are systematically grouped into four categories:small sample with low dimension,small sample with high dimension,large sample with low dimension,and large sample with high dimension.Within this framework,the review examines current research from both methodological and application-oriented perspectives,highlighting main methods,advantages,and limitations.By synthesizing these findings,this review offers a structured overview for scholars and practitioners,serving as a valuable reference for newcomers and experienced researchers seeking to explore and advance data augmentation techniques in industrial processes.
基金supported by the Jiangsu Province IUR Cooperation Project (No.BY2021258)the Wuxi Science and Technology Development Fund Project (No.G20212028)。
文摘An improved cycle-consistent generative adversarial network(CycleGAN) method for defect data augmentation based on feature fusion and self attention residual module is proposed to address the insufficiency of defect sample data for light guide plate(LGP) in production,as well as the problem of minor defects.Two optimizations are made to the generator of CycleGAN:fusion of low resolution features obtained from partial up-sampling and down-sampling with high-resolution features,combination of self attention mechanism with residual network structure to replace the original residual module.Qualitative and quantitative experiments were conducted to compare different data augmentation methods,and the results show that the defect images of the LGP generated by the improved network were more realistic,and the accuracy of the you only look once version 5(YOLOv5) detection network for the LGP was improved by 5.6%,proving the effectiveness and accuracy of the proposed method.
基金supported by the King Abdullah University of Science and Technology(KAUST)。
文摘Seismic data denoising is a critical process usually applied at various stages of the seismic processing workflow,as our ability to mitigate noise in seismic data affects the quality of our subsequent analyses.However,finding an optimal balance between preserving seismic signals and effectively reducing seismic noise presents a substantial challenge.In this study,we introduce a multi-stage deep learning model,trained in a self-supervised manner,designed specifically to suppress seismic noise while minimizing signal leakage.This model operates as a patch-based approach,extracting overlapping patches from the noisy data and converting them into 1D vectors for input.It consists of two identical sub-networks,each configured differently.Inspired by the transformer architecture,each sub-network features an embedded block that comprises two fully connected layers,which are utilized for feature extraction from the input patches.After reshaping,a multi-head attention module enhances the model’s focus on significant features by assigning higher attention weights to them.The key difference between the two sub-networks lies in the number of neurons within their fully connected layers.The first sub-network serves as a strong denoiser with a small number of neurons,effectively attenuating seismic noise;in contrast,the second sub-network functions as a signal-add-back model,using a larger number of neurons to retrieve some of the signal that was not preserved in the output of the first sub-network.The proposed model produces two outputs,each corresponding to one of the sub-networks,and both sub-networks are optimized simultaneously using the noisy data as the label for both outputs.Evaluations conducted on both synthetic and field data demonstrate the model’s effectiveness in suppressing seismic noise with minimal signal leakage,outperforming some benchmark methods.
基金funded by the Bavarian State Ministry of Science,Research and Art(Grant number:H.2-F1116.WE/52/2)。
文摘In order to address the widespread data shortage problem in battery research,this paper proposes a generative adversarial network model that combines it with deep convolutional networks,the Wasserstein distance,and the gradient penalty to achieve data augmentation.To lower the threshold for implementing the proposed method,transfer learning is further introduced.The W-DC-GAN-GP-TL framework is thereby formed.This framework is evaluated on 3 different publicly available datasets to judge the quality of generated data.Through visual comparisons and the examination of two visualization methods(probability density function(PDF)and principal component analysis(PCA)),it is demonstrated that the generated data is hard to distinguish from the real data.The application of generated data for training a battery state model using transfer learning is further evaluated.Specifically,Bi-GRU-based and Transformer-based methods are implemented on 2 separate datasets for estimating state of health(SOH)and state of charge(SOC),respectively.The results indicate that the proposed framework demonstrates satisfactory performance in different scenarios:for the data replacement scenario,where real data are removed and replaced with generated data,the state estimator accuracy decreases only slightly;for the data enhancement scenario,the estimator accuracy is further improved.The estimation accuracy of SOH and SOC is as low as 0.69%and 0.58%root mean square error(RMSE)after applying the proposed framework.This framework provides a reliable method for enriching battery measurement data.It is a generalized framework capable of generating a variety of time series data.
文摘Bird species classification is not only a challenging topic in artificial intelligence but also a domain closely related to environmental protection and ecological research.Additionally,performing edge computing on low-level devices using small neural networks can be an important research direction.In this paper,we use the EfficientNetV2B0 model for bird species classification,applying transfer learning on a dataset of 525 bird species.We also employ the BiRefNet model to remove backgrounds from images in the training set.The generated background-removed images are mixed with the original training set as a form of data augmentation.We aim for these background-removed images to help the model focus on key features,and by combining data augmentation with transfer learning,we trained a highly accurate and efficient bird species classification model.The training process is divided into a transfer learning stage and a fine-tuning stage.In the transfer learning stage,only the newly added custom layers are trained;while in the fine-tuning stage,all pre-trained layers except for the batch normalization layers are fine-tuned.According to the experimental results,the proposed model not only has an advantage in size compared to other models but also outperforms them in various metrics.The training results show that the proposed model achieved an accuracy of 99.54%and a precision of 99.62%,demonstrating that it achieves both lightweight design and high accuracy.To confirm the credibility of the results,we use heatmaps to interpret the model.The heatmaps show that our model can clearly highlight the image feature area.In addition,we also perform the 10-fold cross-validation on the model to verify its credibility.Finally,this paper proposes a model with low training cost and high accuracy,making it suitable for deployment on edge computing devices to provide lighter and more convenient services.
基金supported by the National Natural Science Foundation of China(Grant No.42101089)Sichuan Science and Technology Program(2022YFS0586)the Open Fund of Key Laboratory of Mountain Hazards and Earth Surface Processes Chinese Academy of Sciences.
文摘Rainfall-induced shallow landslides pose one of significant geological hazards,necessitating precise monitoring and prediction for effective disaster mitigation.Most studies on landslide prediction have focused on optimizing machine learning(ML)algorithms,very limited attention has been paid to enhancing data quality for improved predictive performance.This study employs strategic data augmentation(DA)techniques to enhance the accuracy of shallow landslide prediction.Using five DA methods including singular spectrum analysis(SSA),moving averages(MA),wavelet denoising(WD),variational mode decomposition(VMD),and linear interpolation(LI),we utilize strategies such as smoothing,denoising,trend decomposition,and synthetic data generation to improve the training dataset.Four machine learning algorithms,i.e.artificial neural network(ANN),recurrent neural network(RNN),one-dimensional convolutional neural network(CNN1D),and long short-term memory(LSTM),are used to forecast landslide displacement.The case study of a landslide in southwest China shows the effectiveness of our approach in predicting landslide displacements,despite the inherent limitations of the monitoring dataset.VMD proves the most effective for smoothing and denoising,improving R^(2),RMSE,and MAPE by 172.16%,71.82%,and 98.9%,respectively.SSA addresses missing data,while LI is effective with limited data samples,improving metrics by 21.6%,52.59%,and 47.87%,respectively.This study demonstrates the potential of DA techniques to mitigate the impact of data defects on landslide prediction accuracy,with implications for similar cases.
基金Postgraduate Innovation Top notch Talent Training Project of Hunan Province,Grant/Award Number:CX20220045Scientific Research Project of National University of Defense Technology,Grant/Award Number:22-ZZCX-07+2 种基金New Era Education Quality Project of Anhui Province,Grant/Award Number:2023cxcysj194National Natural Science Foundation of China,Grant/Award Numbers:62201597,62205372,1210456foundation of Hefei Comprehensive National Science Center,Grant/Award Number:KY23C502。
文摘Large-scale point cloud datasets form the basis for training various deep learning networks and achieving high-quality network processing tasks.Due to the diversity and robustness constraints of the data,data augmentation(DA)methods are utilised to expand dataset diversity and scale.However,due to the complex and distinct characteristics of LiDAR point cloud data from different platforms(such as missile-borne and vehicular LiDAR data),directly applying traditional 2D visual domain DA methods to 3D data can lead to networks trained using this approach not robustly achieving the corresponding tasks.To address this issue,the present study explores DA for missile-borne LiDAR point cloud using a Monte Carlo(MC)simulation method that closely resembles practical application.Firstly,the model of multi-sensor imaging system is established,taking into account the joint errors arising from the platform itself and the relative motion during the imaging process.A distortion simulation method based on MC simulation for augmenting missile-borne LiDAR point cloud data is proposed,underpinned by an analysis of combined errors between different modal sensors,achieving high-quality augmentation of point cloud data.The effectiveness of the proposed method in addressing imaging system errors and distortion simulation is validated using the imaging scene dataset constructed in this paper.Comparative experiments between the proposed point cloud DA algorithm and the current state-of-the-art algorithms in point cloud detection and single object tracking tasks demonstrate that the proposed method can improve the network performance obtained from unaugmented datasets by over 17.3%and 17.9%,surpassing SOTA performance of current point cloud DA algorithms.
基金supported by National Natural Science Foundation of China(61673186 and 61871196)Beijing Normal University Education Reform Project(jx2024040)Guangdong Undergraduate Universities Teaching Quality and Reform Project(jx2024309).
文摘Data augmentation plays an important role in boosting the performance of 3D models,while very few studies handle the 3D point cloud data with this technique.Global augmentation and cut-paste are commonly used augmentation techniques for point clouds,where global augmentation is applied to the entire point cloud of the scene,and cut-paste samples objects from other frames into the current frame.Both types of data augmentation can improve performance,but the cut-paste technique cannot effectively deal with the occlusion relationship between the foreground object and the background scene and the rationality of object sampling,which may be counterproductive and may hurt the overall performance.In addition,LiDAR is susceptible to signal loss,external occlusion,extreme weather and other factors,which can easily cause object shape changes,while global augmentation and cut-paste cannot effectively enhance the robustness of the model.To this end,we propose Syn-Aug,a synchronous data augmentation framework for LiDAR-based 3D object detection.Specifically,we first propose a novel rendering-based object augmentation technique(Ren-Aug)to enrich training data while enhancing scene realism.Second,we propose a local augmentation technique(Local-Aug)to generate local noise by rotating and scaling objects in the scene while avoiding collisions,which can improve generalisation performance.Finally,we make full use of the structural information of 3D labels to make the model more robust by randomly changing the geometry of objects in the training frames.We verify the proposed framework with four different types of 3D object detectors.Experimental results show that our proposed Syn-Aug significantly improves the performance of various 3D object detectors in the KITTI and nuScenes datasets,proving the effectiveness and generality of Syn-Aug.On KITTI,four different types of baseline models using Syn-Aug improved mAP by 0.89%,1.35%,1.61%and 1.14%respectively.On nuScenes,four different types of baseline models using Syn-Aug improved mAP by 14.93%,10.42%,8.47%and 6.81%respectively.The code is available at https://github.com/liuhuaijjin/Syn-Aug.
基金Natural Science Foundation of Zhejiang Province,Grant/Award Number:LY23F020025Science and Technology Commissioner Program of Huzhou,Grant/Award Number:2023GZ42Sichuan Provincial Science and Technology Support Program,Grant/Award Numbers:2023ZHCG0005,2023ZHCG0008。
文摘Data augmentation plays an important role in training deep neural model by expanding the size and diversity of the dataset.Initially,data augmentation mainly involved some simple transformations of images.Later,in order to increase the diversity and complexity of data,more advanced methods appeared and evolved to sophisticated generative models.However,these methods required a mass of computation of training or searching.In this paper,a novel training-free method that utilises the Pre-Trained Segment Anything Model(SAM)model as a data augmentation tool(PTSAM-DA)is proposed to generate the augmented annotations for images.Without the need for training,it obtains prompt boxes from the original annotations and then feeds the boxes to the pre-trained SAM to generate diverse and improved annotations.In this way,annotations are augmented more ingenious than simple manipulations without incurring huge computation for training a data augmentation model.Multiple comparative experiments on three datasets are conducted,including an in-house dataset,ADE20K and COCO2017.On this in-house dataset,namely Agricultural Plot Segmentation Dataset,maximum improvements of 3.77%and 8.92%are gained in two mainstream metrics,mIoU and mAcc,respectively.Consequently,large vision models like SAM are proven to be promising not only in image segmentation but also in data augmentation.
基金support of the“National R&D Project for Smart Construction Technology (Grant No.RS-2020-KA157074)”funded by the Korea Agency for Infrastructure Technology Advancement under the Ministry of Land,Infrastructure and Transport,and managed by the Korea Expressway Corporation.
文摘The widespread adoption of tunnel boring machines(TBMs)has led to an increased focus on disc cutter wear,including both normal and abnormal types,for efficient and safe TBM excavation.However,abnormal wear has yet to be thoroughly investigated,primarily due to the complexity of considering mixed ground conditions and the imbalance in the number of instances between the two types of wear.This study developed a prediction model for abnormal TBM disc cutter wear,considering mixed ground conditions,by employing interpretable machine learning with data augmentation.An equivalent elastic modulus was used to consider the characteristics of mixed ground conditions,and wear data was obtained from 65 cutterhead intervention(CHI)reports covering both mixed ground and hard rock sections.With a balanced training dataset obtained by data augmentation,an extreme gradient boosting(XGB)model delivered acceptable results with an accuracy of 0.94,an F1-score of 0.808,and a recall of 0.8.In addition,the accuracy for each individual disc cutter exhibited low variability.When employing data augmentation,a significant improvement in recall was observed compared to when it was not used,although the difference in accuracy and F1-score was marginal.The subsequent model interpretation revealed the chamber pressure,cutter installation radius,and torque as significant contributors.Specifically,a threshold in chamber pressure was observed,which could induce abnormal wear.The study also explored how elevated values of these influential contributors correlate with abnormal wear.The proposed model offers a valuable tool for planning the replacement of abnormally worn disc cutters,enhancing the safety and efficiency of TBM operations.
基金supported in part by the National Key Research and Development Program of China(No.2022YFB4500800)the National Science Foundation of China(No.42071431).
文摘With the emergence of new attack techniques,traffic classifiers usually fail to maintain the expected performance in real-world network environments.In order to have sufficient generalizability to deal with unknown malicious samples,they require a large number of new samples for retraining.Considering the cost of data collection and labeling,data augmentation is an ideal solution.We propose an optimized noise-based traffic data augmentation system,ONTDAS.The system uses a gradient-based searching algorithm and an improved Bayesian optimizer to obtain optimized noise.The noise is injected into the original samples for data augmentation.Then,an improved bagging algorithm is used to integrate all the base traffic classifiers trained on noised datasets.The experiments verify ONTDAS on 6 types of base classifiers and 4 publicly available datasets respectively.The results show that ONTDAS can effectively enhance the traffic classifiers’performance and significantly improve their generalizability on unknown malicious samples.The system can also alleviate dataset imbalance.Moreover,the performance of ONTDAS is significantly superior to the existing data augmentation methods mentioned.
基金supported by the German Research Foundation(DFG)under the project“Efficient Sensor-Based Condition Monitoring Methodology for the Detection and Localization of Faults on the Railway Track(ConMoRAIL)”,Grant No.515687155.
文摘Supervised learning classification has arisen as a powerful tool to perform data-driven fault diagnosis in dynamical systems,achieving astonishing results.This approach assumes the availability of extensive,diverse and labeled data corpora for train-ing.However,in some applications it may be difficult or not feasible to obtain a large and balanced dataset including enough representative instances of the fault behaviors of interest.This fact leads to the issues of data scarcity and class imbalance,greatly affecting the performance of supervised learning classifiers.Datasets from railway systems are usually both,scarce and imbalanced,turning supervised learning-based fault diagnosis into a highly challenging task.This article addresses time-series data augmentation for fault diagnosis purposes and presents two application cases in the context of railway track.The case studies employ generative adversarial networks(GAN)schemes to produce realistic synthetic samples of geometrical and structural track defects.The goal is to generate samples that enhance fault diagnosis performance;therefore,major attention was paid not only in the generation process,but also in the synthesis quality assessment,to guarantee the suitability of the samples for training of supervised learning classification models.In the first application,a convolutional classifier achieved a test accuracy of 87.5%for the train on synthetic,test on real(TSTR)scenario,while,in the second application,a fully-connected classifier achieved 96.18%in test accuracy for TSTR.The results indicate that the proposed augmentation approach produces samples having equivalent statistical characteristics and leading to a similar classification behavior as real data.
文摘Medical image classification is crucial in disease diagnosis,treatment planning,and clinical decisionmaking.We introduced a novel medical image classification approach that integrates Bayesian Random Semantic Data Augmentation(BSDA)with a Vision Mamba-based model for medical image classification(MedMamba),enhanced by residual connection blocks,we named the model BSDA-Mamba.BSDA augments medical image data semantically,enhancing the model’s generalization ability and classification performance.MedMamba,a deep learning-based state space model,excels in capturing long-range dependencies in medical images.By incorporating residual connections,BSDA-Mamba further improves feature extraction capabilities.Through comprehensive experiments on eight medical image datasets,we demonstrate that BSDA-Mamba outperforms existing models in accuracy,area under the curve,and F1-score.Our results highlight BSDA-Mamba’s potential as a reliable tool for medical image analysis,particularly in handling diverse imaging modalities from X-rays to MRI.The open-sourcing of our model’s code and datasets,will facilitate the reproduction and extension of our work.
基金supported by the National Natural Science Foundation of China(No.61772401)the Fundamental Research Funds for the Central Universities(No.RW180177)supported by the Science and Technology on Communication Information Security Control Laboratory。
文摘Deep learning has been fully verified and accepted in the field of electromagnetic signal classification. However, in many specific scenarios, such as radio resource management for aircraft communications, labeled data are difficult to obtain, which makes the best deep learning methods at present seem almost powerless, because these methods need a large amount of labeled data for training. When the training dataset is small, it is highly possible to fall into overfitting, which causes performance degradation of the deep neural network. For few-shot electromagnetic signal classification, data augmentation is one of the most intuitive countermeasures. In this work, a generative adversarial network based on the data augmentation method is proposed to achieve better classification performance for electromagnetic signals. Based on the similarity principle, a screening mechanism is established to obtain high-quality generated signals. Then, a data union augmentation algorithm is designed by introducing spatiotemporally flipped shapes of the signal. To verify the effectiveness of the proposed data augmentation algorithm, experiments are conducted on the RADIOML 2016.04C dataset and real-world ACARS dataset. The experimental results show that the proposed method significantly improves the performance of few-shot electromagnetic signal classification.
基金the National Natural Science Foundation of China(51965008)Science and Technology projects of Guizhou[2018]2168Excellent Young Researcher Project of Guizhou[2017]5630.
文摘With the advent of deep learning,self-driving schemes based on deep learning are becoming more and more popular.Robust perception-action models should learn from data with different scenarios and real behaviors,while current end-to-end model learning is generally limited to training of massive data,innovation of deep network architecture,and learning in-situ model in a simulation environment.Therefore,we introduce a new image style transfer method into data augmentation,and improve the diversity of limited data by changing the texture,contrast ratio and color of the image,and then it is extended to the scenarios that the model has been unobserved before.Inspired by rapid style transfer and artistic style neural algorithms,we propose an arbitrary style generation network architecture,including style transfer network,style learning network,style loss network and multivariate Gaussian distribution function.The style embedding vector is randomly sampled from the multivariate Gaussian distribution and linearly interpolated with the embedded vector predicted by the input image on the style learning network,which provides a set of normalization constants for the style transfer network,and finally realizes the diversity of the image style.In order to verify the effectiveness of the method,image classification and simulation experiments were performed separately.Finally,we built a small-sized smart car experiment platform,and apply the data augmentation technology based on image style transfer drive to the experiment of automatic driving for the first time.The experimental results show that:(1)The proposed scheme can improve the prediction accuracy of the end-to-end model and reduce the model’s error accumulation;(2)the method based on image style transfer provides a new scheme for data augmentation technology,and also provides a solution for the high cost that many deep models rely heavily on a large number of label data.
基金the National Key R&D Program of China(No.2018YFB0704404)the Guangdong Basic and Applied Basic Research Foundation(No.2020A1515110798)+1 种基金the National Natural Science Foundation of China(Grant Nos.91860115)the Stable Supporting Fund of Shenzhen(GXWD20201230155427003-20200728114835006)。
文摘A data augmentation technique is employed in the current work on a training dataset of 610 bulk metallic glasses(BMGs),which are randomly selected from 762 collected data.An ensemble machine learning(ML)model is developed on augmented training dataset and tested by the rest 152 data.The result shows that ML model has the ability to predict the maximal diameter Dmaxof BMGs more accurate than all reported ML models.In addition,the novel ML model gives the glass forming ability(GFA)rules:average atomic radius ranging from 140 pm to 165 pm,the value of TT/(T-T)(T-T)being higher than 2.5,the entropy of mixing being higher than 10 J/K/mol,and the enthalpy of mixing ranging from-32 k J/mol to-26 k J/mol.ML model is interpretative,thereby deepening the understanding of GFA.
基金This work was supported by the Science and Technology Innovation 2030-Key Project of“New Generation Artificial Intelligence”of China under Grant 2018AAA0102303the Natural Science Foundation for Distinguished Young Scholars of Jiangsu Province(No.BK20190030)the National Natural Science Foundation of China(No.61631020,No.61871398,No.61931011 and No.U20B2038).
文摘This paper investigates the problem of data scarcity in spectrum prediction.A cognitive radio equipment may frequently switch the target frequency as the electromagnetic environment changes.The previously trained model for prediction often cannot maintain a good performance when facing small amount of historical data of the new target frequency.Moreover,the cognitive radio equipment usually implements the dynamic spectrum access in real time which means the time to recollect the data of the new task frequency band and retrain the model is very limited.To address the above issues,we develop a crossband data augmentation framework for spectrum prediction by leveraging the recent advances of generative adversarial network(GAN)and deep transfer learning.Firstly,through the similarity measurement,we pre-train a GAN model using the historical data of the frequency band that is the most similar to the target frequency band.Then,through the data augmentation by feeding the small amount of the target data into the pre-trained GAN,temporal-spectral residual network is further trained using deep transfer learning and the generated data with high similarity from GAN.Finally,experiment results demonstrate the effectiveness of the proposed framework.
基金This work was supported by the National Nat-420 ural Science Foundation of China(Grant Nos.11822208,11988102,42111772297,91852205)the Fundamental Research Funds for the 422 central Universities.
文摘Inspired by the iterative procedure of computing mean fields with known Reynolds stresses(Guo et al.,Theor Appl Mech Lett,2021),we proposed a way to achieve data augmentation by utilizing the intermediate mean fields after proper selections.We also proposed modifications to the Tensor Basis Neural Network(Ling et al.,J Fluid Mech,2016)model.With the modification of the learning targets and the inclusions of wall distance and logarithm of normalized eddy viscosity in the model inputs,the modified version of the model with augmented training datasets shows better performance on Reynolds stress predictions for two dimensional incompressible flow over periodic hills under different geometries.Furthermore,better propagated mean velocity fields can be achieved,showing better agreements with the direct numerical simulations(DNS)results.
基金Ahmed Alhussen would like to thank the Deanship of Scientific Research at Majmaah University for supporting this work under Project No.R-2022-####.
文摘A brain tumor is a lethal neurological disease that affects the average performance of the brain and can be fatal.In India,around 15 million cases are diagnosed yearly.To mitigate the seriousness of the tumor it is essential to diagnose at the beginning.Notwithstanding,the manual evaluation process utilizing Magnetic Resonance Imaging(MRI)causes a few worries,remarkably inefficient and inaccurate brain tumor diagnoses.Similarly,the examination process of brain tumors is intricate as they display high unbalance in nature like shape,size,appearance,and location.Therefore,a precise and expeditious prognosis of brain tumors is essential for implementing the of an implicit treatment.Several computer models adapted to diagnose the tumor,but the accuracy of the model needs to be tested.Considering all the above mentioned things,this work aims to identify the best classification system by considering the prediction accuracy out of Alex-Net,ResNet 50,and Inception V3.Data augmentation is performed on the database and fed into the three convolutions neural network(CNN)models.A comparison line is drawn between the three models based on accuracy and performance.An accuracy of 96.2%is obtained for AlexNet with augmentation and performed better than ResNet 50 and Inception V3 for the 120th epoch.With the suggested model with higher accuracy,it is highly reliable if brain tumors are diagnosed with available datasets.