The spatial offset of bridge has a significant impact on the safety,comfort,and durability of high-speed railway(HSR)operations,so it is crucial to rapidly and effectively detect the spatial offset of operational HSR ...The spatial offset of bridge has a significant impact on the safety,comfort,and durability of high-speed railway(HSR)operations,so it is crucial to rapidly and effectively detect the spatial offset of operational HSR bridges.Drive-by monitoring of bridge uneven settlement demonstrates significant potential due to its practicality,cost-effectiveness,and efficiency.However,existing drive-by methods for detecting bridge offset have limitations such as reliance on a single data source,low detection accuracy,and the inability to identify lateral deformations of bridges.This paper proposes a novel drive-by inspection method for spatial offset of HSR bridge based on multi-source data fusion of comprehensive inspection train.Firstly,dung beetle optimizer-variational mode decomposition was employed to achieve adaptive decomposition of non-stationary dynamic signals,and explore the hidden temporal relationships in the data.Subsequently,a long short-term memory neural network was developed to achieve feature fusion of multi-source signal and accurate prediction of spatial settlement of HSR bridge.A dataset of track irregularities and CRH380A high-speed train responses was generated using a 3D train-track-bridge interaction model,and the accuracy and effectiveness of the proposed hybrid deep learning model were numerically validated.Finally,the reliability of the proposed drive-by inspection method was further validated by analyzing the actual measurement data obtained from comprehensive inspection train.The research findings indicate that the proposed approach enables rapid and accurate detection of spatial offset in HSR bridge,ensuring the long-term operational safety of HSR bridges.展开更多
In traditional medicine and ethnomedicine,medicinal plants have long been recognized as the basis for materials in therapeutic applications worldwide.In particular,the remarkable curative effect of traditional Chinese...In traditional medicine and ethnomedicine,medicinal plants have long been recognized as the basis for materials in therapeutic applications worldwide.In particular,the remarkable curative effect of traditional Chinese medicine during corona virus disease 2019(COVID-19)pandemic has attracted extensive attention globally.Medicinal plants have,therefore,become increasingly popular among the public.However,with increasing demand for and profit with medicinal plants,commercial fraudulent events such as adulteration or counterfeits sometimes occur,which poses a serious threat to the clinical outcomes and interests of consumers.With rapid advances in artificial intelligence,machine learning can be used to mine information on various medicinal plants to establish an ideal resource database.We herein present a review that mainly introduces common machine learning algorithms and discusses their application in multi-source data analysis of medicinal plants.The combination of machine learning algorithms and multi-source data analysis facilitates a comprehensive analysis and aids in the effective evaluation of the quality of medicinal plants.The findings of this review provide new possibilities for promoting the development and utilization of medicinal plants.展开更多
With the recent increase in data volume and diversity,traditional text representation techniques are struggling to capture context,particularly in environments with sparse data.To address these challenges,this study p...With the recent increase in data volume and diversity,traditional text representation techniques are struggling to capture context,particularly in environments with sparse data.To address these challenges,this study proposes a new model,the Masked Joint Representation Model(MJRM).MJRM approximates the original hypothesis by leveraging multiple elements in a limited context.It dynamically adapts to changes in characteristics based on data distribution through three main components.First,masking-based representation learning,termed selective dynamic masking,integrates topic modeling and sentiment clustering to generate and train multiple instances across different data subsets,whose predictions are then aggregated with optimized weights.This design alleviates sparsity,suppresses noise,and preserves contextual structures.Second,regularization-based improvements are applied.Third,techniques for addressing sparse data are used to perform final inference.As a result,MJRM improves performance by up to 4%compared to existing AI techniques.In our experiments,we analyzed the contribution of each factor,demonstrating that masking,dynamic learning,and aggregating multiple instances complement each other to improve performance.This demonstrates that a masking-based multi-learning strategy is effective for context-aware sparse text classification,and can be useful even in challenging situations such as data shortage or data distribution variations.We expect that the approach can be extended to diverse fields such as sentiment analysis,spam filtering,and domain-specific document classification.展开更多
Near real-time maize phenology monitoring is crucial for field management,cropping system adjustments,and yield estimation.Most phenological monitoring methods are post-seasonal and heavily rely on high-frequency time...Near real-time maize phenology monitoring is crucial for field management,cropping system adjustments,and yield estimation.Most phenological monitoring methods are post-seasonal and heavily rely on high-frequency time-series data.These methods are not applicable on the unmanned aerial vehicle(UAV)platform due to the high cost of acquiring time-series UAV images and the shortage of UAV-based phenological monitoring methods.To address these challenges,we employed the Synthetic Minority Oversampling Technique(SMOTE)for sample augmentation,aiming to resolve the small sample modelling problem.Moreover,we utilized enhanced"separation"and"compactness"feature selection methods to identify input features from multiple data sources.In this process,we incorporated dynamic multi-source data fusion strategies,involving Vegetation index(VI),Color index(CI),and Texture features(TF).A two-stage neural network that combines Convolutional Neural Network(CNN)and Long Short-Term Memory Network(LSTM)is proposed to identify maize phenological stages(including sowing,seedling,jointing,trumpet,tasseling,maturity,and harvesting)on UAV platforms.The results indicate that the dataset generated by SMOTE closely resembles the measured dataset.Among dynamic data fusion strategies,the VI-TF combination proves to be most effective,with CI-TF and VI-CI combinations following behind.Notably,as more data sources are integrated,the model's demand for input features experiences a significant decline.In particular,the CNN-LSTM model,based on the fusion of three data sources,exhibited remarkable reliability when validating the three datasets.For Dataset 1(Beijing Xiaotangshan,2023:Data from 12 UAV Flight Missions),the model achieved an overall accuracy(OA)of 86.53%.Additionally,its precision(Pre),recall(Rec),F1 score(F1),false acceptance rate(FAR),and false rejection rate(FRR)were 0.89,0.89,0.87,0.11,and 0.11,respectively.The model also showed strong generalizability in Dataset 2(Beijing Xiaotangshan,2023:Data from 6 UAV Flight Missions)and Dataset 3(Beijing Xiaotangshan,2022:Data from 4 UAV Flight Missions),with OAs of 89.4%and 85%,respectively.Meanwhile,the model has a low demand for input featu res,requiring only 54.55%(99 of all featu res).The findings of this study not only offer novel insights into near real-time crop phenology monitoring,but also provide technical support for agricultural field management and cropping system adaptation.展开更多
Accurate monitoring of track irregularities is very helpful to improving the vehicle operation quality and to formulating appropriate track maintenance strategies.Existing methods have the problem that they rely on co...Accurate monitoring of track irregularities is very helpful to improving the vehicle operation quality and to formulating appropriate track maintenance strategies.Existing methods have the problem that they rely on complex signal processing algorithms and lack multi-source data analysis.Driven by multi-source measurement data,including the axle box,the bogie frame and the carbody accelerations,this paper proposes a track irregularities monitoring network(TIMNet)based on deep learning methods.TIMNet uses the feature extraction capability of convolutional neural networks and the sequence map-ping capability of the long short-term memory model to explore the mapping relationship between vehicle accelerations and track irregularities.The particle swarm optimization algorithm is used to optimize the network parameters,so that both the vertical and lateral track irregularities can be accurately identified in the time and spatial domains.The effectiveness and superiority of the proposed TIMNet is analyzed under different simulation conditions using a vehicle dynamics model.Field tests are conducted to prove the availability of the proposed TIMNet in quantitatively monitoring vertical and lateral track irregularities.Furthermore,comparative tests show that the TIMNet has a better fitting degree and timeliness in monitoring track irregularities(vertical R2 of 0.91,lateral R2 of 0.84 and time cost of 10 ms),compared to other classical regression.The test also proves that the TIMNet has a better anti-interference ability than other regression models.展开更多
There is a large amount of information in the network data that we canexploit. It is difficult for classical community detection algorithms to handle network data with sparse topology. Representation learning of netwo...There is a large amount of information in the network data that we canexploit. It is difficult for classical community detection algorithms to handle network data with sparse topology. Representation learning of network data is usually paired with clustering algorithms to solve the community detection problem.Meanwhile, there is always an unpredictable distribution of class clusters outputby graph representation learning. Therefore, we propose an improved densitypeak clustering algorithm (ILDPC) for the community detection problem, whichimproves the local density mechanism in the original algorithm and can betteraccommodate class clusters of different shapes. And we study the communitydetection in network data. The algorithm is paired with the benchmark modelGraph sample and aggregate (GraphSAGE) to show the adaptability of ILDPCfor community detection. The plotted decision diagram shows that the ILDPCalgorithm is more discriminative in selecting density peak points compared tothe original algorithm. Finally, the performance of K-means and other clusteringalgorithms on this benchmark model is compared, and the algorithm is proved tobe more suitable for community detection in sparse networks with the benchmarkmodel on the evaluation criterion F1-score. The sensitivity of the parameters ofthe ILDPC algorithm to the low-dimensional vector set output by the benchmarkmodel GraphSAGE is also analyzed.展开更多
Time-domain airborne electromagnetic(AEM)data are frequently subject to interference from various types of noise,which can reduce the data quality and affect data inversion and interpretation.Traditional denoising met...Time-domain airborne electromagnetic(AEM)data are frequently subject to interference from various types of noise,which can reduce the data quality and affect data inversion and interpretation.Traditional denoising methods primarily deal with data directly,without analyzing the data in detail;thus,the results are not always satisfactory.In this paper,we propose a method based on dictionary learning for EM data denoising.This method uses dictionary learning to perform feature analysis and to extract and reconstruct the true signal.In the process of dictionary learning,the random noise is fi ltered out as residuals.To verify the eff ectiveness of this dictionary learning approach for denoising,we use a fi xed overcomplete discrete cosine transform(ODCT)dictionary algorithm,the method-of-optimal-directions(MOD)dictionary learning algorithm,and the K-singular value decomposition(K-SVD)dictionary learning algorithm to denoise decay curves at single points and to denoise profi le data for diff erent time channels in time-domain AEM.The results show obvious diff erences among the three dictionaries for denoising AEM data,with the K-SVD dictionary achieving the best performance.展开更多
For reservoirs with complex non-Gaussian geological characteristics,such as carbonate reservoirs or reservoirs with sedimentary facies distribution,it is difficult to implement history matching directly,especially for...For reservoirs with complex non-Gaussian geological characteristics,such as carbonate reservoirs or reservoirs with sedimentary facies distribution,it is difficult to implement history matching directly,especially for the ensemble-based data assimilation methods.In this paper,we propose a multi-source information fused generative adversarial network(MSIGAN)model,which is used for parameterization of the complex geologies.In MSIGAN,various information such as facies distribution,microseismic,and inter-well connectivity,can be integrated to learn the geological features.And two major generative models in deep learning,variational autoencoder(VAE)and generative adversarial network(GAN)are combined in our model.Then the proposed MSIGAN model is integrated into the ensemble smoother with multiple data assimilation(ESMDA)method to conduct history matching.We tested the proposed method on two reservoir models with fluvial facies.The experimental results show that the proposed MSIGAN model can effectively learn the complex geological features,which can promote the accuracy of history matching.展开更多
We respond to Schaeben et al.’s1 comment on our paper,“Machine Learning Enhanced Analysis of EBSD Data for Texture Representation.”While their observations are factually correct,they do not disprove our results.Our...We respond to Schaeben et al.’s1 comment on our paper,“Machine Learning Enhanced Analysis of EBSD Data for Texture Representation.”While their observations are factually correct,they do not disprove our results.Our method,TACS,preserves the full distribution of crystallographic orientations and is validated with real-world data.We emphasize the importance of empirical validation over theoretical constructs in assessing machine learning methods’practical performance.展开更多
Deep learning has empowered traffic prediction models to integrate diverse auxiliary data sources,such as weather and temporal features,for enhanced forecasting accuracy.However,existing approaches often suffer from l...Deep learning has empowered traffic prediction models to integrate diverse auxiliary data sources,such as weather and temporal features,for enhanced forecasting accuracy.However,existing approaches often suffer from limited generality and scalability,and the field lacks a unified benchmark for fair model comparison.This absence hinders consistent performance evaluation,slows the development of robust and adaptable models,and makes it challenging to quantify the incremental benefits of different auxiliary data sources.To address these issues,we present MltAuxTSPP,a unified benchmark framework for deep learning-based traffic state prediction with multi-source auxiliary data.The framework features a standardized data container and a fusion embedding module,enabling consistent utilization of heterogeneous data and improving scalability.It produces unified hidden representations that can be seamlessly adopted by various downstream models,ensuring fair and reproducible comparisons under identical conditions.Extensive experiments on real-world datasets demonstrate that MltAuxTSPP effectively leverages weather and temporal features to improve long-term forecast performance and offers a practical and reproducible foundation for advancing research in traffic state prediction.展开更多
Our concerns apply to the inadequate ways statistical distributions of crystallographic orientations are compared and occasionally confirmed to agree sufficiently well.The authors of“Machine learning enhanced analysi...Our concerns apply to the inadequate ways statistical distributions of crystallographic orientations are compared and occasionally confirmed to agree sufficiently well.The authors of“Machine learning enhanced analysis of EBSD data for texture representation”1 suggest a method to replace an EBSD dataset of crystallographic orientations with a much smaller synthetic dataset preserving the texture.They claim that their“texture adaptive clustering and sampling”algorithm generates datasets of a few hundred crystallographic orientations,realizing an equivalent crystallographic orientation distribution as the initial dataset.To prove the principle and substantiate their claim of equivalent orientation distributions,the authors content themselves with(i)a visual inspection of the crystallographic pole density function,in fact,of three crystallographic“pole figures”and(ii)Kolmogorov–Smirnov tests for each of the three Euler angles of the crystallographic orientations individually.However,these criteria are insufficient to confirm equivalence of orientation distributions,they do not provide scientific evidence to substantiate the authors’claim that“texture adaptive clustering and sampling”generates crystallographic orientations in terms of their Euler angles representing the same texture.展开更多
Big data analytic techniques associated with machine learning algorithms are playing an increasingly important role in various application fields,including stock market investment.However,few studies have focused on f...Big data analytic techniques associated with machine learning algorithms are playing an increasingly important role in various application fields,including stock market investment.However,few studies have focused on forecasting daily stock market returns,especially when using powerful machine learning techniques,such as deep neural networks(DNNs),to perform the analyses.DNNs employ various deep learning algorithms based on the combination of network structure,activation function,and model parameters,with their performance depending on the format of the data representation.This paper presents a comprehensive big data analytics process to predict the daily return direction of the SPDR S&P 500 ETF(ticker symbol:SPY)based on 60 financial and economic features.DNNs and traditional artificial neural networks(ANNs)are then deployed over the entire preprocessed but untransformed dataset,along with two datasets transformed via principal component analysis(PCA),to predict the daily direction of future stock market index returns.While controlling for overfitting,a pattern for the classification accuracy of the DNNs is detected and demonstrated as the number of the hidden layers increases gradually from 12 to 1000.Moreover,a set of hypothesis testing procedures are implemented on the classification,and the simulation results show that the DNNs using two PCA-represented datasets give significantly higher classification accuracy than those using the entire untransformed dataset,as well as several other hybrid machine learning algorithms.In addition,the trading strategies guided by the DNN classification process based on PCA-represented data perform slightly better than the others tested,including in a comparison against two standard benchmarks.展开更多
Iced transmission line galloping poses a significant threat to the safety and reliability of power systems,leading directly to line tripping,disconnections,and power outages.Existing early warning methods of iced tran...Iced transmission line galloping poses a significant threat to the safety and reliability of power systems,leading directly to line tripping,disconnections,and power outages.Existing early warning methods of iced transmission line galloping suffer from issues such as reliance on a single data source,neglect of irregular time series,and lack of attention-based closed-loop feedback,resulting in high rates of missed and false alarms.To address these challenges,we propose an Internet of Things(IoT)empowered early warning method of transmission line galloping that integrates time series data from optical fiber sensing and weather forecast.Initially,the method applies a primary adaptive weighted fusion to the IoT empowered optical fiber real-time sensing data and weather forecast data,followed by a secondary fusion based on a Back Propagation(BP)neural network,and uses the K-medoids algorithm for clustering the fused data.Furthermore,an adaptive irregular time series perception adjustment module is introduced into the traditional Gated Recurrent Unit(GRU)network,and closed-loop feedback based on attentionmechanism is employed to update network parameters through gradient feedback of the loss function,enabling closed-loop training and time series data prediction of the GRU network model.Subsequently,considering various types of prediction data and the duration of icing,an iced transmission line galloping risk coefficient is established,and warnings are categorized based on this coefficient.Finally,using an IoT-driven realistic dataset of iced transmission line galloping,the effectiveness of the proposed method is validated through multi-dimensional simulation scenarios.展开更多
Image clustering has received significant attention due to the growing importance of image recognition.Researchers have explored Riemannian manifold clustering,which is capable of capturing the non-linear shapes found...Image clustering has received significant attention due to the growing importance of image recognition.Researchers have explored Riemannian manifold clustering,which is capable of capturing the non-linear shapes found in real-world datasets.However,the complexity of image data poses substantial challenges for modelling and feature extraction.Traditional methods such as covariance matrices and linear subspace have shown promise in image modelling,and they are still in their early stages and suffer from certain limitations.However,these include the uncertainty of representing data using only one Riemannian manifold,limited feature extraction capacity of single kernel functions,and resulting incomplete data representation and redundancy.To overcome these limitations,the authors propose a novel approach called join multiple Riemannian manifold representation and multi-kernel non-redundancy for image clustering(MRMNR-MKC).It combines covariance matrices with linear subspace to represent data and applies multiple kernel functions to map the non-linear structural data into a reproducing kernel Hilbert space,enabling linear model analysis for image clustering.Additionally,the authors use matrix-induced regularisation to improve the clustering kernel selection process by reducing redundancy and assigning lower weights to identical kernels.Finally,the authors also conducted numerous experiments to evaluate the performance of our approach,confirming its superiority to state-of-the-art methods on three benchmark datasets.展开更多
With the rapid development of Global Positioning System(GPS),Global System for Mobile Communications(GSM),and the widespread application of mobile devices,a massive amount of trajectory data have been generated.Curren...With the rapid development of Global Positioning System(GPS),Global System for Mobile Communications(GSM),and the widespread application of mobile devices,a massive amount of trajectory data have been generated.Current trajectory data processing methods typically require input in the form of fixed-length vectors,making it crucial to convert variable-length trajectory data into fixed-length,low-dimensional embedding vectors.Trajectory representation learning aims to transform trajectory data into more expressive and interpretable representations.This paper provides a comprehensive review of the research progress,methodologies,and applications of trajectory representation learning.First,it categorizes and introduces the key techniques of trajectory representation learning and summarizes the available public trajectory datasets.Then,it classifies trajectory representation learning methods based on various downstream tasks,with a focus on their principles,advantages,limitations,and application scenarios in trajectory similarity computation,similar trajectory search,trajectory clustering,and trajectory prediction.Additionally,representative model structures and principles in each task are analyzed,along with the characteristics and advantages of different methods in each task.Last,the challenges faced by current trajectory representation learning methods are analyzed,including data sparsity,multimodality,model optimization,and privacy protection,while potential research directions and methodologies to address these challenges are explored.展开更多
Data-driven machine learning, especially deep learning technology, is becoming an important tool for handling big data issues in bioinformatics. In machine learning, DNA sequences are often converted to numerical valu...Data-driven machine learning, especially deep learning technology, is becoming an important tool for handling big data issues in bioinformatics. In machine learning, DNA sequences are often converted to numerical values for data representation and feature learning in various applications. Similar conversion occurs in Genomic Signal Processing(GSP), where genome sequences are transformed into numerical sequences for signal extraction and recognition. This kind of conversion is also called encoding scheme. The diverse encoding schemes can greatly affect the performance of GSP applications and machine learning models. This paper aims to collect,analyze, discuss, and summarize the existing encoding schemes of genome sequence particularly in GSP as well as other genome analysis applications to provide a comprehensive reference for the genomic data representation and feature learning in machine learning.展开更多
Deep learning algorithms increasingly support automated systems in areas such as human activity recognition and purchase recommendation.We identify a current trend in which data is transformed first into abstract visu...Deep learning algorithms increasingly support automated systems in areas such as human activity recognition and purchase recommendation.We identify a current trend in which data is transformed first into abstract visualizations and then processed by a computer vision deep learning pipeline.We call this VisuaLization As Intermediate Representation(VLAIR)and believe that it can be instrumental to support accurate recognition in a number of fields while also enhancing humans’ability to interpret deep learning models for debugging purposes or for personal use.In this paper we describe the potential advantages of this approach and explore various visualization mappings and deep learning architectures.We evaluate several VLAIR alternatives for a specific problem(human activity recognition in an apartment)and show that VLAIR attains classification accuracy above classical machine learning algorithms and several other non-image-based deep learning algorithms with several data representations.展开更多
representation that can identify and isolate different potential variables hidden in the highdimensional observations.Disentangled representation learning can capture information about a single change factor and contr...representation that can identify and isolate different potential variables hidden in the highdimensional observations.Disentangled representation learning can capture information about a single change factor and control it by the corresponding potential subspace,providing a robust representation for complex changes in the data.In this paper,we first introduce and analyze the current status of research on disentangled representation and its causal mechanisms and summarize three crucial properties of disentangled representation.Then,disentangled representation learning algorithms are classified into four categories and outlined in terms of both mathematical description and applicability.Subsequently,the loss functions and objective evaluation metrics commonly used in existing work on disentangled representation are classified.Finally,the paper summarizes representative applications of disentangled representation learning in the field of remote sensing and discusses its future development.展开更多
基金sponsored by the National Natural Science Foundation of China(Grant No.52178100).
文摘The spatial offset of bridge has a significant impact on the safety,comfort,and durability of high-speed railway(HSR)operations,so it is crucial to rapidly and effectively detect the spatial offset of operational HSR bridges.Drive-by monitoring of bridge uneven settlement demonstrates significant potential due to its practicality,cost-effectiveness,and efficiency.However,existing drive-by methods for detecting bridge offset have limitations such as reliance on a single data source,low detection accuracy,and the inability to identify lateral deformations of bridges.This paper proposes a novel drive-by inspection method for spatial offset of HSR bridge based on multi-source data fusion of comprehensive inspection train.Firstly,dung beetle optimizer-variational mode decomposition was employed to achieve adaptive decomposition of non-stationary dynamic signals,and explore the hidden temporal relationships in the data.Subsequently,a long short-term memory neural network was developed to achieve feature fusion of multi-source signal and accurate prediction of spatial settlement of HSR bridge.A dataset of track irregularities and CRH380A high-speed train responses was generated using a 3D train-track-bridge interaction model,and the accuracy and effectiveness of the proposed hybrid deep learning model were numerically validated.Finally,the reliability of the proposed drive-by inspection method was further validated by analyzing the actual measurement data obtained from comprehensive inspection train.The research findings indicate that the proposed approach enables rapid and accurate detection of spatial offset in HSR bridge,ensuring the long-term operational safety of HSR bridges.
基金supported by the National Natural Science Foundation of China(Grant No.:U2202213)the Special Program for the Major Science and Technology Projects of Yunnan Province,China(Grant Nos.:202102AE090051-1-01,and 202202AE090001).
文摘In traditional medicine and ethnomedicine,medicinal plants have long been recognized as the basis for materials in therapeutic applications worldwide.In particular,the remarkable curative effect of traditional Chinese medicine during corona virus disease 2019(COVID-19)pandemic has attracted extensive attention globally.Medicinal plants have,therefore,become increasingly popular among the public.However,with increasing demand for and profit with medicinal plants,commercial fraudulent events such as adulteration or counterfeits sometimes occur,which poses a serious threat to the clinical outcomes and interests of consumers.With rapid advances in artificial intelligence,machine learning can be used to mine information on various medicinal plants to establish an ideal resource database.We herein present a review that mainly introduces common machine learning algorithms and discusses their application in multi-source data analysis of medicinal plants.The combination of machine learning algorithms and multi-source data analysis facilitates a comprehensive analysis and aids in the effective evaluation of the quality of medicinal plants.The findings of this review provide new possibilities for promoting the development and utilization of medicinal plants.
基金supported by the SungKyunKwan University and the BK21 FOUR(Graduate School Innovation)funded by the Ministry of Education(MOE,Korea)and National Research Foundation of Korea(NRF).
文摘With the recent increase in data volume and diversity,traditional text representation techniques are struggling to capture context,particularly in environments with sparse data.To address these challenges,this study proposes a new model,the Masked Joint Representation Model(MJRM).MJRM approximates the original hypothesis by leveraging multiple elements in a limited context.It dynamically adapts to changes in characteristics based on data distribution through three main components.First,masking-based representation learning,termed selective dynamic masking,integrates topic modeling and sentiment clustering to generate and train multiple instances across different data subsets,whose predictions are then aggregated with optimized weights.This design alleviates sparsity,suppresses noise,and preserves contextual structures.Second,regularization-based improvements are applied.Third,techniques for addressing sparse data are used to perform final inference.As a result,MJRM improves performance by up to 4%compared to existing AI techniques.In our experiments,we analyzed the contribution of each factor,demonstrating that masking,dynamic learning,and aggregating multiple instances complement each other to improve performance.This demonstrates that a masking-based multi-learning strategy is effective for context-aware sparse text classification,and can be useful even in challenging situations such as data shortage or data distribution variations.We expect that the approach can be extended to diverse fields such as sentiment analysis,spam filtering,and domain-specific document classification.
基金supported by grants from the National Key Research and Development Program of China(2022YFD2001103)the National Natural Science Foundation of China(42371373)。
文摘Near real-time maize phenology monitoring is crucial for field management,cropping system adjustments,and yield estimation.Most phenological monitoring methods are post-seasonal and heavily rely on high-frequency time-series data.These methods are not applicable on the unmanned aerial vehicle(UAV)platform due to the high cost of acquiring time-series UAV images and the shortage of UAV-based phenological monitoring methods.To address these challenges,we employed the Synthetic Minority Oversampling Technique(SMOTE)for sample augmentation,aiming to resolve the small sample modelling problem.Moreover,we utilized enhanced"separation"and"compactness"feature selection methods to identify input features from multiple data sources.In this process,we incorporated dynamic multi-source data fusion strategies,involving Vegetation index(VI),Color index(CI),and Texture features(TF).A two-stage neural network that combines Convolutional Neural Network(CNN)and Long Short-Term Memory Network(LSTM)is proposed to identify maize phenological stages(including sowing,seedling,jointing,trumpet,tasseling,maturity,and harvesting)on UAV platforms.The results indicate that the dataset generated by SMOTE closely resembles the measured dataset.Among dynamic data fusion strategies,the VI-TF combination proves to be most effective,with CI-TF and VI-CI combinations following behind.Notably,as more data sources are integrated,the model's demand for input features experiences a significant decline.In particular,the CNN-LSTM model,based on the fusion of three data sources,exhibited remarkable reliability when validating the three datasets.For Dataset 1(Beijing Xiaotangshan,2023:Data from 12 UAV Flight Missions),the model achieved an overall accuracy(OA)of 86.53%.Additionally,its precision(Pre),recall(Rec),F1 score(F1),false acceptance rate(FAR),and false rejection rate(FRR)were 0.89,0.89,0.87,0.11,and 0.11,respectively.The model also showed strong generalizability in Dataset 2(Beijing Xiaotangshan,2023:Data from 6 UAV Flight Missions)and Dataset 3(Beijing Xiaotangshan,2022:Data from 4 UAV Flight Missions),with OAs of 89.4%and 85%,respectively.Meanwhile,the model has a low demand for input featu res,requiring only 54.55%(99 of all featu res).The findings of this study not only offer novel insights into near real-time crop phenology monitoring,but also provide technical support for agricultural field management and cropping system adaptation.
基金supported by the Sichuan Science and Technology Program(Nos.2024JDRC0100 and 2023YFQ0091)the National Natural Science Foundation of China(Nos.U21A20167 and 52475138)the Scientific Research Foundation of the State Key Laboratory of Rail Transit Vehicle System(No.2024RVL-T08).
文摘Accurate monitoring of track irregularities is very helpful to improving the vehicle operation quality and to formulating appropriate track maintenance strategies.Existing methods have the problem that they rely on complex signal processing algorithms and lack multi-source data analysis.Driven by multi-source measurement data,including the axle box,the bogie frame and the carbody accelerations,this paper proposes a track irregularities monitoring network(TIMNet)based on deep learning methods.TIMNet uses the feature extraction capability of convolutional neural networks and the sequence map-ping capability of the long short-term memory model to explore the mapping relationship between vehicle accelerations and track irregularities.The particle swarm optimization algorithm is used to optimize the network parameters,so that both the vertical and lateral track irregularities can be accurately identified in the time and spatial domains.The effectiveness and superiority of the proposed TIMNet is analyzed under different simulation conditions using a vehicle dynamics model.Field tests are conducted to prove the availability of the proposed TIMNet in quantitatively monitoring vertical and lateral track irregularities.Furthermore,comparative tests show that the TIMNet has a better fitting degree and timeliness in monitoring track irregularities(vertical R2 of 0.91,lateral R2 of 0.84 and time cost of 10 ms),compared to other classical regression.The test also proves that the TIMNet has a better anti-interference ability than other regression models.
基金The National Natural Science Foundation of China(No.61762031)The Science and Technology Major Project of Guangxi Province(NO.AA19046004)The Natural Science Foundation of Guangxi(No.2021JJA170130).
文摘There is a large amount of information in the network data that we canexploit. It is difficult for classical community detection algorithms to handle network data with sparse topology. Representation learning of network data is usually paired with clustering algorithms to solve the community detection problem.Meanwhile, there is always an unpredictable distribution of class clusters outputby graph representation learning. Therefore, we propose an improved densitypeak clustering algorithm (ILDPC) for the community detection problem, whichimproves the local density mechanism in the original algorithm and can betteraccommodate class clusters of different shapes. And we study the communitydetection in network data. The algorithm is paired with the benchmark modelGraph sample and aggregate (GraphSAGE) to show the adaptability of ILDPCfor community detection. The plotted decision diagram shows that the ILDPCalgorithm is more discriminative in selecting density peak points compared tothe original algorithm. Finally, the performance of K-means and other clusteringalgorithms on this benchmark model is compared, and the algorithm is proved tobe more suitable for community detection in sparse networks with the benchmarkmodel on the evaluation criterion F1-score. The sensitivity of the parameters ofthe ILDPC algorithm to the low-dimensional vector set output by the benchmarkmodel GraphSAGE is also analyzed.
基金financially supported the Strategic Priority Research Program of the Chinese Academy of Sciences (No. XDA14020102)the National Natural Science Foundation of China (Nos. 41774125,41530320 and 41804098)the Key National Research Project of China (Nos. 2016YFC0303100,2017YFC0601900)。
文摘Time-domain airborne electromagnetic(AEM)data are frequently subject to interference from various types of noise,which can reduce the data quality and affect data inversion and interpretation.Traditional denoising methods primarily deal with data directly,without analyzing the data in detail;thus,the results are not always satisfactory.In this paper,we propose a method based on dictionary learning for EM data denoising.This method uses dictionary learning to perform feature analysis and to extract and reconstruct the true signal.In the process of dictionary learning,the random noise is fi ltered out as residuals.To verify the eff ectiveness of this dictionary learning approach for denoising,we use a fi xed overcomplete discrete cosine transform(ODCT)dictionary algorithm,the method-of-optimal-directions(MOD)dictionary learning algorithm,and the K-singular value decomposition(K-SVD)dictionary learning algorithm to denoise decay curves at single points and to denoise profi le data for diff erent time channels in time-domain AEM.The results show obvious diff erences among the three dictionaries for denoising AEM data,with the K-SVD dictionary achieving the best performance.
基金supported by the National Natural Science Foundation of China under Grant 51722406,52074340,and 51874335the Shandong Provincial Natural Science Foundation under Grant JQ201808+5 种基金The Fundamental Research Funds for the Central Universities under Grant 18CX02097Athe Major Scientific and Technological Projects of CNPC under Grant ZD2019-183-008the Science and Technology Support Plan for Youth Innovation of University in Shandong Province under Grant 2019KJH002the National Research Council of Science and Technology Major Project of China under Grant 2016ZX05025001-006111 Project under Grant B08028Sinopec Science and Technology Project under Grant P20050-1
文摘For reservoirs with complex non-Gaussian geological characteristics,such as carbonate reservoirs or reservoirs with sedimentary facies distribution,it is difficult to implement history matching directly,especially for the ensemble-based data assimilation methods.In this paper,we propose a multi-source information fused generative adversarial network(MSIGAN)model,which is used for parameterization of the complex geologies.In MSIGAN,various information such as facies distribution,microseismic,and inter-well connectivity,can be integrated to learn the geological features.And two major generative models in deep learning,variational autoencoder(VAE)and generative adversarial network(GAN)are combined in our model.Then the proposed MSIGAN model is integrated into the ensemble smoother with multiple data assimilation(ESMDA)method to conduct history matching.We tested the proposed method on two reservoir models with fluvial facies.The experimental results show that the proposed MSIGAN model can effectively learn the complex geological features,which can promote the accuracy of history matching.
基金funding for this work was provided by the Center for Extreme Events in Structurally Evolving Material under U.S.Army CCAPG-RTPDivision contract numberW011NF-23-2-0073The funder played no role in the study design,data collection,analysis,and interpretation of data,or the writing of this manuscript.
文摘We respond to Schaeben et al.’s1 comment on our paper,“Machine Learning Enhanced Analysis of EBSD Data for Texture Representation.”While their observations are factually correct,they do not disprove our results.Our method,TACS,preserves the full distribution of crystallographic orientations and is validated with real-world data.We emphasize the importance of empirical validation over theoretical constructs in assessing machine learning methods’practical performance.
基金supported by the National Science and Technology Major Project of China(No.2022ZD0120003)。
文摘Deep learning has empowered traffic prediction models to integrate diverse auxiliary data sources,such as weather and temporal features,for enhanced forecasting accuracy.However,existing approaches often suffer from limited generality and scalability,and the field lacks a unified benchmark for fair model comparison.This absence hinders consistent performance evaluation,slows the development of robust and adaptable models,and makes it challenging to quantify the incremental benefits of different auxiliary data sources.To address these issues,we present MltAuxTSPP,a unified benchmark framework for deep learning-based traffic state prediction with multi-source auxiliary data.The framework features a standardized data container and a fusion embedding module,enabling consistent utilization of heterogeneous data and improving scalability.It produces unified hidden representations that can be seamlessly adopted by various downstream models,ensuring fair and reproducible comparisons under identical conditions.Extensive experiments on real-world datasets demonstrate that MltAuxTSPP effectively leverages weather and temporal features to improve long-term forecast performance and offers a practical and reproducible foundation for advancing research in traffic state prediction.
文摘Our concerns apply to the inadequate ways statistical distributions of crystallographic orientations are compared and occasionally confirmed to agree sufficiently well.The authors of“Machine learning enhanced analysis of EBSD data for texture representation”1 suggest a method to replace an EBSD dataset of crystallographic orientations with a much smaller synthetic dataset preserving the texture.They claim that their“texture adaptive clustering and sampling”algorithm generates datasets of a few hundred crystallographic orientations,realizing an equivalent crystallographic orientation distribution as the initial dataset.To prove the principle and substantiate their claim of equivalent orientation distributions,the authors content themselves with(i)a visual inspection of the crystallographic pole density function,in fact,of three crystallographic“pole figures”and(ii)Kolmogorov–Smirnov tests for each of the three Euler angles of the crystallographic orientations individually.However,these criteria are insufficient to confirm equivalence of orientation distributions,they do not provide scientific evidence to substantiate the authors’claim that“texture adaptive clustering and sampling”generates crystallographic orientations in terms of their Euler angles representing the same texture.
文摘Big data analytic techniques associated with machine learning algorithms are playing an increasingly important role in various application fields,including stock market investment.However,few studies have focused on forecasting daily stock market returns,especially when using powerful machine learning techniques,such as deep neural networks(DNNs),to perform the analyses.DNNs employ various deep learning algorithms based on the combination of network structure,activation function,and model parameters,with their performance depending on the format of the data representation.This paper presents a comprehensive big data analytics process to predict the daily return direction of the SPDR S&P 500 ETF(ticker symbol:SPY)based on 60 financial and economic features.DNNs and traditional artificial neural networks(ANNs)are then deployed over the entire preprocessed but untransformed dataset,along with two datasets transformed via principal component analysis(PCA),to predict the daily direction of future stock market index returns.While controlling for overfitting,a pattern for the classification accuracy of the DNNs is detected and demonstrated as the number of the hidden layers increases gradually from 12 to 1000.Moreover,a set of hypothesis testing procedures are implemented on the classification,and the simulation results show that the DNNs using two PCA-represented datasets give significantly higher classification accuracy than those using the entire untransformed dataset,as well as several other hybrid machine learning algorithms.In addition,the trading strategies guided by the DNN classification process based on PCA-represented data perform slightly better than the others tested,including in a comparison against two standard benchmarks.
基金research was funded by Science and Technology Project of State Grid Corporation of China under grant number 5200-202319382A-2-3-XG.
文摘Iced transmission line galloping poses a significant threat to the safety and reliability of power systems,leading directly to line tripping,disconnections,and power outages.Existing early warning methods of iced transmission line galloping suffer from issues such as reliance on a single data source,neglect of irregular time series,and lack of attention-based closed-loop feedback,resulting in high rates of missed and false alarms.To address these challenges,we propose an Internet of Things(IoT)empowered early warning method of transmission line galloping that integrates time series data from optical fiber sensing and weather forecast.Initially,the method applies a primary adaptive weighted fusion to the IoT empowered optical fiber real-time sensing data and weather forecast data,followed by a secondary fusion based on a Back Propagation(BP)neural network,and uses the K-medoids algorithm for clustering the fused data.Furthermore,an adaptive irregular time series perception adjustment module is introduced into the traditional Gated Recurrent Unit(GRU)network,and closed-loop feedback based on attentionmechanism is employed to update network parameters through gradient feedback of the loss function,enabling closed-loop training and time series data prediction of the GRU network model.Subsequently,considering various types of prediction data and the duration of icing,an iced transmission line galloping risk coefficient is established,and warnings are categorized based on this coefficient.Finally,using an IoT-driven realistic dataset of iced transmission line galloping,the effectiveness of the proposed method is validated through multi-dimensional simulation scenarios.
基金National Natural Science Foundation of China(62273290,62072391,and 61572419).
文摘Image clustering has received significant attention due to the growing importance of image recognition.Researchers have explored Riemannian manifold clustering,which is capable of capturing the non-linear shapes found in real-world datasets.However,the complexity of image data poses substantial challenges for modelling and feature extraction.Traditional methods such as covariance matrices and linear subspace have shown promise in image modelling,and they are still in their early stages and suffer from certain limitations.However,these include the uncertainty of representing data using only one Riemannian manifold,limited feature extraction capacity of single kernel functions,and resulting incomplete data representation and redundancy.To overcome these limitations,the authors propose a novel approach called join multiple Riemannian manifold representation and multi-kernel non-redundancy for image clustering(MRMNR-MKC).It combines covariance matrices with linear subspace to represent data and applies multiple kernel functions to map the non-linear structural data into a reproducing kernel Hilbert space,enabling linear model analysis for image clustering.Additionally,the authors use matrix-induced regularisation to improve the clustering kernel selection process by reducing redundancy and assigning lower weights to identical kernels.Finally,the authors also conducted numerous experiments to evaluate the performance of our approach,confirming its superiority to state-of-the-art methods on three benchmark datasets.
基金supported by the National Natural Science Foundation of China(Grant No.61772249).
文摘With the rapid development of Global Positioning System(GPS),Global System for Mobile Communications(GSM),and the widespread application of mobile devices,a massive amount of trajectory data have been generated.Current trajectory data processing methods typically require input in the form of fixed-length vectors,making it crucial to convert variable-length trajectory data into fixed-length,low-dimensional embedding vectors.Trajectory representation learning aims to transform trajectory data into more expressive and interpretable representations.This paper provides a comprehensive review of the research progress,methodologies,and applications of trajectory representation learning.First,it categorizes and introduces the key techniques of trajectory representation learning and summarizes the available public trajectory datasets.Then,it classifies trajectory representation learning methods based on various downstream tasks,with a focus on their principles,advantages,limitations,and application scenarios in trajectory similarity computation,similar trajectory search,trajectory clustering,and trajectory prediction.Additionally,representative model structures and principles in each task are analyzed,along with the characteristics and advantages of different methods in each task.Last,the challenges faced by current trajectory representation learning methods are analyzed,including data sparsity,multimodality,model optimization,and privacy protection,while potential research directions and methodologies to address these challenges are explored.
基金supports from the Department of Computing Sciences, State University of New York College at Brockport
文摘Data-driven machine learning, especially deep learning technology, is becoming an important tool for handling big data issues in bioinformatics. In machine learning, DNA sequences are often converted to numerical values for data representation and feature learning in various applications. Similar conversion occurs in Genomic Signal Processing(GSP), where genome sequences are transformed into numerical sequences for signal extraction and recognition. This kind of conversion is also called encoding scheme. The diverse encoding schemes can greatly affect the performance of GSP applications and machine learning models. This paper aims to collect,analyze, discuss, and summarize the existing encoding schemes of genome sequence particularly in GSP as well as other genome analysis applications to provide a comprehensive reference for the genomic data representation and feature learning in machine learning.
文摘Deep learning algorithms increasingly support automated systems in areas such as human activity recognition and purchase recommendation.We identify a current trend in which data is transformed first into abstract visualizations and then processed by a computer vision deep learning pipeline.We call this VisuaLization As Intermediate Representation(VLAIR)and believe that it can be instrumental to support accurate recognition in a number of fields while also enhancing humans’ability to interpret deep learning models for debugging purposes or for personal use.In this paper we describe the potential advantages of this approach and explore various visualization mappings and deep learning architectures.We evaluate several VLAIR alternatives for a specific problem(human activity recognition in an apartment)and show that VLAIR attains classification accuracy above classical machine learning algorithms and several other non-image-based deep learning algorithms with several data representations.
基金supported by the National Natural Science Foundation of China(Nos.61825103,62202349)the Natural Science Foundation of Hubei Province(Nos.2022CFB352,2020CFA001)the Key Research&Development of Hubei Province(No.2020BIB006).
文摘representation that can identify and isolate different potential variables hidden in the highdimensional observations.Disentangled representation learning can capture information about a single change factor and control it by the corresponding potential subspace,providing a robust representation for complex changes in the data.In this paper,we first introduce and analyze the current status of research on disentangled representation and its causal mechanisms and summarize three crucial properties of disentangled representation.Then,disentangled representation learning algorithms are classified into four categories and outlined in terms of both mathematical description and applicability.Subsequently,the loss functions and objective evaluation metrics commonly used in existing work on disentangled representation are classified.Finally,the paper summarizes representative applications of disentangled representation learning in the field of remote sensing and discusses its future development.