Attempts have been made to modulate motor sequence learning(MSL)through repetitive transcranial magnetic stimulation,targeting different sites within the sensorimotor network.However,the target with the optimum modula...Attempts have been made to modulate motor sequence learning(MSL)through repetitive transcranial magnetic stimulation,targeting different sites within the sensorimotor network.However,the target with the optimum modulatory effect on neural plasticity associated with MSL remains unclarified.This study was therefore designed to compare the role of the left primary motor cortex and the left supplementary motor area proper(SMAp)in modulating MSL across different complexity levels and for both hands,as well as the associated neuroplasticity by applying intermittent theta burst stimulation together with the electroencephalogram and concurrent transcranial magnetic stimulation.Our data demonstrated the role of SMAp stimulation in modulating neural communication to support MSL,which is achieved by facilitating regional activation and orchestrating neural coupling across distributed brain regions,particularly in interhemispheric connections.These findings may have important clinical implications,particularly for motor rehabilitation in populations such as post-stroke patients.展开更多
Machine learning(ML)efficiently and accurately processes dense seismic array data,improving earthquake catalog creation,which is crucial for understanding earthquake sequences and fault systems;analyzing its reliabili...Machine learning(ML)efficiently and accurately processes dense seismic array data,improving earthquake catalog creation,which is crucial for understanding earthquake sequences and fault systems;analyzing its reliability is also essential.An M5.8 earthquake struck Alxa Left Banner,Inner Mongolia,China on April 15,2015,a region with limited CENC monitoring capabilities,making analysis challenging.However,abundant data from ChinArray provided valuable observations for assessing the event.This study leveraged ChinArray data from the 2015 Alxa Left Banner earthquake sequence,employing machine learning(specifically PhaseNet,a deep learning method,and GaMMA,a Bayesian approach)for automated seismic phase picking,association,and location analysis.Our generated catalog,comprising 10,432 phases from 708 events,is roughly ten times larger than the CENC catalog,encompassing all CENC events with strong consistency.A slight magnitude overestimation is observed only at lower magnitudes.Furthermore,the catalog adheres to the Gutenberg-Richter and Omori laws spatially,temporally,and in magnitude distribution,demonstrating its high reliability.Double-difference tomography refined locations for 366 events,yielding a more compact spatial distribution with horizontal errors within 100m,vertical errors within 300m,and travel-time residuals within 0.05s.Depths predominantly range from 10-30km.Aftershocks align primarily NEE,with the mainshock east of the aftershock zone.The near-vertical main fault plane dips northwestward,exhibiting a Y-shaped branching structure,converging at depth and expanding towards the surface.FOCMEC analysis,using first motion and amplitude ratios,yielded focal mechanism solutions for 10 events,including the mainshock.These solutions consistently indicate a strike-slip mechanism with a minor extensional component.Integrating the earthquake sequence's spatial distribution and focal mechanisms suggests the seismogenic structure is a negative flower structure,consistent with the Dengkou-Benjing fault.Comparing the CENC and ML-generated catalogs using the maximum curvature(MAXC)method reveals a 0.6 decrease in completeness magnitude(M_(C)).However,magnitude-frequency distribution discrepancies above the MAXC-estimated M_(C)suggest MAXC may underestimate both M_(C)and the b-value.This study analyzes the 2015 Alxa Left Banner M5.8 earthquake using a reliable,MLgenerated earthquake catalog,revealing detailed information about the sequence,faulting structure,aftershock distribution,and stress characteristics.展开更多
Rice(Oryza sativa)is a staple food for more than half of the world's population and a critical crop for global agriculture.Understanding the regulatory mechanisms that control gene expression in the rice genome is...Rice(Oryza sativa)is a staple food for more than half of the world's population and a critical crop for global agriculture.Understanding the regulatory mechanisms that control gene expression in the rice genome is fundamental for advancing agricultural productivity and food security.In mechanism,cis-regulatory elements(including promoters,enhancers,silencers,and insulators)are key DNA sequences whose activities determine the spatial and temporal expression patterns of nearby genes(Yocca and Edger,2022;Schmitz et al.,2022).展开更多
Objective To establish and validate a novel diabetic retinopathy(DR)risk-prediction model using a whole-exome sequencing(WES)-based machine learning(ML)method.Methods WES was performed to identify potential single nuc...Objective To establish and validate a novel diabetic retinopathy(DR)risk-prediction model using a whole-exome sequencing(WES)-based machine learning(ML)method.Methods WES was performed to identify potential single nucleotide polymorphism(SNP)or mutation sites in a DR pedigree comprising 10 members.A prediction model was established and validated in a cohort of 420 type 2 diabetic patients based on both genetic and demographic features.The contribution of each feature was assessed using Shapley Additive explanation analysis.The efficacies of the models with and without SNP were compared.Results WES revealed that seven SNPs/mutations(rs116911833 in TRIM7,1997T>C in LRBA,1643T>C in PRMT10,rs117858678 in C9orf152,rs201922794 in CLDN25,rs146694895 in SH3GLB2,and rs201407189 in FANCC)were associated with DR.Notably,the model including rs146694895 and rs201407189 achieved better performance in predicting DR(accuracy:80.2%;sensitivity:83.3%;specificity:76.7%;area under the receiver operating characteristic curve[AUC]:80.0%)than the model without these SNPs(accuracy:79.4%;sensitivity:80.3%;specificity:78.3%;AUC:79.3%).Conclusion Novel SNP sites associated with DR were identified in the DR pedigree.Inclusion of rs146694895 and rs201407189 significantly enhanced the performance of the ML-based DR prediction model.展开更多
Computer-aided design(CAD)software continues to be a crucial tool in digital twin application and manufacturing,facilitating the design of various products.We present a novel CAD generation method,an agent that constr...Computer-aided design(CAD)software continues to be a crucial tool in digital twin application and manufacturing,facilitating the design of various products.We present a novel CAD generation method,an agent that constructs the CAD sequences containing the sketch-and-extrude modelling operations efficiently and with high quality.Starting from the sketch and extrusion operation sequences,we utilise the transformer encoder to encode them into different disentangled codebooks to represent their distribution properties while considering their correlations.Then,a combination of auto-regressive and non-autoregressive samplers is trained to sample the code for CAD sequence con-struction.Extensive experiments demonstrate that our model generates diverse and high-quality CAD models.We also show some cases of real digital twin applications and indicate that our generated model can be used as the data source for the digital twin platform,exhibiting designers'potential.展开更多
The use of dynamic programming(DP)algorithms to learn Bayesian network structures is limited by their high space complexity and difficulty in learning the structure of large-scale networks.Therefore,this study propose...The use of dynamic programming(DP)algorithms to learn Bayesian network structures is limited by their high space complexity and difficulty in learning the structure of large-scale networks.Therefore,this study proposes a DP algorithm based on node block sequence constraints.The proposed algorithm constrains the traversal process of the parent graph by using the M-sequence matrix to considerably reduce the time consumption and space complexity by pruning the traversal process of the order graph using the node block sequence.Experimental results show that compared with existing DP algorithms,the proposed algorithm can obtain learning results more efficiently with less than 1%loss of accuracy,and can be used for learning larger-scale networks.展开更多
Inferring phylogenetic trees from molecular sequences is a cornerstone of evolutionary biology.Many standard phylogenetic methods(such as maximum-likelihood[ML])rely on explicit models of sequence evolution and thus o...Inferring phylogenetic trees from molecular sequences is a cornerstone of evolutionary biology.Many standard phylogenetic methods(such as maximum-likelihood[ML])rely on explicit models of sequence evolution and thus often suffer from model misspecification or inadequacy.The on-rising deep learning(DL)techniques offer a powerful alternative.Deep learning employs multi-layered artificial neural networks to progressively transform input data into more abstract and complex representations.DL methods can autonomously uncover meaningful patterns from data,thereby bypassing potential biases introduced by predefined features(Franklin,2005;Murphy,2012).Recent efforts have aimed to apply deep neural networks(DNNs)to phylogenetics,with a growing number of applications in tree reconstruction(Suvorov et al.,2020;Zou et al.,2020;Nesterenko et al.,2022;Smith and Hahn,2023;Wang et al.,2023),substitution model selection(Abadi et al.,2020;Burgstaller-Muehlbacher et al.,2023),and diversification rate inference(Voznica et al.,2022;Lajaaiti et al.,2023;Lambert et al.,2023).In phylogenetic tree reconstruction,PhyDL(Zou et al.,2020)and Tree_learning(Suvorov et al.,2020)are two notable DNN-based programs designed to infer unrooted quartet trees directly from alignments of four amino acid(AA)and DNA sequences,respectively.展开更多
Many bioinformatics applications require determining the class of a newly sequenced Deoxyribonucleic acid(DNA)sequence,making DNA sequence classification an integral step in performing bioinformatics analysis,where la...Many bioinformatics applications require determining the class of a newly sequenced Deoxyribonucleic acid(DNA)sequence,making DNA sequence classification an integral step in performing bioinformatics analysis,where large biomedical datasets are transformed into valuable knowledge.Existing methods rely on a feature extraction step and suffer from high computational time requirements.In contrast,newer approaches leveraging deep learning have shown significant promise in enhancing accuracy and efficiency.In this paper,we investigate the performance of various deep learning architectures:Convolutional Neural Network(CNN),CNN-Long Short-Term Memory(CNNLSTM),CNN-Bidirectional Long Short-Term Memory(CNN-BiLSTM),Residual Network(ResNet),and InceptionV3 for DNA sequence classification.Various numerical and visual data representation techniques are utilized to represent the input datasets,including:label encoding,k-mer sentence encoding,k-mer one-hot vector,Frequency Chaos Game Representation(FCGR)and 5-Color Map(ColorSquare).Three datasets are used for the training of the models including H3,H4 and DNA Sequence Dataset(Yeast,Human,Arabidopsis Thaliana).Experiments are performed to determine which combination of DNA representation and deep learning architecture yields improved performance for the classification task.Our results indicate that using a hybrid CNN-LSTM neural network trained on DNA sequences represented as one-hot encoded k-mer sequences yields the best performance,achieving an accuracy of 92.1%.展开更多
Background Breed identification is useful in a variety of biological contexts.Breed identification usually involves two stages,i.e.,detection of breed-informative SNPs and breed assignment.For both stages,there are se...Background Breed identification is useful in a variety of biological contexts.Breed identification usually involves two stages,i.e.,detection of breed-informative SNPs and breed assignment.For both stages,there are several methods proposed.However,what is the optimal combination of these methods remain unclear.In this study,using the whole genome sequence data available for 13 cattle breeds from Run 8 of the 1,000 Bull Genomes Project,we compared the combinations of three methods(Delta,FST,and In)for breed-informative SNP detection and five machine learning methods(KNN,SVM,RF,NB,and ANN)for breed assignment with respect to different reference population sizes and difference numbers of most breed-informative SNPs.In addition,we evaluated the accuracy of breed identification using SNP chip data of different densities.Results We found that all combinations performed quite well with identification accuracies over 95%in all scenarios.However,there was no combination which performed the best and robust across all scenarios.We proposed to inte-grate the three breed-informative detection methods,named DFI,and integrate the three machine learning methods,KNN,SVM,and RF,named KSR.We found that the combination of these two integrated methods outperformed the other combinations with accuracies over 99%in most cases and was very robust in all scenarios.The accuracies from using SNP chip data were only slightly lower than that from using sequence data in most cases.Conclusions The current study showed that the combination of DFI and KSR was the optimal strategy.Using sequence data resulted in higher accuracies than using chip data in most cases.However,the differences were gener-ally small.In view of the cost of genotyping,using chip data is also a good option for breed identification.展开更多
As an essential function of encrypted Internet traffic analysis,encrypted traffic service classification can support both coarse-grained network service traffic management and security supervision.However,the traditio...As an essential function of encrypted Internet traffic analysis,encrypted traffic service classification can support both coarse-grained network service traffic management and security supervision.However,the traditional plaintext-based Deep Packet Inspection(DPI)method cannot be applied to such a classification.Moreover,machine learning-based existing methods encounter two problems during feature selection:complex feature overcost processing and Transport Layer Security(TLS)version discrepancy.In this paper,we consider differences between encryption network protocol stacks and propose a composite deep learning-based method in multiprotocol environments using a sliding multiple Protocol Data Unit(multiPDU)length sequence as features by fully utilizing the Markov property in a multiPDU length sequence and maintaining suitability with a TLS-1.3 environment.Control experiments show that both Length-Sensitive(LS)composite deep learning model using a capsule neural network and LS-long short time memory achieve satisfactory effectiveness in F1-score and performance.Owing to faster feature extraction,our method is suitable for actual network environments and superior to state-of-the-art methods.展开更多
An idea of estimating the direct sequence spread spectrum(DSSS) signal pseudo-noise(PN) sequence is presented. Without the apriority knowledge about the DSSS signal in the non-cooperation condition, we propose a s...An idea of estimating the direct sequence spread spectrum(DSSS) signal pseudo-noise(PN) sequence is presented. Without the apriority knowledge about the DSSS signal in the non-cooperation condition, we propose a self-organizing feature map(SOFM) neural network algorithm to detect and identify the PN sequence. A non-supervised learning algorithm is proposed according the Kohonen rule in SOFM. The blind algorithm can also estimate the PN sequence in a low signal-to-noise(SNR) and computer simulation demonstrates that the algorithm is effective. Compared with the traditional correlation algorithm based on slip-correlation, the proposed algorithm's bit error rate(BER) and complexity are lower.展开更多
Production performance prediction of tight gas reservoirs is crucial to the estimation of ultimate recovery,which has an important impact on gas field development planning and economic evaluation.Owing to the model’s...Production performance prediction of tight gas reservoirs is crucial to the estimation of ultimate recovery,which has an important impact on gas field development planning and economic evaluation.Owing to the model’s simplicity,the decline curve analysis method has been widely used to predict production performance.The advancement of deep-learning methods provides an intelligent way of analyzing production performance in tight gas reservoirs.In this paper,a sequence learning method to improve the accuracy and efficiency of tight gas production forecasting is proposed.The sequence learning methods used in production performance analysis herein include the recurrent neural network(RNN),long short-term memory(LSTM)neural network,and gated recurrent unit(GRU)neural network,and their performance in the tight gas reservoir production prediction is investigated and compared.To further improve the performance of the sequence learning method,the hyperparameters in the sequence learning methods are optimized through a particle swarm optimization algorithm,which can greatly simplify the optimization process of the neural network model in an automated manner.Results show that the optimized GRU and RNN models have more compact neural network structures than the LSTM model and that the GRU is more efficiently trained.The predictive performance of LSTM and GRU is similar,and both are better than the RNN and the decline curve analysis model and thus can be used to predict tight gas production.展开更多
Aortic dissection(AD)is one of the most serious diseases with high mortality,and its diagnosis mainly depends on computed tomography(CT)results.Most existing automatic diagnosis methods of AD are only suitable for AD ...Aortic dissection(AD)is one of the most serious diseases with high mortality,and its diagnosis mainly depends on computed tomography(CT)results.Most existing automatic diagnosis methods of AD are only suitable for AD recognition,which usually require preselection of CT images and cannot be further classified to different types.In this work,we constructed a dataset of 105 cases with a total of 49021 slices,including 31043 slices expertlevel annotation and proposed a two-stage AD diagnosis structure based on sequence information and deep learning.The proposed region of interest(RoI)extraction algorithm based on sequence information(RESI)can realize high-precision for RoI identification in the first stage.Then DenseNet-121 is applied for further diagnosis.Specially,the proposed method can judge the type of AD without preselection of CT images.The experimental results show that the accuracy of Stanford typing classification of AD is 89.19%,and the accuracy at the slice-level reaches 97.41%,which outperform the state-ofart methods.It can provide important decision-making information for the determination of further surgical treatment plan for patients.展开更多
With the fast development of business logic and information technology, today's best solutions are tomorrow's legacy systems. In China, the situation in the education domain follows the same path. Currently, there e...With the fast development of business logic and information technology, today's best solutions are tomorrow's legacy systems. In China, the situation in the education domain follows the same path. Currently, there exists a number of e-learning legacy assets with accumulated practical business experience, such as program resource, usage behaviour data resource, and so on. In order to use these legacy assets adequately and efficiently, we should not only utilize the explicit assets but also discover the hidden assets. The usage behaviour data resource is the set of practical operation sequences requested by all users. The hidden patterns in this data resource will provide users' practical experiences, which can benefit the service composition in service-oriented architecture (SOA) migration. Namely, these discovered patterns will be the candidate composite services (coarse-grained) in SOA systems. Although data mining techniques have been used for software engineering tasks, little is known about how they can be used for service composition of migrating an e-learning legacy system (MELS) to SOA. In this paper, we propose a service composition approach based on sequence mining techniques for MELS. Composite services found by this approach will be the complementation of business logic analysis results of MELS. The core of this approach is to develop an appropriate sequence mining algorithm for mining related data collected from an e-learning legacy system. According to the features of execution trace data on usage behaviour from this e-learning legacy system and needs of further pattern analysis, we propose a sequential mining algorithm to mine this kind of data of tile legacy system. For validation, this approach has been applied to the corresponding real data, which was collected from the e-learning legacy system; meanwhile, some investigation questionnaires were set up to collect satisfaction data. The investigation result is 90% the same with the result obtained through our approach.展开更多
This paper proposes how to learn and generate multiple action sequences of a humanoid robot. At first, all the basic action sequences, also called primitive behaviors, are learned by a recurrent neural network with pa...This paper proposes how to learn and generate multiple action sequences of a humanoid robot. At first, all the basic action sequences, also called primitive behaviors, are learned by a recurrent neural network with parametric bias (RNNPB) and the value of the internal nodes which are parametric bias (PB) determining the output with different primitive behaviors are obtained. The training of the RNN uses back propagation through time (BPTT) method. After that, to generate the learned behaviors, or a more complex behavior which is the combination of the primitive behaviors, a reinforcement learning algorithm: Q-learning (QL) is adopt to determine which PB value is adaptive for the generation. Finally, using a real humanoid robot, the proposed method was confirmed its effectiveness by the results of experiment.展开更多
This paper studies a single machine scheduling problem with time-dependent learning and setup times. Time-dependent learning means that the actual processing time of a job is a function of the sum of the normal proces...This paper studies a single machine scheduling problem with time-dependent learning and setup times. Time-dependent learning means that the actual processing time of a job is a function of the sum of the normal processing times of the jobs already scheduled. The setup time of a job is proportional to the length of the already processed jobs, that is, past-sequence-dependent (psd) setup time. We show that the addressed problem remains polynomially solvable for the objectives, i.e., minimization of the total completion time and minimization of the total weighted completion time. We also show that the smallest processing time (SPT) rule provides the optimum sequence for the addressed problem.展开更多
Online sensing can provide useful information in monitoring applications,for example,machine health monitoring,structural condition monitoring,environmental monitoring,and many more.Missing data is generally a signifi...Online sensing can provide useful information in monitoring applications,for example,machine health monitoring,structural condition monitoring,environmental monitoring,and many more.Missing data is generally a significant issue in the sensory data that is collected online by sensing systems,which may affect the goals of monitoring programs.In this paper,a sequence-to-sequence learning model based on a recurrent neural network(RNN)architecture is presented.In the proposed method,multivariate time series of the monitored parameters is embedded into the neural network through layer-by-layer encoders where the hidden features of the inputs are adaptively extracted.Afterwards,predictions of the missing data are generated by network decoders,which are one-step-ahead predictive data sequences of the monitored parameters.The prediction performance of the proposed model is validated based on a real-world sensory dataset.The experimental results demonstrate the performance of the proposed RNN-encoder-decoder model with its capability in sequence-to-sequence learning for online imputation of sensory data.展开更多
Data-driven approaches such as neural networks are increasingly used for deep excavations due to the growing amount of available monitoring data in practical projects.However,most neural network models only use the da...Data-driven approaches such as neural networks are increasingly used for deep excavations due to the growing amount of available monitoring data in practical projects.However,most neural network models only use the data from a single monitoring point and neglect the spatial relationships between multiple monitoring points.Besides,most models lack flexibility in providing predictions for multiple days after monitoring activity.This study proposes a sequence-to-sequence(seq2seq)two-dimensional(2D)convolutional long short-term memory neural network(S2SCL2D)for predicting the spatiotemporal wall deflections induced by deep excavations.The model utilizes the data from all monitoring points on the entire wall and extracts spatiotemporal features from data by combining the 2D convolutional layers and long short-term memory(LSTM)layers.The S2SCL2D model achieves a long-term prediction of wall deflections through a recursive seq2seq structure.The excavation depth,which has a significant impact on wall deflections,is also considered using a feature fusion method.An excavation project in Hangzhou,China,is used to illustrate the proposed model.The results demonstrate that the S2SCL2D model has superior prediction accuracy and robustness than that of the LSTM and S2SCL1D(one-dimensional)models.The prediction model demonstrates a strong generalizability when applied to an adjacent excavation.Based on the long-term prediction results,practitioners can plan and allocate resources in advance to address the potential engineering issues.展开更多
In recent years,deep learning methods have gradually been applied to prediction tasks related to Arctic sea ice concentration,but relatively little research has been conducted for larger spatial and temporal scales,ma...In recent years,deep learning methods have gradually been applied to prediction tasks related to Arctic sea ice concentration,but relatively little research has been conducted for larger spatial and temporal scales,mainly due to the limited time coverage of observations and reanalysis data.Meanwhile,deep learning predictions of sea ice thickness(SIT)have yet to receive ample attention.In this study,two data-driven deep learning(DL)models are built based on the ConvLSTM and fully convolutional U-net(FC-Unet)algorithms and trained using CMIP6 historical simulations for transfer learning and fine-tuned using reanalysis/observations.These models enable monthly predictions of Arctic SIT without considering the complex physical processes involved.Through comprehensive assessments of prediction skills by season and region,the results suggest that using a broader set of CMIP6 data for transfer learning,as well as incorporating multiple climate variables as predictors,contribute to better prediction results,although both DL models can effectively predict the spatiotemporal features of SIT anomalies.Regarding the predicted SIT anomalies of the FC-Unet model,the spatial correlations with reanalysis reach an average level of 89%over all months,while the temporal anomaly correlation coefficients are close to unity in most cases.The models also demonstrate robust performances in predicting SIT and SIE during extreme events.The effectiveness and reliability of the proposed deep transfer learning models in predicting Arctic SIT can facilitate more accurate pan-Arctic predictions,aiding climate change research and real-time business applications.展开更多
基金supported by grants from the Zhejiang Provincial Natural Science Foundation(LGJ22H180001)Zhejiang Medical and Health Science and Technology Project(2021KY249)the National Key R&D Program of China(2017YFC1310000).
文摘Attempts have been made to modulate motor sequence learning(MSL)through repetitive transcranial magnetic stimulation,targeting different sites within the sensorimotor network.However,the target with the optimum modulatory effect on neural plasticity associated with MSL remains unclarified.This study was therefore designed to compare the role of the left primary motor cortex and the left supplementary motor area proper(SMAp)in modulating MSL across different complexity levels and for both hands,as well as the associated neuroplasticity by applying intermittent theta burst stimulation together with the electroencephalogram and concurrent transcranial magnetic stimulation.Our data demonstrated the role of SMAp stimulation in modulating neural communication to support MSL,which is achieved by facilitating regional activation and orchestrating neural coupling across distributed brain regions,particularly in interhemispheric connections.These findings may have important clinical implications,particularly for motor rehabilitation in populations such as post-stroke patients.
基金funded by the Inner Mongolia Natural Science Foundation(No.2024MS04021)the Science and Technology Plan of Inner Mongolia Autonomous Region(No.2023YFSH0004)the Director Fund of the Inner Mongolia Autonomous Region Seismological Bureau(No.2023GG01,No.2023GG02,No.2023MS05,No.2023QN13)。
文摘Machine learning(ML)efficiently and accurately processes dense seismic array data,improving earthquake catalog creation,which is crucial for understanding earthquake sequences and fault systems;analyzing its reliability is also essential.An M5.8 earthquake struck Alxa Left Banner,Inner Mongolia,China on April 15,2015,a region with limited CENC monitoring capabilities,making analysis challenging.However,abundant data from ChinArray provided valuable observations for assessing the event.This study leveraged ChinArray data from the 2015 Alxa Left Banner earthquake sequence,employing machine learning(specifically PhaseNet,a deep learning method,and GaMMA,a Bayesian approach)for automated seismic phase picking,association,and location analysis.Our generated catalog,comprising 10,432 phases from 708 events,is roughly ten times larger than the CENC catalog,encompassing all CENC events with strong consistency.A slight magnitude overestimation is observed only at lower magnitudes.Furthermore,the catalog adheres to the Gutenberg-Richter and Omori laws spatially,temporally,and in magnitude distribution,demonstrating its high reliability.Double-difference tomography refined locations for 366 events,yielding a more compact spatial distribution with horizontal errors within 100m,vertical errors within 300m,and travel-time residuals within 0.05s.Depths predominantly range from 10-30km.Aftershocks align primarily NEE,with the mainshock east of the aftershock zone.The near-vertical main fault plane dips northwestward,exhibiting a Y-shaped branching structure,converging at depth and expanding towards the surface.FOCMEC analysis,using first motion and amplitude ratios,yielded focal mechanism solutions for 10 events,including the mainshock.These solutions consistently indicate a strike-slip mechanism with a minor extensional component.Integrating the earthquake sequence's spatial distribution and focal mechanisms suggests the seismogenic structure is a negative flower structure,consistent with the Dengkou-Benjing fault.Comparing the CENC and ML-generated catalogs using the maximum curvature(MAXC)method reveals a 0.6 decrease in completeness magnitude(M_(C)).However,magnitude-frequency distribution discrepancies above the MAXC-estimated M_(C)suggest MAXC may underestimate both M_(C)and the b-value.This study analyzes the 2015 Alxa Left Banner M5.8 earthquake using a reliable,MLgenerated earthquake catalog,revealing detailed information about the sequence,faulting structure,aftershock distribution,and stress characteristics.
基金supported by the National Natural Science Foundation of China(32070656)。
文摘Rice(Oryza sativa)is a staple food for more than half of the world's population and a critical crop for global agriculture.Understanding the regulatory mechanisms that control gene expression in the rice genome is fundamental for advancing agricultural productivity and food security.In mechanism,cis-regulatory elements(including promoters,enhancers,silencers,and insulators)are key DNA sequences whose activities determine the spatial and temporal expression patterns of nearby genes(Yocca and Edger,2022;Schmitz et al.,2022).
基金supported by the National Natural Science Foundation of China[Grant No.62206185]。
文摘Objective To establish and validate a novel diabetic retinopathy(DR)risk-prediction model using a whole-exome sequencing(WES)-based machine learning(ML)method.Methods WES was performed to identify potential single nucleotide polymorphism(SNP)or mutation sites in a DR pedigree comprising 10 members.A prediction model was established and validated in a cohort of 420 type 2 diabetic patients based on both genetic and demographic features.The contribution of each feature was assessed using Shapley Additive explanation analysis.The efficacies of the models with and without SNP were compared.Results WES revealed that seven SNPs/mutations(rs116911833 in TRIM7,1997T>C in LRBA,1643T>C in PRMT10,rs117858678 in C9orf152,rs201922794 in CLDN25,rs146694895 in SH3GLB2,and rs201407189 in FANCC)were associated with DR.Notably,the model including rs146694895 and rs201407189 achieved better performance in predicting DR(accuracy:80.2%;sensitivity:83.3%;specificity:76.7%;area under the receiver operating characteristic curve[AUC]:80.0%)than the model without these SNPs(accuracy:79.4%;sensitivity:80.3%;specificity:78.3%;AUC:79.3%).Conclusion Novel SNP sites associated with DR were identified in the DR pedigree.Inclusion of rs146694895 and rs201407189 significantly enhanced the performance of the ML-based DR prediction model.
基金National Key Research and Development Program of China,Grant/Award Number:2022YFF0904303Beijing Science and Technology Planning Project,Grant/Award Number:Z221100006322003National Natural Science Foundation of China,Grant/Award Number:61932003。
文摘Computer-aided design(CAD)software continues to be a crucial tool in digital twin application and manufacturing,facilitating the design of various products.We present a novel CAD generation method,an agent that constructs the CAD sequences containing the sketch-and-extrude modelling operations efficiently and with high quality.Starting from the sketch and extrusion operation sequences,we utilise the transformer encoder to encode them into different disentangled codebooks to represent their distribution properties while considering their correlations.Then,a combination of auto-regressive and non-autoregressive samplers is trained to sample the code for CAD sequence con-struction.Extensive experiments demonstrate that our model generates diverse and high-quality CAD models.We also show some cases of real digital twin applications and indicate that our generated model can be used as the data source for the digital twin platform,exhibiting designers'potential.
基金Shaanxi Science Fund for Distinguished Young Scholars,Grant/Award Number:2024JC-JCQN-57Xi’an Science and Technology Plan Project,Grant/Award Number:2023JH-QCYJQ-0086+2 种基金Scientific Research Program Funded by Education Department of Shaanxi Provincial Government,Grant/Award Number:P23JP071Engineering Technology Research Center of Shaanxi Province for Intelligent Testing and Reliability Evaluation of Electronic Equipments,Grant/Award Number:2023-ZC-GCZX-00472022 Shaanxi University Youth Innovation Team Project。
文摘The use of dynamic programming(DP)algorithms to learn Bayesian network structures is limited by their high space complexity and difficulty in learning the structure of large-scale networks.Therefore,this study proposes a DP algorithm based on node block sequence constraints.The proposed algorithm constrains the traversal process of the parent graph by using the M-sequence matrix to considerably reduce the time consumption and space complexity by pruning the traversal process of the order graph using the node block sequence.Experimental results show that compared with existing DP algorithms,the proposed algorithm can obtain learning results more efficiently with less than 1%loss of accuracy,and can be used for learning larger-scale networks.
基金supported by the National Key R&D Program of China(2022YFD1401600)the National Science Foundation for Distinguished Young Scholars of Zhejang Province,China(LR23C140001)supported by the Key Area Research and Development Program of Guangdong Province,China(2018B020205003 and 2020B0202090001).
文摘Inferring phylogenetic trees from molecular sequences is a cornerstone of evolutionary biology.Many standard phylogenetic methods(such as maximum-likelihood[ML])rely on explicit models of sequence evolution and thus often suffer from model misspecification or inadequacy.The on-rising deep learning(DL)techniques offer a powerful alternative.Deep learning employs multi-layered artificial neural networks to progressively transform input data into more abstract and complex representations.DL methods can autonomously uncover meaningful patterns from data,thereby bypassing potential biases introduced by predefined features(Franklin,2005;Murphy,2012).Recent efforts have aimed to apply deep neural networks(DNNs)to phylogenetics,with a growing number of applications in tree reconstruction(Suvorov et al.,2020;Zou et al.,2020;Nesterenko et al.,2022;Smith and Hahn,2023;Wang et al.,2023),substitution model selection(Abadi et al.,2020;Burgstaller-Muehlbacher et al.,2023),and diversification rate inference(Voznica et al.,2022;Lajaaiti et al.,2023;Lambert et al.,2023).In phylogenetic tree reconstruction,PhyDL(Zou et al.,2020)and Tree_learning(Suvorov et al.,2020)are two notable DNN-based programs designed to infer unrooted quartet trees directly from alignments of four amino acid(AA)and DNA sequences,respectively.
基金funded by the Researchers Supporting Project number(RSPD2025R857),King Saud University,Riyadh,Saudi Arabia.
文摘Many bioinformatics applications require determining the class of a newly sequenced Deoxyribonucleic acid(DNA)sequence,making DNA sequence classification an integral step in performing bioinformatics analysis,where large biomedical datasets are transformed into valuable knowledge.Existing methods rely on a feature extraction step and suffer from high computational time requirements.In contrast,newer approaches leveraging deep learning have shown significant promise in enhancing accuracy and efficiency.In this paper,we investigate the performance of various deep learning architectures:Convolutional Neural Network(CNN),CNN-Long Short-Term Memory(CNNLSTM),CNN-Bidirectional Long Short-Term Memory(CNN-BiLSTM),Residual Network(ResNet),and InceptionV3 for DNA sequence classification.Various numerical and visual data representation techniques are utilized to represent the input datasets,including:label encoding,k-mer sentence encoding,k-mer one-hot vector,Frequency Chaos Game Representation(FCGR)and 5-Color Map(ColorSquare).Three datasets are used for the training of the models including H3,H4 and DNA Sequence Dataset(Yeast,Human,Arabidopsis Thaliana).Experiments are performed to determine which combination of DNA representation and deep learning architecture yields improved performance for the classification task.Our results indicate that using a hybrid CNN-LSTM neural network trained on DNA sequences represented as one-hot encoded k-mer sequences yields the best performance,achieving an accuracy of 92.1%.
基金funded by National Key Research and Development Program of China(2021YFD1200404)the Yangzhou University Interdisciplinary Research Foundation for Animal Science Discipline of Targeted Support(yzuxk202016)the Project of Genetic Improvement for Agricultural Species(Dairy Cattle)of Shandong Province(2019LZGC011).
文摘Background Breed identification is useful in a variety of biological contexts.Breed identification usually involves two stages,i.e.,detection of breed-informative SNPs and breed assignment.For both stages,there are several methods proposed.However,what is the optimal combination of these methods remain unclear.In this study,using the whole genome sequence data available for 13 cattle breeds from Run 8 of the 1,000 Bull Genomes Project,we compared the combinations of three methods(Delta,FST,and In)for breed-informative SNP detection and five machine learning methods(KNN,SVM,RF,NB,and ANN)for breed assignment with respect to different reference population sizes and difference numbers of most breed-informative SNPs.In addition,we evaluated the accuracy of breed identification using SNP chip data of different densities.Results We found that all combinations performed quite well with identification accuracies over 95%in all scenarios.However,there was no combination which performed the best and robust across all scenarios.We proposed to inte-grate the three breed-informative detection methods,named DFI,and integrate the three machine learning methods,KNN,SVM,and RF,named KSR.We found that the combination of these two integrated methods outperformed the other combinations with accuracies over 99%in most cases and was very robust in all scenarios.The accuracies from using SNP chip data were only slightly lower than that from using sequence data in most cases.Conclusions The current study showed that the combination of DFI and KSR was the optimal strategy.Using sequence data resulted in higher accuracies than using chip data in most cases.However,the differences were gener-ally small.In view of the cost of genotyping,using chip data is also a good option for breed identification.
基金supported by the General Program of the National Natural Science Foundation of China under Grant No.62172093the National Key R&D Program of China under Grant No.2018YFB1800602+1 种基金2019 Industrial Internet Innovation and Development Project,Ministry of Industry and Information Technology(MIIT)under Grant No.6709010003Ministry of Education-China Mobile Research Fund under Grant No.MCM20180506。
文摘As an essential function of encrypted Internet traffic analysis,encrypted traffic service classification can support both coarse-grained network service traffic management and security supervision.However,the traditional plaintext-based Deep Packet Inspection(DPI)method cannot be applied to such a classification.Moreover,machine learning-based existing methods encounter two problems during feature selection:complex feature overcost processing and Transport Layer Security(TLS)version discrepancy.In this paper,we consider differences between encryption network protocol stacks and propose a composite deep learning-based method in multiprotocol environments using a sliding multiple Protocol Data Unit(multiPDU)length sequence as features by fully utilizing the Markov property in a multiPDU length sequence and maintaining suitability with a TLS-1.3 environment.Control experiments show that both Length-Sensitive(LS)composite deep learning model using a capsule neural network and LS-long short time memory achieve satisfactory effectiveness in F1-score and performance.Owing to faster feature extraction,our method is suitable for actual network environments and superior to state-of-the-art methods.
基金supported by the National Natural Science Foundation of China under Grant No.61271168
文摘An idea of estimating the direct sequence spread spectrum(DSSS) signal pseudo-noise(PN) sequence is presented. Without the apriority knowledge about the DSSS signal in the non-cooperation condition, we propose a self-organizing feature map(SOFM) neural network algorithm to detect and identify the PN sequence. A non-supervised learning algorithm is proposed according the Kohonen rule in SOFM. The blind algorithm can also estimate the PN sequence in a low signal-to-noise(SNR) and computer simulation demonstrates that the algorithm is effective. Compared with the traditional correlation algorithm based on slip-correlation, the proposed algorithm's bit error rate(BER) and complexity are lower.
基金funded by the Joint Funds of the National Natural Science Foundation of China(U19B6003)the PetroChina Innovation Foundation(Grant No.2020D5007-0203)it was further supported by the Science Foundation of China University of Petroleum,Beijing(Nos.2462021YXZZ010,2462018QZDX13,and 2462020YXZZ028).
文摘Production performance prediction of tight gas reservoirs is crucial to the estimation of ultimate recovery,which has an important impact on gas field development planning and economic evaluation.Owing to the model’s simplicity,the decline curve analysis method has been widely used to predict production performance.The advancement of deep-learning methods provides an intelligent way of analyzing production performance in tight gas reservoirs.In this paper,a sequence learning method to improve the accuracy and efficiency of tight gas production forecasting is proposed.The sequence learning methods used in production performance analysis herein include the recurrent neural network(RNN),long short-term memory(LSTM)neural network,and gated recurrent unit(GRU)neural network,and their performance in the tight gas reservoir production prediction is investigated and compared.To further improve the performance of the sequence learning method,the hyperparameters in the sequence learning methods are optimized through a particle swarm optimization algorithm,which can greatly simplify the optimization process of the neural network model in an automated manner.Results show that the optimized GRU and RNN models have more compact neural network structures than the LSTM model and that the GRU is more efficiently trained.The predictive performance of LSTM and GRU is similar,and both are better than the RNN and the decline curve analysis model and thus can be used to predict tight gas production.
基金This work was supported in part by the National Natural Science Foundation of China(No.62002392)in part by the Key Research and Development Plan of Hunan Province(No.2019SK2022)+2 种基金in part by the Natural Science Foundation of Hunan Province(No.2020JJ4140 and 2020JJ4141)in part by the Science Research Projects of Hunan Provincial Education Department(No.19B584)in part by the Postgraduate Excellent teaching team Project of Hunan Province[Grant[2019]370-133].
文摘Aortic dissection(AD)is one of the most serious diseases with high mortality,and its diagnosis mainly depends on computed tomography(CT)results.Most existing automatic diagnosis methods of AD are only suitable for AD recognition,which usually require preselection of CT images and cannot be further classified to different types.In this work,we constructed a dataset of 105 cases with a total of 49021 slices,including 31043 slices expertlevel annotation and proposed a two-stage AD diagnosis structure based on sequence information and deep learning.The proposed region of interest(RoI)extraction algorithm based on sequence information(RESI)can realize high-precision for RoI identification in the first stage.Then DenseNet-121 is applied for further diagnosis.Specially,the proposed method can judge the type of AD without preselection of CT images.The experimental results show that the accuracy of Stanford typing classification of AD is 89.19%,and the accuracy at the slice-level reaches 97.41%,which outperform the state-ofart methods.It can provide important decision-making information for the determination of further surgical treatment plan for patients.
基金supported by E-learning Platform, National Torch Project (No. z20040010)
文摘With the fast development of business logic and information technology, today's best solutions are tomorrow's legacy systems. In China, the situation in the education domain follows the same path. Currently, there exists a number of e-learning legacy assets with accumulated practical business experience, such as program resource, usage behaviour data resource, and so on. In order to use these legacy assets adequately and efficiently, we should not only utilize the explicit assets but also discover the hidden assets. The usage behaviour data resource is the set of practical operation sequences requested by all users. The hidden patterns in this data resource will provide users' practical experiences, which can benefit the service composition in service-oriented architecture (SOA) migration. Namely, these discovered patterns will be the candidate composite services (coarse-grained) in SOA systems. Although data mining techniques have been used for software engineering tasks, little is known about how they can be used for service composition of migrating an e-learning legacy system (MELS) to SOA. In this paper, we propose a service composition approach based on sequence mining techniques for MELS. Composite services found by this approach will be the complementation of business logic analysis results of MELS. The core of this approach is to develop an appropriate sequence mining algorithm for mining related data collected from an e-learning legacy system. According to the features of execution trace data on usage behaviour from this e-learning legacy system and needs of further pattern analysis, we propose a sequential mining algorithm to mine this kind of data of tile legacy system. For validation, this approach has been applied to the corresponding real data, which was collected from the e-learning legacy system; meanwhile, some investigation questionnaires were set up to collect satisfaction data. The investigation result is 90% the same with the result obtained through our approach.
文摘This paper proposes how to learn and generate multiple action sequences of a humanoid robot. At first, all the basic action sequences, also called primitive behaviors, are learned by a recurrent neural network with parametric bias (RNNPB) and the value of the internal nodes which are parametric bias (PB) determining the output with different primitive behaviors are obtained. The training of the RNN uses back propagation through time (BPTT) method. After that, to generate the learned behaviors, or a more complex behavior which is the combination of the primitive behaviors, a reinforcement learning algorithm: Q-learning (QL) is adopt to determine which PB value is adaptive for the generation. Finally, using a real humanoid robot, the proposed method was confirmed its effectiveness by the results of experiment.
文摘This paper studies a single machine scheduling problem with time-dependent learning and setup times. Time-dependent learning means that the actual processing time of a job is a function of the sum of the normal processing times of the jobs already scheduled. The setup time of a job is proportional to the length of the already processed jobs, that is, past-sequence-dependent (psd) setup time. We show that the addressed problem remains polynomially solvable for the objectives, i.e., minimization of the total completion time and minimization of the total weighted completion time. We also show that the smallest processing time (SPT) rule provides the optimum sequence for the addressed problem.
文摘Online sensing can provide useful information in monitoring applications,for example,machine health monitoring,structural condition monitoring,environmental monitoring,and many more.Missing data is generally a significant issue in the sensory data that is collected online by sensing systems,which may affect the goals of monitoring programs.In this paper,a sequence-to-sequence learning model based on a recurrent neural network(RNN)architecture is presented.In the proposed method,multivariate time series of the monitored parameters is embedded into the neural network through layer-by-layer encoders where the hidden features of the inputs are adaptively extracted.Afterwards,predictions of the missing data are generated by network decoders,which are one-step-ahead predictive data sequences of the monitored parameters.The prediction performance of the proposed model is validated based on a real-world sensory dataset.The experimental results demonstrate the performance of the proposed RNN-encoder-decoder model with its capability in sequence-to-sequence learning for online imputation of sensory data.
基金supported by the National Natural Science Foundation of China(Grant No.42307218)the Foundation of Key Laboratory of Soft Soils and Geoenvironmental Engineering(Zhejiang University),Ministry of Education(Grant No.2022P08)the Natural Science Foundation of Zhejiang Province(Grant No.LTZ21E080001).
文摘Data-driven approaches such as neural networks are increasingly used for deep excavations due to the growing amount of available monitoring data in practical projects.However,most neural network models only use the data from a single monitoring point and neglect the spatial relationships between multiple monitoring points.Besides,most models lack flexibility in providing predictions for multiple days after monitoring activity.This study proposes a sequence-to-sequence(seq2seq)two-dimensional(2D)convolutional long short-term memory neural network(S2SCL2D)for predicting the spatiotemporal wall deflections induced by deep excavations.The model utilizes the data from all monitoring points on the entire wall and extracts spatiotemporal features from data by combining the 2D convolutional layers and long short-term memory(LSTM)layers.The S2SCL2D model achieves a long-term prediction of wall deflections through a recursive seq2seq structure.The excavation depth,which has a significant impact on wall deflections,is also considered using a feature fusion method.An excavation project in Hangzhou,China,is used to illustrate the proposed model.The results demonstrate that the S2SCL2D model has superior prediction accuracy and robustness than that of the LSTM and S2SCL1D(one-dimensional)models.The prediction model demonstrates a strong generalizability when applied to an adjacent excavation.Based on the long-term prediction results,practitioners can plan and allocate resources in advance to address the potential engineering issues.
基金supported by the National Natural Science Foundation of China(Grant Nos.41976193 and 42176243).
文摘In recent years,deep learning methods have gradually been applied to prediction tasks related to Arctic sea ice concentration,but relatively little research has been conducted for larger spatial and temporal scales,mainly due to the limited time coverage of observations and reanalysis data.Meanwhile,deep learning predictions of sea ice thickness(SIT)have yet to receive ample attention.In this study,two data-driven deep learning(DL)models are built based on the ConvLSTM and fully convolutional U-net(FC-Unet)algorithms and trained using CMIP6 historical simulations for transfer learning and fine-tuned using reanalysis/observations.These models enable monthly predictions of Arctic SIT without considering the complex physical processes involved.Through comprehensive assessments of prediction skills by season and region,the results suggest that using a broader set of CMIP6 data for transfer learning,as well as incorporating multiple climate variables as predictors,contribute to better prediction results,although both DL models can effectively predict the spatiotemporal features of SIT anomalies.Regarding the predicted SIT anomalies of the FC-Unet model,the spatial correlations with reanalysis reach an average level of 89%over all months,while the temporal anomaly correlation coefficients are close to unity in most cases.The models also demonstrate robust performances in predicting SIT and SIE during extreme events.The effectiveness and reliability of the proposed deep transfer learning models in predicting Arctic SIT can facilitate more accurate pan-Arctic predictions,aiding climate change research and real-time business applications.