This paper focuses on the numerical solution of a tumor growth model under a data-driven approach.Based on the inherent laws of the data and reasonable assumptions,an ordinary differential equation model for tumor gro...This paper focuses on the numerical solution of a tumor growth model under a data-driven approach.Based on the inherent laws of the data and reasonable assumptions,an ordinary differential equation model for tumor growth is established.Nonlinear fitting is employed to obtain the optimal parameter estimation of the mathematical model,and the numerical solution is carried out using the Matlab software.By comparing the clinical data with the simulation results,a good agreement is achieved,which verifies the rationality and feasibility of the model.展开更多
Permanent magnet synchronous motor(PMSM)is widely used in alternating current servo systems as it provides high eficiency,high power density,and a wide speed regulation range.The servo system is placing higher demands...Permanent magnet synchronous motor(PMSM)is widely used in alternating current servo systems as it provides high eficiency,high power density,and a wide speed regulation range.The servo system is placing higher demands on its control performance.The model predictive control(MPC)algorithm is emerging as a potential high-performance motor control algorithm due to its capability of handling multiple-input and multipleoutput variables and imposed constraints.For the MPC used in the PMSM control process,there is a nonlinear disturbance caused by the change of electromagnetic parameters or load disturbance that may lead to a mismatch between the nominal model and the controlled object,which causes the prediction error and thus affects the dynamic stability of the control system.This paper proposes a data-driven MPC strategy in which the historical data in an appropriate range are utilized to eliminate the impact of parameter mismatch and further improve the control performance.The stability of the proposed algorithm is proved as the simulation demonstrates the feasibility.Compared with the classical MPC strategy,the superiority of the algorithm has also been verified.展开更多
This paper aims to develop Machine Learning algorithms to classify electronic articles related to this phenomenon by retrieving information and topic modelling.The Methodology of this study is categorized into three p...This paper aims to develop Machine Learning algorithms to classify electronic articles related to this phenomenon by retrieving information and topic modelling.The Methodology of this study is categorized into three phases:the Text Classification Approach(TCA),the Proposed Algorithms Interpretation(PAI),andfinally,Information Retrieval Approach(IRA).The TCA reflects the text preprocessing pipeline called a clean corpus.The Global Vec-tors for Word Representation(Glove)pre-trained model,FastText,Term Frequency-Inverse Document Fre-quency(TF-IDF),and Bag-of-Words(BOW)for extracting the features have been interpreted in this research.The PAI manifests the Bidirectional Long Short-Term Memory(Bi-LSTM)and Convolutional Neural Network(CNN)to classify the COVID-19 news.Again,the IRA explains the mathematical interpretation of Latent Dirich-let Allocation(LDA),obtained for modelling the topic of Information Retrieval(IR).In this study,99%accuracy was obtained by performing K-fold cross-validation on Bi-LSTM with Glove.A comparative analysis between Deep Learning and Machine Learning based on feature extraction and computational complexity exploration has been performed in this research.Furthermore,some text analyses and the most influential aspects of each document have been explored in this study.We have utilized Bidirectional Encoder Representations from Trans-formers(BERT)as a Deep Learning mechanism in our model training,but the result has not been uncovered satisfactory.However,the proposed system can be adjustable in the real-time news classification of COVID-19.展开更多
Due to the signal reflection and diffraction,site-specific unmodeled errors like multipath effect and Non-Line-of-Sight reception are significant error sources in Global Navigation Satellite System since they cannot b...Due to the signal reflection and diffraction,site-specific unmodeled errors like multipath effect and Non-Line-of-Sight reception are significant error sources in Global Navigation Satellite System since they cannot be easily mitigated.However,how to characterize and model the internal mechanisms and external influences of these site-specific unmodeled errors are still to be investigated.Therefore,we propose a method for characterizing and modeling site-specific unmodeled errors under reflection and diffraction using a data-driven approach.Specifically,we first consider all the popular potential features,which generate the site-specific unmodeled errors.We then use the random forest regression to comprehensively analyze the correlations between the site-specific unmodeled errors and the potential features.We finally characterize and model the site-specific unmodeled errors.Two 7-consecutive datasets dominated by signal reflection and diffraction were conducted.The results show that there are significant differences in the correlations with potential features.They are highly related to the application scenarios,observation types,and satellite types.Notably,the innovation vector often shows a strong correlation with the code site-specific unmodeled errors.For the phase site-specific unmodeled errors,they have high correlations with elevation,azimuth,number of visible satellites,and between-frequency differenced phase observations.In the environments of reflection and diffraction,the sum of the correlations of the top six potential features can reach approximately 88.5 and 87.7%,respectively.Meanwhile,these correlations are stable for different observation types and satellite types.With the integration of a transformer model with the random forest method,a high-precision unmodeled error prediction model is established,demonstrating the necessity to include multiple features for accurate and efficient characterization and modeling of site-specific unmodeled errors.展开更多
Manufacturers are striving to achieve higher energy efficiency without compromising production performance and quality standards.Parallel-serial structures,commonly found in modern production systems,offer a unique ba...Manufacturers are striving to achieve higher energy efficiency without compromising production performance and quality standards.Parallel-serial structures,commonly found in modern production systems,offer a unique balance of flexibility and efficiency by combining parallel processes with sequential workflows.However,their inherent complexity poses significant challenges,particularly in optimizing energy efficiency and ensuring consistent product quality.In data-driven manufacturing environments,it is not clear how to leverage production data to enhance the energy efficiency of production systems.Therefore,this paper studied a data-driven approach to improving energy efficiency in parallel-serial production lines with product quality issues.Firstly,the authors developed a data-driven performance analysis method to evaluate the effects of disruption events,such as energy-saving control actions,machine breakdowns,and product quality failures,on system throughput and energy consumption.Secondly,a periodic energy-saving control method was developed to enhance system energy efficiency using a non-linear programming model.To reduce complexity and improve computational efficiency,the model was simplified by leveraging the intrinsic properties of parallel-serial production lines and solved using an adaptive genetic algorithm.Finally,the effectiveness of the proposed data-driven approach was validated through case studies,providing actionable insights into achieving data-driven energy efficiency optimization in complex production systems.展开更多
Building integrated energy systems(BIESs)are pivotal for enhancing energy efficiency by accounting for a significant proportion of global energy consumption.Two key barriers that reduce the BIES operational efficiency...Building integrated energy systems(BIESs)are pivotal for enhancing energy efficiency by accounting for a significant proportion of global energy consumption.Two key barriers that reduce the BIES operational efficiency mainly lie in the renewable generation uncertainty and operational non-convexity of combined heat and power(CHP)units.To this end,this paper proposes a soft actor-critic(SAC)algorithm to solve the scheduling problem of BIES,which overcomes the model non-convexity and shows advantages in robustness and generalization.This paper also adopts a temporal fusion transformer(TFT)to enhance the optimal solution for the SAC algorithm by forecasting the renewable generation and energy demand.The TFT can effectively capture the complex temporal patterns and dependencies that span multiple steps.Furthermore,its forecasting results are interpretable due to the employment of a self-attention layer so as to assist in more trustworthy decision-making in the SAC algorithm.The proposed hybrid data-driven approach integrating TFT and SAC algorithm,i.e.,TFT-SAC approach,is trained and tested on a real-world dataset to validate its superior performance in reducing the energy cost and computational time compared with the benchmark approaches.The generalization performance for the scheduling policy,as well as the sensitivity analysis,are examined in the case studies.展开更多
Numerous influence factors will lead to the inaccurate prediction of fatigue crack propagation(FCP)life of the metal structure based on the existing FCP model,while the prediction method based on machine learning(ML)a...Numerous influence factors will lead to the inaccurate prediction of fatigue crack propagation(FCP)life of the metal structure based on the existing FCP model,while the prediction method based on machine learning(ML)and data-driven approach can provide a new idea for accurately predicting the FCP life of the metal structure.In response to the inconvenience of the online prediction method and the inaccu-racy of the offline prediction method,an improved offline prediction method based on data feedback is presented in this paper.FCP tests of reduced scale models of balcony opening corners in a cruise ship are conducted to obtain experimental data with respect to the a-N curves.The crack length corresponding to the cycle is trained using a support vector regression(SVR)and back propagation neural network(BP NN)algorithms.FCP prediction lives of test specimens are performed according to the online,offline,and improved offline prediction methods.Effects of the number of feedback data,the sequence length(SL)in the input set,and the cycle interval on prediction accuracy are discussed.The generalization ability of the proposed method is validated by comparing the prediction results with the experimental data in the literature.The larger the number of feedback data,the higher the prediction accuracy.The results show that 1/5 and 1/2 feedback data are needed in the SVR and BP NN algorithm with SL is 5,respectively.Furthermore,the SVR algorithm and SL=5 are recommended for FCP life prediction using the improved offline prediction method.展开更多
The COVID-19 pandemic has significantly impacted global health,social,and economic situations since its emergence in December 2019.The primary focus of this study is to propose a distinct vaccination policy and assess...The COVID-19 pandemic has significantly impacted global health,social,and economic situations since its emergence in December 2019.The primary focus of this study is to propose a distinct vaccination policy and assess its impact on controlling COVID-19 transmission in Malaysia using a Bayesian data-driven approach,concentrating on the year 2021.We employ a compartmental Susceptible-Exposed-Infected-Recovered-Vaccinated(SEIRV)model,incorporating a time-varying transmission rate and a data-driven method for its estimation through an Exploratory Data Analysis(EDA)approach.While no vaccine guarantees total immunity against the disease,and vaccine immunity wanes over time,it is critical to include and accurately estimate vaccine efficacy,as well as a constant vaccine immunity decay or wane factor,to better simulate the dynamics of vaccine-induced protection over time.Based on the distribution and effectiveness of vaccines,we integrated a data-driven estimation of vaccine efficacy,calculated at 75%for Malaysia,underscoring the model's realism and relevance to the specific context of the country.The Bayesian inference framework is used to assimilate various data sources and account for underlying uncertainties in model parameters.The model is fitted to real-world data from Malaysia to analyze disease spread trends and evaluate the effectiveness of our proposed vaccination policy.Our findings reveal that this distinct vaccination policy,which emphasizes an accelerated vaccination rate during the initial stages of the program,is highly effective in mitigating the spread of COVID-19 and substantially reducing the pandemic peak and new infections.The study found that vaccinating 57–66%of the population(as opposed to 76%in the real data)with a better vaccination policy such as proposed here is able to significantly reduce the number of new infections and ultimately reduce the costs associated with new infections.The study contributes to the development of a robust and informative representation of COVID-19 transmission and vaccination,offering valuable insights for policymakers on the potential benefits and limitations of different vaccination policies,particularly highlighting the importance of a well-planned and efficient vaccination rollout strategy.While the methodology used in this study is specifically applied to national data from Malaysia,its successful application to local regions within Malaysia,such as Selangor and Johor,indicates its adaptability and potential for broader application.This demonstrates the model's adaptability for policy assessment and improvement across various demographic and epidemiological landscapes,implying its usefulness for similar datasets from various geographical regions.展开更多
Non-learning based motion and path planning of an Unmanned Aerial Vehicle(UAV)is faced with low computation efficiency,mapping memory occupation and local optimization problems.This article investigates the challenge ...Non-learning based motion and path planning of an Unmanned Aerial Vehicle(UAV)is faced with low computation efficiency,mapping memory occupation and local optimization problems.This article investigates the challenge of quadrotor control using offline reinforcement learning.By establishing a data-driven learning paradigm that operates without real-environment interaction,the proposed workflow offers a safer approach than traditional reinforcement learning,making it particularly suited for UAV control in industrial scenarios.The introduced algorithm evaluates dataset uncertainty and employs a pessimistic estimation to foster offline deep reinforcement learning.Experiments highlight the algorithm's superiority over traditional online reinforcement learning methods,especially when learning from offline datasets.Furthermore,the article emphasizes the importance of a more general behavior policy.In evaluations,the trained policy demonstrated versatility by adeptly navigating diverse obstacles,underscoring its real-world applicability.展开更多
The electricity and steam integrated energy systems,which can capture waste heat and improve the overall energy efficiency,have been widely utilised in industrial parks.However,intensive and frequent changes in demand...The electricity and steam integrated energy systems,which can capture waste heat and improve the overall energy efficiency,have been widely utilised in industrial parks.However,intensive and frequent changes in demands would lead to model parameters with strong time-varying characteristics.This paper proposes a hybrid physics and data-driven framework for online joint state and parameter estimation of steam and electricity integrated energy system.Based on the physical non-linear state space models for the electricity network(EN)and steam heating network(SHN),relevance vector machine is developed to learn parameters'dynamic characteristics with respect to model states,which is embedded with physical models.Then,the online joint state and parameter estimation based on unscented Kalman filter is proposed,which would be learnt recursively to capture the spatiotemporal transient characteristics between electricity and SHNs.The IEEE 39-bus EN and the 29-nodes SHN are employed to verify the effectiveness of the proposed method.The experimental results validate that the pro-posed method can provide a higher estimation accuracy than the state-of-the-art approaches.展开更多
We propose an integrated method of data-driven and mechanism models for well logging formation evaluation,explicitly focusing on predicting reservoir parameters,such as porosity and water saturation.Accurately interpr...We propose an integrated method of data-driven and mechanism models for well logging formation evaluation,explicitly focusing on predicting reservoir parameters,such as porosity and water saturation.Accurately interpreting these parameters is crucial for effectively exploring and developing oil and gas.However,with the increasing complexity of geological conditions in this industry,there is a growing demand for improved accuracy in reservoir parameter prediction,leading to higher costs associated with manual interpretation.The conventional logging interpretation methods rely on empirical relationships between logging data and reservoir parameters,which suffer from low interpretation efficiency,intense subjectivity,and suitability for ideal conditions.The application of artificial intelligence in the interpretation of logging data provides a new solution to the problems existing in traditional methods.It is expected to improve the accuracy and efficiency of the interpretation.If large and high-quality datasets exist,data-driven models can reveal relationships of arbitrary complexity.Nevertheless,constructing sufficiently large logging datasets with reliable labels remains challenging,making it difficult to apply data-driven models effectively in logging data interpretation.Furthermore,data-driven models often act as“black boxes”without explaining their predictions or ensuring compliance with primary physical constraints.This paper proposes a machine learning method with strong physical constraints by integrating mechanism and data-driven models.Prior knowledge of logging data interpretation is embedded into machine learning regarding network structure,loss function,and optimization algorithm.We employ the Physically Informed Auto-Encoder(PIAE)to predict porosity and water saturation,which can be trained without labeled reservoir parameters using self-supervised learning techniques.This approach effectively achieves automated interpretation and facilitates generalization across diverse datasets.展开更多
Cardiovascular diseases(CVDs)remain the leading cause of morbidity and mortality worldwide,necessitating innovative diagnostic and prognostic strategies.Traditional biomarkers like C-reactive protein,uric acid,troponi...Cardiovascular diseases(CVDs)remain the leading cause of morbidity and mortality worldwide,necessitating innovative diagnostic and prognostic strategies.Traditional biomarkers like C-reactive protein,uric acid,troponin,and natriuretic peptides play crucial roles in CVD management,yet they are often limited by sensitivity and specificity constraints.This narrative review critically examines the emerging landscape of cardiac biomarkers and advocates for a multiple-marker approach to enhance early detection,prognosis,and risk stratification of CVD.In recent years,several novel biomarkers have shown promise in revolutionizing CVD diagnostics.Gamma-glutamyltransferase,microRNAs,endothelial microparticles,placental growth factor,trimethylamine N-oxide,retinol-binding protein 4,copeptin,heart-type fatty acid-binding protein,galectin-3,growth differentiation factor-15,soluble suppression of tumorigenicity 2,fibroblast growth factor 23,and adrenomedullin have emerged as significant indicators of CV health.These biomarkers provide insights into various pathophysiological processes,such as oxidative stress,endothelial dysfunction,inflammation,metabolic disturbances,and myocardial injury.The integration of these novel biomarkers with traditional ones offers a more comprehensive understanding of CVD mechanisms.This multiple-marker approach can improve diagnostic accuracy,allowing for better risk stratification and more personalized treatment strategies.This review underscores the need for continued research to validate the clinical utility of these biomarkers and their potential incorporation into routine clinical practice.By leveraging the strengths of both traditional and novel biomarkers,precise therapeutic plans can be developed,thereby improving the management and prognosis of patients with CVDs.The ongoing exploration and validation of these biomarkers are crucial for advancing CV care and addressing the limitations of current diagnostic tools.展开更多
The progressive loss of dopaminergic neurons in affected patient brains is one of the pathological features of Parkinson's disease,the second most common human neurodegenerative disease.Although the detailed patho...The progressive loss of dopaminergic neurons in affected patient brains is one of the pathological features of Parkinson's disease,the second most common human neurodegenerative disease.Although the detailed pathogenesis accounting for dopaminergic neuron degeneration in Parkinson's disease is still unclear,the advancement of stem cell approaches has shown promise for Parkinson's disease research and therapy.The induced pluripotent stem cells have been commonly used to generate dopaminergic neurons,which has provided valuable insights to improve our understanding of Parkinson's disease pathogenesis and contributed to anti-Parkinson's disease therapies.The current review discusses the practical approaches and potential applications of induced pluripotent stem cell techniques for generating and differentiating dopaminergic neurons from induced pluripotent stem cells.The benefits of induced pluripotent stem cell-based research are highlighted.Various dopaminergic neuron differentiation protocols from induced pluripotent stem cells are compared.The emerging three-dimension-based brain organoid models compared with conventional two-dimensional cell culture are evaluated.Finally,limitations,challenges,and future directions of induced pluripotent stem cell–based approaches are analyzed and proposed,which will be significant to the future application of induced pluripotent stem cell-related techniques for Parkinson's disease.展开更多
This article provides a comprehensive analysis of the study by Hou et al,focusing on the complex interplay between psychological and physical factors in the postoperative recovery(POR)of patients with perianal disease...This article provides a comprehensive analysis of the study by Hou et al,focusing on the complex interplay between psychological and physical factors in the postoperative recovery(POR)of patients with perianal diseases.The study sheds light on how illness perception,anxiety,and depression significantly influence recovery outcomes.Hou et al developed a predictive model that demonstrated high accuracy in identifying patients at risk of poor recovery.The article explores the critical role of pre-operative psychological assessment,highlighting the need for mental health support and personalized recovery plans in enhancing POR quality.A multidisciplinary approach,integrating mental health professionals with surgeons,anesthesiologists,and other specialists,is emphasized to ensure comprehensive care for patients.The study’s findings serve as a call to integrate psychological care into surgical practice to optimize outcomes for patients with perianal diseases.展开更多
Objective: To explore the effect of Health Action Process Approach (HAPA) theory in patients with type D personality psoriasis. Methods: A total of 66 patients with type D personality psoriasis admitted to the dermato...Objective: To explore the effect of Health Action Process Approach (HAPA) theory in patients with type D personality psoriasis. Methods: A total of 66 patients with type D personality psoriasis admitted to the dermatology department of a top-three hospital in Jingzhou City from November 2022 to July 2023 were selected and divided into control group and test group with 33 cases in each group by random number table method. The control group received routine health education, and the experimental group received health education based on the HAPA theory. Chronic disease self-efficacy scale, hospital anxiety and depression scale and skin disease quality of life scale were used to evaluate the effect of intervention. Results: After 3 months of intervention, the scores of self-efficacy in experimental group were higher than those in control group (P P Conclusion: Health education based on the theory of HAPA can enhance the self-efficacy of patients with type D personality psoriasis, relieve negative emotions and improve their quality of life.展开更多
A data-driven model ofmultiple variable cutting(M-VCUT)level set-based substructure is proposed for the topology optimization of lattice structures.TheM-VCUTlevel setmethod is used to represent substructures,enriching...A data-driven model ofmultiple variable cutting(M-VCUT)level set-based substructure is proposed for the topology optimization of lattice structures.TheM-VCUTlevel setmethod is used to represent substructures,enriching their diversity of configuration while ensuring connectivity.To construct the data-driven model of substructure,a database is prepared by sampling the space of substructures spanned by several substructure prototypes.Then,for each substructure in this database,the stiffness matrix is condensed so that its degrees of freedomare reduced.Thereafter,the data-drivenmodel of substructures is constructed through interpolationwith compactly supported radial basis function(CS-RBF).The inputs of the data-driven model are the design variables of topology optimization,and the outputs are the condensed stiffness matrix and volume of substructures.During the optimization,this data-driven model is used,thus avoiding repeated static condensation that would requiremuch computation time.Several numerical examples are provided to verify the proposed method.展开更多
The outstanding comprehensive mechanical properties of newly developed hybrid lattice structures make them useful in engineering applications for bearing multiple mechanical loads.Additive-manufacturing technologies m...The outstanding comprehensive mechanical properties of newly developed hybrid lattice structures make them useful in engineering applications for bearing multiple mechanical loads.Additive-manufacturing technologies make it possible to fabricate these highly spatially programmable structures and greatly enhance the freedom in their design.However,traditional analytical methods do not sufficiently reflect the actual vibration-damping mechanism of lattice structures and are limited by their high computational cost.In this study,a hybrid lattice structure consisting of various cells was designed based on quasi-static and vibration experiments.Subsequently,a novel parametric design method based on a data-driven approach was developed for hybrid lattices with engineered properties.The response surface method was adopted to define the sensitive optimization target.A prediction model for the lattice geometric parameters and vibration properties was established using a backpropagation neural network.Then,it was integrated into the genetic algorithm to create the optimal hybrid lattice with varying geometric features and the required wide-band vibration-damping characteristics.Validation experiments were conducted,demonstrating that the optimized hybrid lattice can achieve the target properties.In addition,the data-driven parametric design method can reduce computation time and be widely applied to complex structural designs when analytical and empirical solutions are unavailable.展开更多
The Underwater Acoustic(UWA)channel is bandwidth-constrained and experiences doubly selective fading.It is challenging to acquire perfect channel knowledge for Orthogonal Frequency Division Multiplexing(OFDM)communica...The Underwater Acoustic(UWA)channel is bandwidth-constrained and experiences doubly selective fading.It is challenging to acquire perfect channel knowledge for Orthogonal Frequency Division Multiplexing(OFDM)communications using a finite number of pilots.On the other hand,Deep Learning(DL)approaches have been very successful in wireless OFDM communications.However,whether they will work underwater is still a mystery.For the first time,this paper compares two categories of DL-based UWA OFDM receivers:the DataDriven(DD)method,which performs as an end-to-end black box,and the Model-Driven(MD)method,also known as the model-based data-driven method,which combines DL and expert OFDM receiver knowledge.The encoder-decoder framework and Convolutional Neural Network(CNN)structure are employed to establish the DD receiver.On the other hand,an unfolding-based Minimum Mean Square Error(MMSE)structure is adopted for the MD receiver.We analyze the characteristics of different receivers by Monte Carlo simulations under diverse communications conditions and propose a strategy for selecting a proper receiver under different communication scenarios.Field trials in the pool and sea are also conducted to verify the feasibility and advantages of the DL receivers.It is observed that DL receivers perform better than conventional receivers in terms of bit error rate.展开更多
When assessing seismic liquefaction potential with data-driven models,addressing the uncertainties of establishing models,interpreting cone penetration tests(CPT)data and decision threshold is crucial for avoiding bia...When assessing seismic liquefaction potential with data-driven models,addressing the uncertainties of establishing models,interpreting cone penetration tests(CPT)data and decision threshold is crucial for avoiding biased data selection,ameliorating overconfident models,and being flexible to varying practical objectives,especially when the training and testing data are not identically distributed.A workflow characterized by leveraging Bayesian methodology was proposed to address these issues.Employing a Multi-Layer Perceptron(MLP)as the foundational model,this approach was benchmarked against empirical methods and advanced algorithms for its efficacy in simplicity,accuracy,and resistance to overfitting.The analysis revealed that,while MLP models optimized via maximum a posteriori algorithm suffices for straightforward scenarios,Bayesian neural networks showed great potential for preventing overfitting.Additionally,integrating decision thresholds through various evaluative principles offers insights for challenging decisions.Two case studies demonstrate the framework's capacity for nuanced interpretation of in situ data,employing a model committee for a detailed evaluation of liquefaction potential via Monte Carlo simulations and basic statistics.Overall,the proposed step-by-step workflow for analyzing seismic liquefaction incorporates multifold testing and real-world data validation,showing improved robustness against overfitting and greater versatility in addressing practical challenges.This research contributes to the seismic liquefaction assessment field by providing a structured,adaptable methodology for accurate and reliable analysis.展开更多
Hydraulic fracturing technology has achieved remarkable results in improving the production of tight gas reservoirs,but its effectiveness is under the joint action of multiple factors of complexity.Traditional analysi...Hydraulic fracturing technology has achieved remarkable results in improving the production of tight gas reservoirs,but its effectiveness is under the joint action of multiple factors of complexity.Traditional analysis methods have limitations in dealing with these complex and interrelated factors,and it is difficult to fully reveal the actual contribution of each factor to the production.Machine learning-based methods explore the complex mapping relationships between large amounts of data to provide datadriven insights into the key factors driving production.In this study,a data-driven PCA-RF-VIM(Principal Component Analysis-Random Forest-Variable Importance Measures)approach of analyzing the importance of features is proposed to identify the key factors driving post-fracturing production.Four types of parameters,including log parameters,geological and reservoir physical parameters,hydraulic fracturing design parameters,and reservoir stimulation parameters,were inputted into the PCA-RF-VIM model.The model was trained using 6-fold cross-validation and grid search,and the relative importance ranking of each factor was finally obtained.In order to verify the validity of the PCA-RF-VIM model,a consolidation model that uses three other independent data-driven methods(Pearson correlation coefficient,RF feature significance analysis method,and XGboost feature significance analysis method)are applied to compare with the PCA-RF-VIM model.A comparison the two models shows that they contain almost the same parameters in the top ten,with only minor differences in one parameter.In combination with the reservoir characteristics,the reasonableness of the PCA-RF-VIM model is verified,and the importance ranking of the parameters by this method is more consistent with the reservoir characteristics of the study area.Ultimately,the ten parameters are selected as the controlling factors that have the potential to influence post-fracturing gas production,as the combined importance of these top ten parameters is 91.95%on driving natural gas production.Analyzing and obtaining these ten controlling factors provides engineers with a new insight into the reservoir selection for fracturing stimulation and fracturing parameter optimization to improve fracturing efficiency and productivity.展开更多
基金National Natural Science Foundation of China(Project No.:12371428)Projects of the Provincial College Students’Innovation and Training Program in 2024(Project No.:S202413023106,S202413023110)。
文摘This paper focuses on the numerical solution of a tumor growth model under a data-driven approach.Based on the inherent laws of the data and reasonable assumptions,an ordinary differential equation model for tumor growth is established.Nonlinear fitting is employed to obtain the optimal parameter estimation of the mathematical model,and the numerical solution is carried out using the Matlab software.By comparing the clinical data with the simulation results,a good agreement is achieved,which verifies the rationality and feasibility of the model.
文摘Permanent magnet synchronous motor(PMSM)is widely used in alternating current servo systems as it provides high eficiency,high power density,and a wide speed regulation range.The servo system is placing higher demands on its control performance.The model predictive control(MPC)algorithm is emerging as a potential high-performance motor control algorithm due to its capability of handling multiple-input and multipleoutput variables and imposed constraints.For the MPC used in the PMSM control process,there is a nonlinear disturbance caused by the change of electromagnetic parameters or load disturbance that may lead to a mismatch between the nominal model and the controlled object,which causes the prediction error and thus affects the dynamic stability of the control system.This paper proposes a data-driven MPC strategy in which the historical data in an appropriate range are utilized to eliminate the impact of parameter mismatch and further improve the control performance.The stability of the proposed algorithm is proved as the simulation demonstrates the feasibility.Compared with the classical MPC strategy,the superiority of the algorithm has also been verified.
文摘This paper aims to develop Machine Learning algorithms to classify electronic articles related to this phenomenon by retrieving information and topic modelling.The Methodology of this study is categorized into three phases:the Text Classification Approach(TCA),the Proposed Algorithms Interpretation(PAI),andfinally,Information Retrieval Approach(IRA).The TCA reflects the text preprocessing pipeline called a clean corpus.The Global Vec-tors for Word Representation(Glove)pre-trained model,FastText,Term Frequency-Inverse Document Fre-quency(TF-IDF),and Bag-of-Words(BOW)for extracting the features have been interpreted in this research.The PAI manifests the Bidirectional Long Short-Term Memory(Bi-LSTM)and Convolutional Neural Network(CNN)to classify the COVID-19 news.Again,the IRA explains the mathematical interpretation of Latent Dirich-let Allocation(LDA),obtained for modelling the topic of Information Retrieval(IR).In this study,99%accuracy was obtained by performing K-fold cross-validation on Bi-LSTM with Glove.A comparative analysis between Deep Learning and Machine Learning based on feature extraction and computational complexity exploration has been performed in this research.Furthermore,some text analyses and the most influential aspects of each document have been explored in this study.We have utilized Bidirectional Encoder Representations from Trans-formers(BERT)as a Deep Learning mechanism in our model training,but the result has not been uncovered satisfactory.However,the proposed system can be adjustable in the real-time news classification of COVID-19.
基金funded by the National Natural Science Foundation of China(42374014,42004014).
文摘Due to the signal reflection and diffraction,site-specific unmodeled errors like multipath effect and Non-Line-of-Sight reception are significant error sources in Global Navigation Satellite System since they cannot be easily mitigated.However,how to characterize and model the internal mechanisms and external influences of these site-specific unmodeled errors are still to be investigated.Therefore,we propose a method for characterizing and modeling site-specific unmodeled errors under reflection and diffraction using a data-driven approach.Specifically,we first consider all the popular potential features,which generate the site-specific unmodeled errors.We then use the random forest regression to comprehensively analyze the correlations between the site-specific unmodeled errors and the potential features.We finally characterize and model the site-specific unmodeled errors.Two 7-consecutive datasets dominated by signal reflection and diffraction were conducted.The results show that there are significant differences in the correlations with potential features.They are highly related to the application scenarios,observation types,and satellite types.Notably,the innovation vector often shows a strong correlation with the code site-specific unmodeled errors.For the phase site-specific unmodeled errors,they have high correlations with elevation,azimuth,number of visible satellites,and between-frequency differenced phase observations.In the environments of reflection and diffraction,the sum of the correlations of the top six potential features can reach approximately 88.5 and 87.7%,respectively.Meanwhile,these correlations are stable for different observation types and satellite types.With the integration of a transformer model with the random forest method,a high-precision unmodeled error prediction model is established,demonstrating the necessity to include multiple features for accurate and efficient characterization and modeling of site-specific unmodeled errors.
基金supported in part by Major Project of the National Social Science Fund of China,under Grant No.23&ZD050in part by National Natural Science Foundation of China(NSFC),under Grant Nos.72402031 and 71971052+2 种基金in part by Open Project Program of State Key Laboratory of Massive Personalized Customization System and Technology,under Grant No.H&C-MPC-2023-04-03in part by the Fundamental Research Funds for the Central Universities,under Grant No.N25ZJL015the Joint Funds of the Natural Science Foundation of Liaoning,under Grant No.2023-BSBA-139.
文摘Manufacturers are striving to achieve higher energy efficiency without compromising production performance and quality standards.Parallel-serial structures,commonly found in modern production systems,offer a unique balance of flexibility and efficiency by combining parallel processes with sequential workflows.However,their inherent complexity poses significant challenges,particularly in optimizing energy efficiency and ensuring consistent product quality.In data-driven manufacturing environments,it is not clear how to leverage production data to enhance the energy efficiency of production systems.Therefore,this paper studied a data-driven approach to improving energy efficiency in parallel-serial production lines with product quality issues.Firstly,the authors developed a data-driven performance analysis method to evaluate the effects of disruption events,such as energy-saving control actions,machine breakdowns,and product quality failures,on system throughput and energy consumption.Secondly,a periodic energy-saving control method was developed to enhance system energy efficiency using a non-linear programming model.To reduce complexity and improve computational efficiency,the model was simplified by leveraging the intrinsic properties of parallel-serial production lines and solved using an adaptive genetic algorithm.Finally,the effectiveness of the proposed data-driven approach was validated through case studies,providing actionable insights into achieving data-driven energy efficiency optimization in complex production systems.
文摘Building integrated energy systems(BIESs)are pivotal for enhancing energy efficiency by accounting for a significant proportion of global energy consumption.Two key barriers that reduce the BIES operational efficiency mainly lie in the renewable generation uncertainty and operational non-convexity of combined heat and power(CHP)units.To this end,this paper proposes a soft actor-critic(SAC)algorithm to solve the scheduling problem of BIES,which overcomes the model non-convexity and shows advantages in robustness and generalization.This paper also adopts a temporal fusion transformer(TFT)to enhance the optimal solution for the SAC algorithm by forecasting the renewable generation and energy demand.The TFT can effectively capture the complex temporal patterns and dependencies that span multiple steps.Furthermore,its forecasting results are interpretable due to the employment of a self-attention layer so as to assist in more trustworthy decision-making in the SAC algorithm.The proposed hybrid data-driven approach integrating TFT and SAC algorithm,i.e.,TFT-SAC approach,is trained and tested on a real-world dataset to validate its superior performance in reducing the energy cost and computational time compared with the benchmark approaches.The generalization performance for the scheduling policy,as well as the sensitivity analysis,are examined in the case studies.
文摘Numerous influence factors will lead to the inaccurate prediction of fatigue crack propagation(FCP)life of the metal structure based on the existing FCP model,while the prediction method based on machine learning(ML)and data-driven approach can provide a new idea for accurately predicting the FCP life of the metal structure.In response to the inconvenience of the online prediction method and the inaccu-racy of the offline prediction method,an improved offline prediction method based on data feedback is presented in this paper.FCP tests of reduced scale models of balcony opening corners in a cruise ship are conducted to obtain experimental data with respect to the a-N curves.The crack length corresponding to the cycle is trained using a support vector regression(SVR)and back propagation neural network(BP NN)algorithms.FCP prediction lives of test specimens are performed according to the online,offline,and improved offline prediction methods.Effects of the number of feedback data,the sequence length(SL)in the input set,and the cycle interval on prediction accuracy are discussed.The generalization ability of the proposed method is validated by comparing the prediction results with the experimental data in the literature.The larger the number of feedback data,the higher the prediction accuracy.The results show that 1/5 and 1/2 feedback data are needed in the SVR and BP NN algorithm with SL is 5,respectively.Furthermore,the SVR algorithm and SL=5 are recommended for FCP life prediction using the improved offline prediction method.
基金his work was funded by the James Watt PhD Scholarship program supported by Heriot-Watt University.
文摘The COVID-19 pandemic has significantly impacted global health,social,and economic situations since its emergence in December 2019.The primary focus of this study is to propose a distinct vaccination policy and assess its impact on controlling COVID-19 transmission in Malaysia using a Bayesian data-driven approach,concentrating on the year 2021.We employ a compartmental Susceptible-Exposed-Infected-Recovered-Vaccinated(SEIRV)model,incorporating a time-varying transmission rate and a data-driven method for its estimation through an Exploratory Data Analysis(EDA)approach.While no vaccine guarantees total immunity against the disease,and vaccine immunity wanes over time,it is critical to include and accurately estimate vaccine efficacy,as well as a constant vaccine immunity decay or wane factor,to better simulate the dynamics of vaccine-induced protection over time.Based on the distribution and effectiveness of vaccines,we integrated a data-driven estimation of vaccine efficacy,calculated at 75%for Malaysia,underscoring the model's realism and relevance to the specific context of the country.The Bayesian inference framework is used to assimilate various data sources and account for underlying uncertainties in model parameters.The model is fitted to real-world data from Malaysia to analyze disease spread trends and evaluate the effectiveness of our proposed vaccination policy.Our findings reveal that this distinct vaccination policy,which emphasizes an accelerated vaccination rate during the initial stages of the program,is highly effective in mitigating the spread of COVID-19 and substantially reducing the pandemic peak and new infections.The study found that vaccinating 57–66%of the population(as opposed to 76%in the real data)with a better vaccination policy such as proposed here is able to significantly reduce the number of new infections and ultimately reduce the costs associated with new infections.The study contributes to the development of a robust and informative representation of COVID-19 transmission and vaccination,offering valuable insights for policymakers on the potential benefits and limitations of different vaccination policies,particularly highlighting the importance of a well-planned and efficient vaccination rollout strategy.While the methodology used in this study is specifically applied to national data from Malaysia,its successful application to local regions within Malaysia,such as Selangor and Johor,indicates its adaptability and potential for broader application.This demonstrates the model's adaptability for policy assessment and improvement across various demographic and epidemiological landscapes,implying its usefulness for similar datasets from various geographical regions.
基金supported by the National Natural Science Foundation of China(No.52272382)the Aeronautical Science Foundation of China(No.20200017051001)the Fundamental Research Funds for the Central Universities,China。
文摘Non-learning based motion and path planning of an Unmanned Aerial Vehicle(UAV)is faced with low computation efficiency,mapping memory occupation and local optimization problems.This article investigates the challenge of quadrotor control using offline reinforcement learning.By establishing a data-driven learning paradigm that operates without real-environment interaction,the proposed workflow offers a safer approach than traditional reinforcement learning,making it particularly suited for UAV control in industrial scenarios.The introduced algorithm evaluates dataset uncertainty and employs a pessimistic estimation to foster offline deep reinforcement learning.Experiments highlight the algorithm's superiority over traditional online reinforcement learning methods,especially when learning from offline datasets.Furthermore,the article emphasizes the importance of a more general behavior policy.In evaluations,the trained policy demonstrated versatility by adeptly navigating diverse obstacles,underscoring its real-world applicability.
基金National Natural Sciences Foundation of China,Grant/Award Numbers:62125302,62203087Sci-Tech Talent Innovation Support Program of Dalian,Grant/Award Number:2022RG03+1 种基金Liaoning Revitalization Talents Program,Grant/Award Number:XLYC2002087Young Elite Scientist Sponsorship Program by CAST,Grant/Award Number:YESS20220018。
文摘The electricity and steam integrated energy systems,which can capture waste heat and improve the overall energy efficiency,have been widely utilised in industrial parks.However,intensive and frequent changes in demands would lead to model parameters with strong time-varying characteristics.This paper proposes a hybrid physics and data-driven framework for online joint state and parameter estimation of steam and electricity integrated energy system.Based on the physical non-linear state space models for the electricity network(EN)and steam heating network(SHN),relevance vector machine is developed to learn parameters'dynamic characteristics with respect to model states,which is embedded with physical models.Then,the online joint state and parameter estimation based on unscented Kalman filter is proposed,which would be learnt recursively to capture the spatiotemporal transient characteristics between electricity and SHNs.The IEEE 39-bus EN and the 29-nodes SHN are employed to verify the effectiveness of the proposed method.The experimental results validate that the pro-posed method can provide a higher estimation accuracy than the state-of-the-art approaches.
基金supported by National Key Research and Development Program (2019YFA0708301)National Natural Science Foundation of China (51974337)+2 种基金the Strategic Cooperation Projects of CNPC and CUPB (ZLZX2020-03)Science and Technology Innovation Fund of CNPC (2021DQ02-0403)Open Fund of Petroleum Exploration and Development Research Institute of CNPC (2022-KFKT-09)
文摘We propose an integrated method of data-driven and mechanism models for well logging formation evaluation,explicitly focusing on predicting reservoir parameters,such as porosity and water saturation.Accurately interpreting these parameters is crucial for effectively exploring and developing oil and gas.However,with the increasing complexity of geological conditions in this industry,there is a growing demand for improved accuracy in reservoir parameter prediction,leading to higher costs associated with manual interpretation.The conventional logging interpretation methods rely on empirical relationships between logging data and reservoir parameters,which suffer from low interpretation efficiency,intense subjectivity,and suitability for ideal conditions.The application of artificial intelligence in the interpretation of logging data provides a new solution to the problems existing in traditional methods.It is expected to improve the accuracy and efficiency of the interpretation.If large and high-quality datasets exist,data-driven models can reveal relationships of arbitrary complexity.Nevertheless,constructing sufficiently large logging datasets with reliable labels remains challenging,making it difficult to apply data-driven models effectively in logging data interpretation.Furthermore,data-driven models often act as“black boxes”without explaining their predictions or ensuring compliance with primary physical constraints.This paper proposes a machine learning method with strong physical constraints by integrating mechanism and data-driven models.Prior knowledge of logging data interpretation is embedded into machine learning regarding network structure,loss function,and optimization algorithm.We employ the Physically Informed Auto-Encoder(PIAE)to predict porosity and water saturation,which can be trained without labeled reservoir parameters using self-supervised learning techniques.This approach effectively achieves automated interpretation and facilitates generalization across diverse datasets.
文摘Cardiovascular diseases(CVDs)remain the leading cause of morbidity and mortality worldwide,necessitating innovative diagnostic and prognostic strategies.Traditional biomarkers like C-reactive protein,uric acid,troponin,and natriuretic peptides play crucial roles in CVD management,yet they are often limited by sensitivity and specificity constraints.This narrative review critically examines the emerging landscape of cardiac biomarkers and advocates for a multiple-marker approach to enhance early detection,prognosis,and risk stratification of CVD.In recent years,several novel biomarkers have shown promise in revolutionizing CVD diagnostics.Gamma-glutamyltransferase,microRNAs,endothelial microparticles,placental growth factor,trimethylamine N-oxide,retinol-binding protein 4,copeptin,heart-type fatty acid-binding protein,galectin-3,growth differentiation factor-15,soluble suppression of tumorigenicity 2,fibroblast growth factor 23,and adrenomedullin have emerged as significant indicators of CV health.These biomarkers provide insights into various pathophysiological processes,such as oxidative stress,endothelial dysfunction,inflammation,metabolic disturbances,and myocardial injury.The integration of these novel biomarkers with traditional ones offers a more comprehensive understanding of CVD mechanisms.This multiple-marker approach can improve diagnostic accuracy,allowing for better risk stratification and more personalized treatment strategies.This review underscores the need for continued research to validate the clinical utility of these biomarkers and their potential incorporation into routine clinical practice.By leveraging the strengths of both traditional and novel biomarkers,precise therapeutic plans can be developed,thereby improving the management and prognosis of patients with CVDs.The ongoing exploration and validation of these biomarkers are crucial for advancing CV care and addressing the limitations of current diagnostic tools.
基金supported by Singapore National Medical Research Council(NMRC)grants,including CS-IRG,HLCA2022(to ZDZ),STaR,OF LCG 000207(to EKT)a Clinical Translational Research Programme in Parkinson's DiseaseDuke-Duke-NUS collaboration pilot grant(to ZDZ)。
文摘The progressive loss of dopaminergic neurons in affected patient brains is one of the pathological features of Parkinson's disease,the second most common human neurodegenerative disease.Although the detailed pathogenesis accounting for dopaminergic neuron degeneration in Parkinson's disease is still unclear,the advancement of stem cell approaches has shown promise for Parkinson's disease research and therapy.The induced pluripotent stem cells have been commonly used to generate dopaminergic neurons,which has provided valuable insights to improve our understanding of Parkinson's disease pathogenesis and contributed to anti-Parkinson's disease therapies.The current review discusses the practical approaches and potential applications of induced pluripotent stem cell techniques for generating and differentiating dopaminergic neurons from induced pluripotent stem cells.The benefits of induced pluripotent stem cell-based research are highlighted.Various dopaminergic neuron differentiation protocols from induced pluripotent stem cells are compared.The emerging three-dimension-based brain organoid models compared with conventional two-dimensional cell culture are evaluated.Finally,limitations,challenges,and future directions of induced pluripotent stem cell–based approaches are analyzed and proposed,which will be significant to the future application of induced pluripotent stem cell-related techniques for Parkinson's disease.
基金Supported by National Research Foundation of Korea,No.NRF-2021S1A5A8062526.
文摘This article provides a comprehensive analysis of the study by Hou et al,focusing on the complex interplay between psychological and physical factors in the postoperative recovery(POR)of patients with perianal diseases.The study sheds light on how illness perception,anxiety,and depression significantly influence recovery outcomes.Hou et al developed a predictive model that demonstrated high accuracy in identifying patients at risk of poor recovery.The article explores the critical role of pre-operative psychological assessment,highlighting the need for mental health support and personalized recovery plans in enhancing POR quality.A multidisciplinary approach,integrating mental health professionals with surgeons,anesthesiologists,and other specialists,is emphasized to ensure comprehensive care for patients.The study’s findings serve as a call to integrate psychological care into surgical practice to optimize outcomes for patients with perianal diseases.
文摘Objective: To explore the effect of Health Action Process Approach (HAPA) theory in patients with type D personality psoriasis. Methods: A total of 66 patients with type D personality psoriasis admitted to the dermatology department of a top-three hospital in Jingzhou City from November 2022 to July 2023 were selected and divided into control group and test group with 33 cases in each group by random number table method. The control group received routine health education, and the experimental group received health education based on the HAPA theory. Chronic disease self-efficacy scale, hospital anxiety and depression scale and skin disease quality of life scale were used to evaluate the effect of intervention. Results: After 3 months of intervention, the scores of self-efficacy in experimental group were higher than those in control group (P P Conclusion: Health education based on the theory of HAPA can enhance the self-efficacy of patients with type D personality psoriasis, relieve negative emotions and improve their quality of life.
基金supported by the National Natural Science Foundation of China(Grant No.12272144).
文摘A data-driven model ofmultiple variable cutting(M-VCUT)level set-based substructure is proposed for the topology optimization of lattice structures.TheM-VCUTlevel setmethod is used to represent substructures,enriching their diversity of configuration while ensuring connectivity.To construct the data-driven model of substructure,a database is prepared by sampling the space of substructures spanned by several substructure prototypes.Then,for each substructure in this database,the stiffness matrix is condensed so that its degrees of freedomare reduced.Thereafter,the data-drivenmodel of substructures is constructed through interpolationwith compactly supported radial basis function(CS-RBF).The inputs of the data-driven model are the design variables of topology optimization,and the outputs are the condensed stiffness matrix and volume of substructures.During the optimization,this data-driven model is used,thus avoiding repeated static condensation that would requiremuch computation time.Several numerical examples are provided to verify the proposed method.
基金supported by National Natural Science Foundation of China(Grant No.52375380)National Key R&D Program of China(Grant No.2022YFB3402200)the Key Project of National Natural Science Foundation of China(Grant No.12032018).
文摘The outstanding comprehensive mechanical properties of newly developed hybrid lattice structures make them useful in engineering applications for bearing multiple mechanical loads.Additive-manufacturing technologies make it possible to fabricate these highly spatially programmable structures and greatly enhance the freedom in their design.However,traditional analytical methods do not sufficiently reflect the actual vibration-damping mechanism of lattice structures and are limited by their high computational cost.In this study,a hybrid lattice structure consisting of various cells was designed based on quasi-static and vibration experiments.Subsequently,a novel parametric design method based on a data-driven approach was developed for hybrid lattices with engineered properties.The response surface method was adopted to define the sensitive optimization target.A prediction model for the lattice geometric parameters and vibration properties was established using a backpropagation neural network.Then,it was integrated into the genetic algorithm to create the optimal hybrid lattice with varying geometric features and the required wide-band vibration-damping characteristics.Validation experiments were conducted,demonstrating that the optimized hybrid lattice can achieve the target properties.In addition,the data-driven parametric design method can reduce computation time and be widely applied to complex structural designs when analytical and empirical solutions are unavailable.
基金funded in part by the National Natural Science Foundation of China under Grant 62401167 and 62192712in part by the Key Laboratory of Marine Environmental Survey Technology and Application,Ministry of Natural Resources,P.R.China under Grant MESTA-2023-B001in part by the Stable Supporting Fund of National Key Laboratory of Underwater Acoustic Technology under Grant JCKYS2022604SSJS007.
文摘The Underwater Acoustic(UWA)channel is bandwidth-constrained and experiences doubly selective fading.It is challenging to acquire perfect channel knowledge for Orthogonal Frequency Division Multiplexing(OFDM)communications using a finite number of pilots.On the other hand,Deep Learning(DL)approaches have been very successful in wireless OFDM communications.However,whether they will work underwater is still a mystery.For the first time,this paper compares two categories of DL-based UWA OFDM receivers:the DataDriven(DD)method,which performs as an end-to-end black box,and the Model-Driven(MD)method,also known as the model-based data-driven method,which combines DL and expert OFDM receiver knowledge.The encoder-decoder framework and Convolutional Neural Network(CNN)structure are employed to establish the DD receiver.On the other hand,an unfolding-based Minimum Mean Square Error(MMSE)structure is adopted for the MD receiver.We analyze the characteristics of different receivers by Monte Carlo simulations under diverse communications conditions and propose a strategy for selecting a proper receiver under different communication scenarios.Field trials in the pool and sea are also conducted to verify the feasibility and advantages of the DL receivers.It is observed that DL receivers perform better than conventional receivers in terms of bit error rate.
文摘When assessing seismic liquefaction potential with data-driven models,addressing the uncertainties of establishing models,interpreting cone penetration tests(CPT)data and decision threshold is crucial for avoiding biased data selection,ameliorating overconfident models,and being flexible to varying practical objectives,especially when the training and testing data are not identically distributed.A workflow characterized by leveraging Bayesian methodology was proposed to address these issues.Employing a Multi-Layer Perceptron(MLP)as the foundational model,this approach was benchmarked against empirical methods and advanced algorithms for its efficacy in simplicity,accuracy,and resistance to overfitting.The analysis revealed that,while MLP models optimized via maximum a posteriori algorithm suffices for straightforward scenarios,Bayesian neural networks showed great potential for preventing overfitting.Additionally,integrating decision thresholds through various evaluative principles offers insights for challenging decisions.Two case studies demonstrate the framework's capacity for nuanced interpretation of in situ data,employing a model committee for a detailed evaluation of liquefaction potential via Monte Carlo simulations and basic statistics.Overall,the proposed step-by-step workflow for analyzing seismic liquefaction incorporates multifold testing and real-world data validation,showing improved robustness against overfitting and greater versatility in addressing practical challenges.This research contributes to the seismic liquefaction assessment field by providing a structured,adaptable methodology for accurate and reliable analysis.
基金funded by the Key Research and Development Program of Shaanxi,China(No.2024GX-YBXM-503)the National Natural Science Foundation of China(No.51974254)。
文摘Hydraulic fracturing technology has achieved remarkable results in improving the production of tight gas reservoirs,but its effectiveness is under the joint action of multiple factors of complexity.Traditional analysis methods have limitations in dealing with these complex and interrelated factors,and it is difficult to fully reveal the actual contribution of each factor to the production.Machine learning-based methods explore the complex mapping relationships between large amounts of data to provide datadriven insights into the key factors driving production.In this study,a data-driven PCA-RF-VIM(Principal Component Analysis-Random Forest-Variable Importance Measures)approach of analyzing the importance of features is proposed to identify the key factors driving post-fracturing production.Four types of parameters,including log parameters,geological and reservoir physical parameters,hydraulic fracturing design parameters,and reservoir stimulation parameters,were inputted into the PCA-RF-VIM model.The model was trained using 6-fold cross-validation and grid search,and the relative importance ranking of each factor was finally obtained.In order to verify the validity of the PCA-RF-VIM model,a consolidation model that uses three other independent data-driven methods(Pearson correlation coefficient,RF feature significance analysis method,and XGboost feature significance analysis method)are applied to compare with the PCA-RF-VIM model.A comparison the two models shows that they contain almost the same parameters in the top ten,with only minor differences in one parameter.In combination with the reservoir characteristics,the reasonableness of the PCA-RF-VIM model is verified,and the importance ranking of the parameters by this method is more consistent with the reservoir characteristics of the study area.Ultimately,the ten parameters are selected as the controlling factors that have the potential to influence post-fracturing gas production,as the combined importance of these top ten parameters is 91.95%on driving natural gas production.Analyzing and obtaining these ten controlling factors provides engineers with a new insight into the reservoir selection for fracturing stimulation and fracturing parameter optimization to improve fracturing efficiency and productivity.