The bandgap is a key parameter for understanding and designing hybrid perovskite material properties,as well as developing photovoltaic devices.Traditional bandgap calculation methods like ultravioletvisible spectrosc...The bandgap is a key parameter for understanding and designing hybrid perovskite material properties,as well as developing photovoltaic devices.Traditional bandgap calculation methods like ultravioletvisible spectroscopy and first-principles calculations are time-and power-consuming,not to mention capturing bandgap change mechanisms for hybrid perovskite materials across a wide range of unknown space.In the present work,an artificial intelligence ensemble comprising two classifiers(with F1 scores of 0.9125 and 0.925)and a regressor(with mean squared error of 0.0014 eV)is constructed to achieve high-precision prediction of the bandgap.The bandgap perovskite dataset is established through highthroughput prediction of bandgaps by the ensemble.Based on the self-built dataset,partial dependence analysis(PDA)is developed to interpret the bandgap influential mechanism.Meanwhile,an interpretable mathematical model with an R^(2)of 0.8417 is generated using the genetic programming symbolic regression(GPSR)technique.The constructed PDA maps agree well with the Shapley Additive exPlanations,the GPSR model,and experiment verification.Through PDA,we reveal the boundary effect,the bowing effect,and their evolution trends with key descriptors.展开更多
This paper proposes a new approach for online power system transient security assessment(TSA)and preventive control based on XGBoost and DC optimal power flow(DCOPF).The novelty of this proposal is that it applies the...This paper proposes a new approach for online power system transient security assessment(TSA)and preventive control based on XGBoost and DC optimal power flow(DCOPF).The novelty of this proposal is that it applies the XGBoost and data selection method based on the 1-norm distance in local feature importance evaluation which can provide a certain model interpretability.The method of SMOTE+ENN is adopted for data rebalancing.The contingency-oriented XGBoost model is trained with databases generated by time domain simulations to represent the transient security constraint in the DCOPF model,which has a relatively fast speed of calculation.The transient security constrained generation rescheduling is implemented with the differential evolution algorithm,which is utilized to optimize the rescheduled generation in the preventive control.Feasibility and effectiveness of the proposed approach are demonstrated on an IEEE 39-bus test system and a 500-bus operational model for South Carolina,USA.展开更多
Ceramic relief mural is a contemporary landscape art that is carefully designed based on human nature,culture,and architectural wall space,combined with social customs,visual sensibility,and art.It may also become the...Ceramic relief mural is a contemporary landscape art that is carefully designed based on human nature,culture,and architectural wall space,combined with social customs,visual sensibility,and art.It may also become the main axis of ceramic art in the future.Taiwan public ceramic relief murals(PCRM)are most distinctive with the PCRM pioneered by Pan-Hsiung Chu of Meinong Kiln in 1987.In addition to breaking through the limitations of traditional public ceramic murals,Chu leveraged local culture and sensibility.The theme of art gives PCRM its unique style and innovative value throughout the Taiwan region.This study mainly analyzes and understands the design image of public ceramic murals,taking Taiwan PCRM’s design and creation as the scope,and applies STEEP analysis,that is,the social,technological,economic,ecological,and political-legal environments are analyzed as core factors;eight main important factors in the artistic design image of ceramic murals are evaluated.Then,interpretive structural modeling(ISM)is used to establish five levels,analyze the four main problems in the main core factor area and the four main target results in the affected factor area;and analyze the problem points and target points as well as their causal relationships.It is expected to sort out the relationship between these factors,obtain the hierarchical relationship of each factor,and provide a reference basis and research methods.展开更多
Artificial intelligence(AI)has emerged as a transformative technology in accelerating drug discovery and development within natural medicines research.Natural medicines,characterized by their complex chemical composit...Artificial intelligence(AI)has emerged as a transformative technology in accelerating drug discovery and development within natural medicines research.Natural medicines,characterized by their complex chemical compositions and multifaceted pharmacological mechanisms,demonstrate widespread application in treating diverse diseases.However,research and development face significant challenges,including component complexity,extraction difficulties,and efficacy validation.AI technology,particularly through deep learning(DL)and machine learning(ML)approaches,enables efficient analysis of extensive datasets,facilitating drug screening,component analysis,and pharmacological mechanism elucidation.The implementation of AI technology demonstrates considerable potential in virtual screening,compound optimization,and synthetic pathway design,thereby enhancing natural medicines’bioavailability and safety profiles.Nevertheless,current applications encounter limitations regarding data quality,model interpretability,and ethical considerations.As AI technologies continue to evolve,natural medicines research and development will achieve greater efficiency and precision,advancing both personalized medicine and contemporary drug development approaches.展开更多
Given the growing concern over global warming and the critical role of carbon dioxide(CO_(2))in this phenomenon,the study of CO_(2)-induced alterations in coal strength has garnered significant attention due to its im...Given the growing concern over global warming and the critical role of carbon dioxide(CO_(2))in this phenomenon,the study of CO_(2)-induced alterations in coal strength has garnered significant attention due to its implications for carbon sequestration.A large number of experiments have proved that CO_(2) interaction time(T),saturation pressure(P)and other parameters have significant effects on coal strength.However,accurate evaluation of CO_(2)-induced alterations in coal strength is still a difficult problem,so it is particularly important to establish accurate and efficient prediction models.This study explored the application of advancedmachine learning(ML)algorithms and Gene Expression Programming(GEP)techniques to predict CO_(2)-induced alterations in coal strength.Sixmodels were developed,including three metaheuristic-optimized XGBoost models(GWO-XGBoost,SSA-XGBoost,PO-XGBoost)and three GEP models(GEP-1,GEP-2,GEP-3).Comprehensive evaluations using multiple metrics revealed that all models demonstrated high predictive accuracy,with the SSA-XGBoost model achieving the best performance(R2—Coefficient of determination=0.99396,RMSE—Root Mean Square Error=0.62102,MAE—Mean Absolute Error=0.36164,MAPE—Mean Absolute Percentage Error=4.8101%,RPD—Residual Predictive Deviation=13.4741).Model interpretability analyses using SHAP(Shapley Additive exPlanations),ICE(Individual Conditional Expectation),and PDP(Partial Dependence Plot)techniques highlighted the dominant role of fixed carbon content(FC)and significant interactions between FC and CO_(2) saturation pressure(P).Theresults demonstrated that the proposedmodels effectively address the challenges of CO_(2)-induced strength prediction,providing valuable insights for geological storage safety and environmental applications.展开更多
This study introduces a comprehensive and automated framework that leverages data-driven method-ologies to address various challenges in shale gas development and production.Specifically,it harnesses the power of Auto...This study introduces a comprehensive and automated framework that leverages data-driven method-ologies to address various challenges in shale gas development and production.Specifically,it harnesses the power of Automated Machine Learning(AutoML)to construct an ensemble model to predict the estimated ultimate recovery(EUR)of shale gas wells.To demystify the“black-box”nature of the ensemble model,KernelSHAP,a kernel-based approach to compute Shapley values,is utilized for elucidating the influential factors that affect shale gas production at both global and local scales.Furthermore,a bi-objective optimization algorithm named NSGA-Ⅱ is seamlessly incorporated to opti-mize hydraulic fracturing designs for production boost and cost control.This innovative framework addresses critical limitations often encountered in applying machine learning(ML)to shale gas pro-duction:the challenge of achieving sufficient model accuracy with limited samples,the multidisciplinary expertise required for developing robust ML models,and the need for interpretability in“black-box”models.Validation with field data from the Fuling shale gas field in the Sichuan Basin substantiates the framework's efficacy in enhancing the precision and applicability of data-driven techniques.The test accuracy of the ensemble ML model reached 83%compared to a maximum of 72%of single ML models.The contribution of each geological and engineering factor to the overall production was quantitatively evaluated.Fracturing design optimization raised EUR by 7%-34%under different production and cost tradeoff scenarios.The results empower domain experts to conduct more precise and objective data-driven analyses and optimizations for shale gas production with minimal expertise in data science.展开更多
Formation pore pressure is the foundation of well plan,and it is related to the safety and efficiency of drilling operations in oil and gas development.However,the traditional method for predicting formation pore pres...Formation pore pressure is the foundation of well plan,and it is related to the safety and efficiency of drilling operations in oil and gas development.However,the traditional method for predicting formation pore pressure involves applying post-drilling measurement data from nearby wells to the target well,which may not accurately reflect the formation pore pressure of the target well.In this paper,a novel method for predicting formation pore pressure ahead of the drill bit by embedding petrophysical theory into machine learning based on seismic and logging-while-drilling(LWD)data was proposed.Gated recurrent unit(GRU)and long short-term memory(LSTM)models were developed and validated using data from three wells in the Bohai Oilfield,and the Shapley additive explanations(SHAP)were utilized to visualize and interpret the models proposed in this study,thereby providing valuable insights into the relative importance and impact of input features.The results show that among the eight models trained in this study,almost all model prediction errors converge to 0.05 g/cm^(3),with the largest root mean square error(RMSE)being 0.03072 and the smallest RMSE being 0.008964.Moreover,continuously updating the model with the increasing training data during drilling operations can further improve accuracy.Compared to other approaches,this study accurately and precisely depicts formation pore pressure,while SHAP analysis guides effective model refinement and feature engineering strategies.This work underscores the potential of integrating advanced machine learning techniques with domain-specific knowledge to enhance predictive accuracy for petroleum engineering applications.展开更多
Characterized by self-monitoring and agile adaptation to fast changing dynamics in complex production environments,smart manufacturing as envisioned under Industry 4.0 aims to improve the throughput and reliability of...Characterized by self-monitoring and agile adaptation to fast changing dynamics in complex production environments,smart manufacturing as envisioned under Industry 4.0 aims to improve the throughput and reliability of production beyond the state-of-the-art.While the widespread application of deep learning(DL)has opened up new opportunities to accomplish the goal,data quality and model interpretability have continued to present a roadblock for the widespread acceptance of DL for real-world applications.This has motivated research on two fronts:data curation,which aims to provide quality data as input for meaningful DL-based analysis,and model interpretation,which intends to reveal the physical reasoning underlying DL model outputs and promote trust from the users.This paper summarizes several key techniques in data curation where breakthroughs in data denoising,outlier detection,imputation,balancing,and semantic annotation have demonstrated the effectiveness in information extraction from noisy,incomplete,insufficient,and/or unannotated data.Also highlighted are model interpretation methods that address the“black-box”nature of DL towards model transparency.展开更多
Nonlinear characteristic fault detection and diagnosis method based on higher-order statistical(HOS) is an effective data-driven method, but the calculation costs much for a large-scale process control system. An HOS-...Nonlinear characteristic fault detection and diagnosis method based on higher-order statistical(HOS) is an effective data-driven method, but the calculation costs much for a large-scale process control system. An HOS-ISM fault diagnosis framework combining interpretative structural model(ISM) and HOS is proposed:(1) the adjacency matrix is determined by partial correlation coefficient;(2) the modified adjacency matrix is defined by directed graph with prior knowledge of process piping and instrument diagram;(3) interpretative structural for large-scale process control system is built by this ISM method; and(4) non-Gaussianity index, nonlinearity index, and total nonlinearity index are calculated dynamically based on interpretative structural to effectively eliminate uncertainty of the nonlinear characteristic diagnostic method with reasonable sampling period and data window. The proposed HOS-ISM fault diagnosis framework is verified by the Tennessee Eastman process and presents improvement for highly non-linear characteristic for selected fault cases.展开更多
For ecological restoration and reconstruction of the degraded area, it is an important premise to correctly understand the degradation factors of the ecosystem in the arid-hot valleys. The factors including vegetation...For ecological restoration and reconstruction of the degraded area, it is an important premise to correctly understand the degradation factors of the ecosystem in the arid-hot valleys. The factors including vegetation degradation, land degradation, arid climate, policy failure, forest fire, rapid population growth, excessive deforestation, overgrazing, steep slope reclamation, economic poverty, engineering construction, lithology, slope, low cultural level, geological hazards, biological disaster, soil properties etc, were selected to study the Yuanmou arid-hot valleys. Based on the interpretative structural model (ISM), it has found out that the degradation factors of the Yuanmou arid-hot valleys were not at the same level but in a multilevel hierarchical system with internal relations, which pointed out that the degradation mode of the arid-hot valleys was "straight (appearance)-penetrating-background". Such researches have important directive significance for the restoration and reconstruction of the arid-hot valleys ecosystem.展开更多
The distributions of local velocity and local phase holdup along the radial direction of pipes are complicated because of gravity differentiation,and the distribution of fluid velocity fi eld changes along the gravity...The distributions of local velocity and local phase holdup along the radial direction of pipes are complicated because of gravity differentiation,and the distribution of fluid velocity fi eld changes along the gravity direction in horizontal wells.Therefore,measuring the mixture flow and water holdup is difficult,resulting in poor interpretation accuracy of the production logging output profile.In this paper,oil–water two-phase flow dynamic simulation logging experiments in horizontal oil–water two-phase fl ow simulation wells were conducted using the Multiple Array Production Suite,which comprises a capacitance array tool(CAT)and a spinner array tool(SAT),and then the response characteristics of SAT and CAT in diff erent fl ow rates and water cut production conditions were studied.According to the response characteristics of CAT in diff erent water holdup ranges,interpolation imaging along the wellbore section determines the water holdup distribution,and then,the oil–water two-phase velocity fi eld in the fl ow section was reconstructed on the basis of the fl ow section water holdup distribution and the logging value of SAT and combined with the rheological equation of viscous fl uid,and the calculation method of the oil–water partial phase fl ow rate in the fl ow section was proposed.This new approach was applied in the experiment data calculations,and the results are basically consistent with the experimental data.The total fl ow rate and water holdup from the calculation are in agreement with the set values in the experiment,suggesting that the method has high accuracy.展开更多
Alarm flood is one of the main problems in the alarm systems of industrial process. Alarm root-cause analysis and alarm prioritization are good for alarm flood reduction. This paper proposes a systematic rationalizati...Alarm flood is one of the main problems in the alarm systems of industrial process. Alarm root-cause analysis and alarm prioritization are good for alarm flood reduction. This paper proposes a systematic rationalization method for multivariate correlated alarms to realize the root cause analysis and alarm prioritization. An information fusion based interpretive structural model is constructed according to the data-driven partial correlation coefficient calculation and process knowledge modification. This hierarchical multi-layer model is helpful in abnormality propagation path identification and root-cause analysis. Revised Likert scale method is adopted to determine the alarm priority and reduce the blindness of alarm handling. As a case study, the Tennessee Eastman process is utilized to show the effectiveness and validity of proposed approach. Alarm system performance comparison shows that our rationalization methodology can reduce the alarm flood to some extent and improve the performance.展开更多
In complex media, especially for seismic prospecting in deep layers in East China and in the mountainous area in West China, due to the complex geological condition, the common-mid-point (CMP) gather of deep reflect...In complex media, especially for seismic prospecting in deep layers in East China and in the mountainous area in West China, due to the complex geological condition, the common-mid-point (CMP) gather of deep reflection event is neither hyperbolic, nor any simple function. If traditional normal move-out (NMO) and stack imaging technology are still used, it is difficult to get a clear stack image. Based on previous techniques on non-hyperbolic stack, it is thought in this paper that no matter how complex the geological condition is, in order to get an optimized stack image, the stack should be non time move-out stack, and any stacking method limited to some kind of curve will be restricted to application conditions. In order to overcome the above-mentioned limit, a new method called optimized non-hyperbolic stack imaging based on interpretation model is presented in this paper. Based on CMP/CRP (Common-Reflection-Point) gather after NMO or pre-stack migration, this method uses the interpretation model of reflectors as constraint, and takes comparability as a distinguishing criterion, and finally forms a residual move-out correction for the gather of constrained model. Numerical simulation indicates that this method could overcome the non hyperbolic problem and get fine stack image.展开更多
Traditional visual interpretation is often inefficient due to its excessively workload professional knowledge and strong subjectivity.Therefore,building an automatic interpretation model on high spatial resolution rem...Traditional visual interpretation is often inefficient due to its excessively workload professional knowledge and strong subjectivity.Therefore,building an automatic interpretation model on high spatial resolution remote sensing images is the key to the quick and efficient interpretation of earthquake-triggered landslides.Aiming at addressing this problem,a landslide interpretation model of high-resolution images based on bag of visual word(BoVW)feature was proposed.The high-resolution images were pre-processed,and then BoVW feature and support vector machine(SVM)was adopted to establish an automatic landslide interpretation model.This model was further compared with the currently widely used Histogram of Oriented Gradient(HoG)feature extraction model.In order to test the effectiveness of the method,typical landslide images were selected to construct a landslide sample library,which was subsequently utilized as the foundation for conducting an experimental study.The results show that the accuracy of landslide extraction using this method reaches as high as 89%,indicating that the method can be used for the automatic interpretation of landslides in disaster-prone areas,and has high practical value for regional disaster prevention and damage reduction.展开更多
In this paper, the structure characteristics of open complex giant systems are concretely analysed in depth, thus the view and its significance to support the meta synthesis engineering with manifold knowledge models...In this paper, the structure characteristics of open complex giant systems are concretely analysed in depth, thus the view and its significance to support the meta synthesis engineering with manifold knowledge models are clarified. Furthermore, the knowledge based multifaceted modeling methodology for open complex giant systems is emphatically studied. The major points are as follows: (1) nonlinear mechanism and general information partition law; (2) from the symmetry and similarity to the acquisition of construction knowledge; (3) structures for hierarchical and nonhierarchical organizations; (4) the integration of manifold knowledge models; (5) the methodology of knowledge based multifaceted modeling.展开更多
Interpretative structural model(ISM) can transform a multivariate problem into several sub-variable problems to analyze a complex industrial structure in a more efficient way by building a multi-level hierarchical str...Interpretative structural model(ISM) can transform a multivariate problem into several sub-variable problems to analyze a complex industrial structure in a more efficient way by building a multi-level hierarchical structure model. To build an ISM of a production system, the partial correlation coefficient method is proposed to obtain the adjacency matrix, which can be transformed to ISM. According to estimation of correlation coefficient, the result can give actual variable correlations and eliminate effects of intermediate variables. Furthermore, this paper proposes an effective approach using ISM to analyze the main factors and basic mechanisms that affect the energy consumption in an ethylene production system. The case study shows that the proposed energy consumption analysis method is valid and efficient in improvement of energy efficiency in ethylene production.展开更多
In enterprise operations,maintaining manual rules for enterprise processes can be expensive,time-consuming,and dependent on specialized domain knowledge in that enterprise domain.Recently,rule-generation has been auto...In enterprise operations,maintaining manual rules for enterprise processes can be expensive,time-consuming,and dependent on specialized domain knowledge in that enterprise domain.Recently,rule-generation has been automated in enterprises,particularly through Machine Learning,to streamline routine tasks.Typically,these machine models are black boxes where the reasons for the decisions are not always transparent,and the end users need to verify the model proposals as a part of the user acceptance testing to trust it.In such scenarios,rules excel over Machine Learning models as the end-users can verify the rules and have more trust.In many scenarios,the truth label changes frequently thus,it becomes difficult for the Machine Learning model to learn till a considerable amount of data has been accumulated,but with rules,the truth can be adapted.This paper presents a novel framework for generating human-understandable rules using the Classification and Regression Tree(CART)decision tree method,which ensures both optimization and user trust in automated decision-making processes.The framework generates comprehensible rules in the form of if condition and then predicts class even in domains where noise is present.The proposed system transforms enterprise operations by automating the production of human-readable rules from structured data,resulting in increased efficiency and transparency.Removing the need for human rule construction saves time and money while guaranteeing that users can readily check and trust the automatic judgments of the system.The remarkable performance metrics of the framework,which achieve 99.85%accuracy and 96.30%precision,further support its efficiency in translating complex data into comprehensible rules,eventually empowering users and enhancing organizational decision-making processes.展开更多
To the Editor,We commend Chen et al[1]for their insightful exploration of the challenges associated with the evaluation of large language models(LLMs)and agents in clinical applications.Their work contributes signific...To the Editor,We commend Chen et al[1]for their insightful exploration of the challenges associated with the evaluation of large language models(LLMs)and agents in clinical applications.Their work contributes significantly to the discourse on integrating artificial intelligence into healthcare,highlighting critical issues such as data privacy,model interpretability,and risk of misinformation.展开更多
With the rapid development of the Internet,network security and data privacy are increasingly valued.Although classical Network Intrusion Detection System(NIDS)based on Deep Learning(DL)models can provide good detecti...With the rapid development of the Internet,network security and data privacy are increasingly valued.Although classical Network Intrusion Detection System(NIDS)based on Deep Learning(DL)models can provide good detection accuracy,but collecting samples for centralized training brings the huge risk of data privacy leakage.Furthermore,the training of supervised deep learning models requires a large number of labeled samples,which is usually cumbersome.The“black-box”problem also makes the DL models of NIDS untrustworthy.In this paper,we propose a trusted Federated Learning(FL)Traffic IDS method called FL-TIDS to address the above-mentioned problems.In FL-TIDS,we design an unsupervised intrusion detection model based on autoencoders that alleviates the reliance on marked samples.At the same time,we use FL for model training to protect data privacy.In addition,we design an improved SHAP interpretable method based on chi-square test to perform interpretable analysis of the trained model.We conducted several experiments to evaluate the proposed FL-TIDS.We first determine experimentally the structure and the number of neurons of the unsupervised AE model.Secondly,we evaluated the proposed method using the UNSW-NB15 and CICIDS2017 datasets.The exper-imental results show that the unsupervised AE model has better performance than the other 7 intrusion detection models in terms of precision,recall and f1-score.Then,federated learning is used to train the intrusion detection model.The experimental results indicate that the model is more accurate than the local learning model.Finally,we use an improved SHAP explainability method based on Chi-square test to analyze the explainability.The analysis results show that the identification characteristics of the model are consistent with the attack characteristics,and the model is reliable.展开更多
Cytotoxicity,usually represented by cell viability,is a crucial parameter for evaluating drug safety in vitro.Accurate prediction of cell viability/cytotoxicity could accelerate drug development in the early stage.In ...Cytotoxicity,usually represented by cell viability,is a crucial parameter for evaluating drug safety in vitro.Accurate prediction of cell viability/cytotoxicity could accelerate drug development in the early stage.In this study,by integrating cellular transcriptome and cell viability data using four machine learning algorithms(support vector machine(SVM),random forest(RF),extreme gradient boosting(XGBoost),and light gradient boosting machine(LightGBM))and two ensemble algorithms(voting and stacking),highly accurate prediction models of 50%and 80%cell viability were developed with area under the receiver operating characteristic curve(AUROC)of 0.90 and 0.84,respectively;these models also showed good performance when utilized for diverse cell lines.Concerning the characterization of the employed Feature Genes,the models were interpreted,and the mechanisms of bioactive compounds with a narrow therapeutic index(NTI)can also be analyzed.In summary,the models established in this research exhibit superior capacity to those of previous studies;these models enable accurate high-safety substance screening via cytotoxicity prediction across cell lines.Moreover,for the first time,Cytotoxicity Signature(CTS)genes were identified,which could provide additional clues for further study of mechanisms of action(MOA),especially for NTI compounds.展开更多
基金supported by the National Research Foundation of Korea(NRF)funded by the Korean government(MSIT)(Grant number:RS-2025-02316700,and RS-2025-00522430)the China Scholarship Council Program。
文摘The bandgap is a key parameter for understanding and designing hybrid perovskite material properties,as well as developing photovoltaic devices.Traditional bandgap calculation methods like ultravioletvisible spectroscopy and first-principles calculations are time-and power-consuming,not to mention capturing bandgap change mechanisms for hybrid perovskite materials across a wide range of unknown space.In the present work,an artificial intelligence ensemble comprising two classifiers(with F1 scores of 0.9125 and 0.925)and a regressor(with mean squared error of 0.0014 eV)is constructed to achieve high-precision prediction of the bandgap.The bandgap perovskite dataset is established through highthroughput prediction of bandgaps by the ensemble.Based on the self-built dataset,partial dependence analysis(PDA)is developed to interpret the bandgap influential mechanism.Meanwhile,an interpretable mathematical model with an R^(2)of 0.8417 is generated using the genetic programming symbolic regression(GPSR)technique.The constructed PDA maps agree well with the Shapley Additive exPlanations,the GPSR model,and experiment verification.Through PDA,we reveal the boundary effect,the bowing effect,and their evolution trends with key descriptors.
基金supported in part by the National Key Research and Development Program of China under Grant 2020YFB0905900.
文摘This paper proposes a new approach for online power system transient security assessment(TSA)and preventive control based on XGBoost and DC optimal power flow(DCOPF).The novelty of this proposal is that it applies the XGBoost and data selection method based on the 1-norm distance in local feature importance evaluation which can provide a certain model interpretability.The method of SMOTE+ENN is adopted for data rebalancing.The contingency-oriented XGBoost model is trained with databases generated by time domain simulations to represent the transient security constraint in the DCOPF model,which has a relatively fast speed of calculation.The transient security constrained generation rescheduling is implemented with the differential evolution algorithm,which is utilized to optimize the rescheduled generation in the preventive control.Feasibility and effectiveness of the proposed approach are demonstrated on an IEEE 39-bus test system and a 500-bus operational model for South Carolina,USA.
文摘Ceramic relief mural is a contemporary landscape art that is carefully designed based on human nature,culture,and architectural wall space,combined with social customs,visual sensibility,and art.It may also become the main axis of ceramic art in the future.Taiwan public ceramic relief murals(PCRM)are most distinctive with the PCRM pioneered by Pan-Hsiung Chu of Meinong Kiln in 1987.In addition to breaking through the limitations of traditional public ceramic murals,Chu leveraged local culture and sensibility.The theme of art gives PCRM its unique style and innovative value throughout the Taiwan region.This study mainly analyzes and understands the design image of public ceramic murals,taking Taiwan PCRM’s design and creation as the scope,and applies STEEP analysis,that is,the social,technological,economic,ecological,and political-legal environments are analyzed as core factors;eight main important factors in the artistic design image of ceramic murals are evaluated.Then,interpretive structural modeling(ISM)is used to establish five levels,analyze the four main problems in the main core factor area and the four main target results in the affected factor area;and analyze the problem points and target points as well as their causal relationships.It is expected to sort out the relationship between these factors,obtain the hierarchical relationship of each factor,and provide a reference basis and research methods.
基金supports from the National Key Research and Development Program of China(No.2020YFE0202200)the National Natural Science Foundation of China(Nos.81903538,82322073,92253303)+1 种基金the Innovation Team and Talents Cultivation Program of National Administration of Traditional Chinese Medicine(No.ZYYCXTD-D-202004)the Science and Technology Commission of Shanghai Municipality(Nos.22ZR1474200,24JS2830200).
文摘Artificial intelligence(AI)has emerged as a transformative technology in accelerating drug discovery and development within natural medicines research.Natural medicines,characterized by their complex chemical compositions and multifaceted pharmacological mechanisms,demonstrate widespread application in treating diverse diseases.However,research and development face significant challenges,including component complexity,extraction difficulties,and efficacy validation.AI technology,particularly through deep learning(DL)and machine learning(ML)approaches,enables efficient analysis of extensive datasets,facilitating drug screening,component analysis,and pharmacological mechanism elucidation.The implementation of AI technology demonstrates considerable potential in virtual screening,compound optimization,and synthetic pathway design,thereby enhancing natural medicines’bioavailability and safety profiles.Nevertheless,current applications encounter limitations regarding data quality,model interpretability,and ethical considerations.As AI technologies continue to evolve,natural medicines research and development will achieve greater efficiency and precision,advancing both personalized medicine and contemporary drug development approaches.
基金partially supported by the National Natural Science Foundation of China(42177164,52474121)the Outstanding Youth Project of Hunan Provincial Department of Education(23B0008).
文摘Given the growing concern over global warming and the critical role of carbon dioxide(CO_(2))in this phenomenon,the study of CO_(2)-induced alterations in coal strength has garnered significant attention due to its implications for carbon sequestration.A large number of experiments have proved that CO_(2) interaction time(T),saturation pressure(P)and other parameters have significant effects on coal strength.However,accurate evaluation of CO_(2)-induced alterations in coal strength is still a difficult problem,so it is particularly important to establish accurate and efficient prediction models.This study explored the application of advancedmachine learning(ML)algorithms and Gene Expression Programming(GEP)techniques to predict CO_(2)-induced alterations in coal strength.Sixmodels were developed,including three metaheuristic-optimized XGBoost models(GWO-XGBoost,SSA-XGBoost,PO-XGBoost)and three GEP models(GEP-1,GEP-2,GEP-3).Comprehensive evaluations using multiple metrics revealed that all models demonstrated high predictive accuracy,with the SSA-XGBoost model achieving the best performance(R2—Coefficient of determination=0.99396,RMSE—Root Mean Square Error=0.62102,MAE—Mean Absolute Error=0.36164,MAPE—Mean Absolute Percentage Error=4.8101%,RPD—Residual Predictive Deviation=13.4741).Model interpretability analyses using SHAP(Shapley Additive exPlanations),ICE(Individual Conditional Expectation),and PDP(Partial Dependence Plot)techniques highlighted the dominant role of fixed carbon content(FC)and significant interactions between FC and CO_(2) saturation pressure(P).Theresults demonstrated that the proposedmodels effectively address the challenges of CO_(2)-induced strength prediction,providing valuable insights for geological storage safety and environmental applications.
基金funded by the National Natural Science Foundation of China(42050104).
文摘This study introduces a comprehensive and automated framework that leverages data-driven method-ologies to address various challenges in shale gas development and production.Specifically,it harnesses the power of Automated Machine Learning(AutoML)to construct an ensemble model to predict the estimated ultimate recovery(EUR)of shale gas wells.To demystify the“black-box”nature of the ensemble model,KernelSHAP,a kernel-based approach to compute Shapley values,is utilized for elucidating the influential factors that affect shale gas production at both global and local scales.Furthermore,a bi-objective optimization algorithm named NSGA-Ⅱ is seamlessly incorporated to opti-mize hydraulic fracturing designs for production boost and cost control.This innovative framework addresses critical limitations often encountered in applying machine learning(ML)to shale gas pro-duction:the challenge of achieving sufficient model accuracy with limited samples,the multidisciplinary expertise required for developing robust ML models,and the need for interpretability in“black-box”models.Validation with field data from the Fuling shale gas field in the Sichuan Basin substantiates the framework's efficacy in enhancing the precision and applicability of data-driven techniques.The test accuracy of the ensemble ML model reached 83%compared to a maximum of 72%of single ML models.The contribution of each geological and engineering factor to the overall production was quantitatively evaluated.Fracturing design optimization raised EUR by 7%-34%under different production and cost tradeoff scenarios.The results empower domain experts to conduct more precise and objective data-driven analyses and optimizations for shale gas production with minimal expertise in data science.
基金supported by the National Natural Science Foundation of China(Grant numbers:52174012,52394250,52394255,52234002,U22B20126,51804322).
文摘Formation pore pressure is the foundation of well plan,and it is related to the safety and efficiency of drilling operations in oil and gas development.However,the traditional method for predicting formation pore pressure involves applying post-drilling measurement data from nearby wells to the target well,which may not accurately reflect the formation pore pressure of the target well.In this paper,a novel method for predicting formation pore pressure ahead of the drill bit by embedding petrophysical theory into machine learning based on seismic and logging-while-drilling(LWD)data was proposed.Gated recurrent unit(GRU)and long short-term memory(LSTM)models were developed and validated using data from three wells in the Bohai Oilfield,and the Shapley additive explanations(SHAP)were utilized to visualize and interpret the models proposed in this study,thereby providing valuable insights into the relative importance and impact of input features.The results show that among the eight models trained in this study,almost all model prediction errors converge to 0.05 g/cm^(3),with the largest root mean square error(RMSE)being 0.03072 and the smallest RMSE being 0.008964.Moreover,continuously updating the model with the increasing training data during drilling operations can further improve accuracy.Compared to other approaches,this study accurately and precisely depicts formation pore pressure,while SHAP analysis guides effective model refinement and feature engineering strategies.This work underscores the potential of integrating advanced machine learning techniques with domain-specific knowledge to enhance predictive accuracy for petroleum engineering applications.
文摘Characterized by self-monitoring and agile adaptation to fast changing dynamics in complex production environments,smart manufacturing as envisioned under Industry 4.0 aims to improve the throughput and reliability of production beyond the state-of-the-art.While the widespread application of deep learning(DL)has opened up new opportunities to accomplish the goal,data quality and model interpretability have continued to present a roadblock for the widespread acceptance of DL for real-world applications.This has motivated research on two fronts:data curation,which aims to provide quality data as input for meaningful DL-based analysis,and model interpretation,which intends to reveal the physical reasoning underlying DL model outputs and promote trust from the users.This paper summarizes several key techniques in data curation where breakthroughs in data denoising,outlier detection,imputation,balancing,and semantic annotation have demonstrated the effectiveness in information extraction from noisy,incomplete,insufficient,and/or unannotated data.Also highlighted are model interpretation methods that address the“black-box”nature of DL towards model transparency.
基金Supported by the National Natural Science Foundation of China(61374166)the Doctoral Fund of Ministry of Education of China(20120010110010)the Natural Science Fund of Ningbo(2012A610001)
文摘Nonlinear characteristic fault detection and diagnosis method based on higher-order statistical(HOS) is an effective data-driven method, but the calculation costs much for a large-scale process control system. An HOS-ISM fault diagnosis framework combining interpretative structural model(ISM) and HOS is proposed:(1) the adjacency matrix is determined by partial correlation coefficient;(2) the modified adjacency matrix is defined by directed graph with prior knowledge of process piping and instrument diagram;(3) interpretative structural for large-scale process control system is built by this ISM method; and(4) non-Gaussianity index, nonlinearity index, and total nonlinearity index are calculated dynamically based on interpretative structural to effectively eliminate uncertainty of the nonlinear characteristic diagnostic method with reasonable sampling period and data window. The proposed HOS-ISM fault diagnosis framework is verified by the Tennessee Eastman process and presents improvement for highly non-linear characteristic for selected fault cases.
基金the National Basic Research Program of China (973 Program) ( 2007CB407206)the National Key Technologies Research and Develop-ment Program in the Eleventh Five-Year Plan of China (2006BAC01A11)
文摘For ecological restoration and reconstruction of the degraded area, it is an important premise to correctly understand the degradation factors of the ecosystem in the arid-hot valleys. The factors including vegetation degradation, land degradation, arid climate, policy failure, forest fire, rapid population growth, excessive deforestation, overgrazing, steep slope reclamation, economic poverty, engineering construction, lithology, slope, low cultural level, geological hazards, biological disaster, soil properties etc, were selected to study the Yuanmou arid-hot valleys. Based on the interpretative structural model (ISM), it has found out that the degradation factors of the Yuanmou arid-hot valleys were not at the same level but in a multilevel hierarchical system with internal relations, which pointed out that the degradation mode of the arid-hot valleys was "straight (appearance)-penetrating-background". Such researches have important directive significance for the restoration and reconstruction of the arid-hot valleys ecosystem.
基金supported by National Natural Science Foundation of China(41474115,42174155)Open Fund of Key Laboratory of Exploration Technologies for Oil and Gas Resources(Yangtze University)Ministry of Education of China(No K2018-02)。
文摘The distributions of local velocity and local phase holdup along the radial direction of pipes are complicated because of gravity differentiation,and the distribution of fluid velocity fi eld changes along the gravity direction in horizontal wells.Therefore,measuring the mixture flow and water holdup is difficult,resulting in poor interpretation accuracy of the production logging output profile.In this paper,oil–water two-phase flow dynamic simulation logging experiments in horizontal oil–water two-phase fl ow simulation wells were conducted using the Multiple Array Production Suite,which comprises a capacitance array tool(CAT)and a spinner array tool(SAT),and then the response characteristics of SAT and CAT in diff erent fl ow rates and water cut production conditions were studied.According to the response characteristics of CAT in diff erent water holdup ranges,interpolation imaging along the wellbore section determines the water holdup distribution,and then,the oil–water two-phase velocity fi eld in the fl ow section was reconstructed on the basis of the fl ow section water holdup distribution and the logging value of SAT and combined with the rheological equation of viscous fl uid,and the calculation method of the oil–water partial phase fl ow rate in the fl ow section was proposed.This new approach was applied in the experiment data calculations,and the results are basically consistent with the experimental data.The total fl ow rate and water holdup from the calculation are in agreement with the set values in the experiment,suggesting that the method has high accuracy.
基金Supported by the National Natural Science Foundation of China(61473026,61104131)the Fundamental Research Funds for the Central Universities(JD1413)
文摘Alarm flood is one of the main problems in the alarm systems of industrial process. Alarm root-cause analysis and alarm prioritization are good for alarm flood reduction. This paper proposes a systematic rationalization method for multivariate correlated alarms to realize the root cause analysis and alarm prioritization. An information fusion based interpretive structural model is constructed according to the data-driven partial correlation coefficient calculation and process knowledge modification. This hierarchical multi-layer model is helpful in abnormality propagation path identification and root-cause analysis. Revised Likert scale method is adopted to determine the alarm priority and reduce the blindness of alarm handling. As a case study, the Tennessee Eastman process is utilized to show the effectiveness and validity of proposed approach. Alarm system performance comparison shows that our rationalization methodology can reduce the alarm flood to some extent and improve the performance.
文摘In complex media, especially for seismic prospecting in deep layers in East China and in the mountainous area in West China, due to the complex geological condition, the common-mid-point (CMP) gather of deep reflection event is neither hyperbolic, nor any simple function. If traditional normal move-out (NMO) and stack imaging technology are still used, it is difficult to get a clear stack image. Based on previous techniques on non-hyperbolic stack, it is thought in this paper that no matter how complex the geological condition is, in order to get an optimized stack image, the stack should be non time move-out stack, and any stacking method limited to some kind of curve will be restricted to application conditions. In order to overcome the above-mentioned limit, a new method called optimized non-hyperbolic stack imaging based on interpretation model is presented in this paper. Based on CMP/CRP (Common-Reflection-Point) gather after NMO or pre-stack migration, this method uses the interpretation model of reflectors as constraint, and takes comparability as a distinguishing criterion, and finally forms a residual move-out correction for the gather of constrained model. Numerical simulation indicates that this method could overcome the non hyperbolic problem and get fine stack image.
基金the National Key R&D Program of China(2019YFC1510700)the Sichuan Science and Technology Program(2022YFS0539)the Geomatics Technology and Application Key Laboratory of Qinghai Province,China(QHDX-2018-07).
文摘Traditional visual interpretation is often inefficient due to its excessively workload professional knowledge and strong subjectivity.Therefore,building an automatic interpretation model on high spatial resolution remote sensing images is the key to the quick and efficient interpretation of earthquake-triggered landslides.Aiming at addressing this problem,a landslide interpretation model of high-resolution images based on bag of visual word(BoVW)feature was proposed.The high-resolution images were pre-processed,and then BoVW feature and support vector machine(SVM)was adopted to establish an automatic landslide interpretation model.This model was further compared with the currently widely used Histogram of Oriented Gradient(HoG)feature extraction model.In order to test the effectiveness of the method,typical landslide images were selected to construct a landslide sample library,which was subsequently utilized as the foundation for conducting an experimental study.The results show that the accuracy of landslide extraction using this method reaches as high as 89%,indicating that the method can be used for the automatic interpretation of landslides in disaster-prone areas,and has high practical value for regional disaster prevention and damage reduction.
文摘In this paper, the structure characteristics of open complex giant systems are concretely analysed in depth, thus the view and its significance to support the meta synthesis engineering with manifold knowledge models are clarified. Furthermore, the knowledge based multifaceted modeling methodology for open complex giant systems is emphatically studied. The major points are as follows: (1) nonlinear mechanism and general information partition law; (2) from the symmetry and similarity to the acquisition of construction knowledge; (3) structures for hierarchical and nonhierarchical organizations; (4) the integration of manifold knowledge models; (5) the methodology of knowledge based multifaceted modeling.
基金Supported by the National Natural Science Foundation of China(61374166,6153303)the Doctoral Fund of Ministry of Education of China(20120010110010)the Fundamental Research Funds for the Central Universities(YS1404,JD1413,ZY1502)
文摘Interpretative structural model(ISM) can transform a multivariate problem into several sub-variable problems to analyze a complex industrial structure in a more efficient way by building a multi-level hierarchical structure model. To build an ISM of a production system, the partial correlation coefficient method is proposed to obtain the adjacency matrix, which can be transformed to ISM. According to estimation of correlation coefficient, the result can give actual variable correlations and eliminate effects of intermediate variables. Furthermore, this paper proposes an effective approach using ISM to analyze the main factors and basic mechanisms that affect the energy consumption in an ethylene production system. The case study shows that the proposed energy consumption analysis method is valid and efficient in improvement of energy efficiency in ethylene production.
文摘In enterprise operations,maintaining manual rules for enterprise processes can be expensive,time-consuming,and dependent on specialized domain knowledge in that enterprise domain.Recently,rule-generation has been automated in enterprises,particularly through Machine Learning,to streamline routine tasks.Typically,these machine models are black boxes where the reasons for the decisions are not always transparent,and the end users need to verify the model proposals as a part of the user acceptance testing to trust it.In such scenarios,rules excel over Machine Learning models as the end-users can verify the rules and have more trust.In many scenarios,the truth label changes frequently thus,it becomes difficult for the Machine Learning model to learn till a considerable amount of data has been accumulated,but with rules,the truth can be adapted.This paper presents a novel framework for generating human-understandable rules using the Classification and Regression Tree(CART)decision tree method,which ensures both optimization and user trust in automated decision-making processes.The framework generates comprehensible rules in the form of if condition and then predicts class even in domains where noise is present.The proposed system transforms enterprise operations by automating the production of human-readable rules from structured data,resulting in increased efficiency and transparency.Removing the need for human rule construction saves time and money while guaranteeing that users can readily check and trust the automatic judgments of the system.The remarkable performance metrics of the framework,which achieve 99.85%accuracy and 96.30%precision,further support its efficiency in translating complex data into comprehensible rules,eventually empowering users and enhancing organizational decision-making processes.
文摘To the Editor,We commend Chen et al[1]for their insightful exploration of the challenges associated with the evaluation of large language models(LLMs)and agents in clinical applications.Their work contributes significantly to the discourse on integrating artificial intelligence into healthcare,highlighting critical issues such as data privacy,model interpretability,and risk of misinformation.
基金supported by National Natural Science Fundation of China under Grant 61972208National Natural Science Fundation(General Program)of China under Grant 61972211+2 种基金National Key Research and Development Project of China under Grant 2020YFB1804700Future Network Innovation Research and Application Projects under Grant No.2021FNA020062021 Jiangsu Postgraduate Research Innovation Plan under Grant No.KYCX210794.
文摘With the rapid development of the Internet,network security and data privacy are increasingly valued.Although classical Network Intrusion Detection System(NIDS)based on Deep Learning(DL)models can provide good detection accuracy,but collecting samples for centralized training brings the huge risk of data privacy leakage.Furthermore,the training of supervised deep learning models requires a large number of labeled samples,which is usually cumbersome.The“black-box”problem also makes the DL models of NIDS untrustworthy.In this paper,we propose a trusted Federated Learning(FL)Traffic IDS method called FL-TIDS to address the above-mentioned problems.In FL-TIDS,we design an unsupervised intrusion detection model based on autoencoders that alleviates the reliance on marked samples.At the same time,we use FL for model training to protect data privacy.In addition,we design an improved SHAP interpretable method based on chi-square test to perform interpretable analysis of the trained model.We conducted several experiments to evaluate the proposed FL-TIDS.We first determine experimentally the structure and the number of neurons of the unsupervised AE model.Secondly,we evaluated the proposed method using the UNSW-NB15 and CICIDS2017 datasets.The exper-imental results show that the unsupervised AE model has better performance than the other 7 intrusion detection models in terms of precision,recall and f1-score.Then,federated learning is used to train the intrusion detection model.The experimental results indicate that the model is more accurate than the local learning model.Finally,we use an improved SHAP explainability method based on Chi-square test to analyze the explainability.The analysis results show that the identification characteristics of the model are consistent with the attack characteristics,and the model is reliable.
基金supported by the National Natural Science Foundation of China(82204472)the CAMS Innovation Fund for Medical Sciences(2022-I2M-1-014,2021-I2M-1-028,and 2022-I2M-2-002,China).
文摘Cytotoxicity,usually represented by cell viability,is a crucial parameter for evaluating drug safety in vitro.Accurate prediction of cell viability/cytotoxicity could accelerate drug development in the early stage.In this study,by integrating cellular transcriptome and cell viability data using four machine learning algorithms(support vector machine(SVM),random forest(RF),extreme gradient boosting(XGBoost),and light gradient boosting machine(LightGBM))and two ensemble algorithms(voting and stacking),highly accurate prediction models of 50%and 80%cell viability were developed with area under the receiver operating characteristic curve(AUROC)of 0.90 and 0.84,respectively;these models also showed good performance when utilized for diverse cell lines.Concerning the characterization of the employed Feature Genes,the models were interpreted,and the mechanisms of bioactive compounds with a narrow therapeutic index(NTI)can also be analyzed.In summary,the models established in this research exhibit superior capacity to those of previous studies;these models enable accurate high-safety substance screening via cytotoxicity prediction across cell lines.Moreover,for the first time,Cytotoxicity Signature(CTS)genes were identified,which could provide additional clues for further study of mechanisms of action(MOA),especially for NTI compounds.