Action model learning has become a hot topic in knowledge engineering for automated planning.A key problem for learning action models is to analyze state changes before and after action executions from observed"p...Action model learning has become a hot topic in knowledge engineering for automated planning.A key problem for learning action models is to analyze state changes before and after action executions from observed"plan traces".To support such an analysis,a new approach is proposed to partition propositions of plan traces into states.First,vector representations of propositions and actions are obtained by training a neural network called Skip-Gram borrowed from the area of natural language processing(NLP).Then,a type of semantic distance among propositions and actions is defined based on their similarity measures in the vector space.Finally,k-means and k-nearest neighbor(kNN)algorithms are exploited to map propositions to states.This approach is called state partition by word vector(SPWV),which is implemented on top of a recent action model learning framework by Rao et al.Experimental results on the benchmark domains show that SPWV leads to a lower error rate of the learnt action model,compared to the probability based approach for state partition that was developed by Rao et al.展开更多
Background Cotton is one of the most important commercial crops after food crops,especially in countries like India,where it’s grown extensively under rainfed conditions.Because of its usage in multiple industries,su...Background Cotton is one of the most important commercial crops after food crops,especially in countries like India,where it’s grown extensively under rainfed conditions.Because of its usage in multiple industries,such as textile,medicine,and automobile industries,it has greater commercial importance.The crop’s performance is greatly influenced by prevailing weather dynamics.As climate changes,assessing how weather changes affect crop performance is essential.Among various techniques that are available,crop models are the most effective and widely used tools for predicting yields.Results This study compares statistical and machine learning models to assess their ability to predict cotton yield across major producing districts of Karnataka,India,utilizing a long-term dataset spanning from 1990 to 2023 that includes yield and weather factors.The artificial neural networks(ANNs)performed superiorly with acceptable yield deviations ranging within±10%during both vegetative stage(F1)and mid stage(F2)for cotton.The model evaluation metrics such as root mean square error(RMSE),normalized root mean square error(nRMSE),and modelling efficiency(EF)were also within the acceptance limits in most districts.Furthermore,the tested ANN model was used to assess the importance of the dominant weather factors influencing crop yield in each district.Specifically,the use of morning relative humidity as an individual parameter and its interaction with maximum and minimum tempera-ture had a major influence on cotton yield in most of the yield predicted districts.These differences highlighted the differential interactions of weather factors in each district for cotton yield formation,highlighting individual response of each weather factor under different soils and management conditions over the major cotton growing districts of Karnataka.Conclusions Compared with statistical models,machine learning models such as ANNs proved higher efficiency in forecasting the cotton yield due to their ability to consider the interactive effects of weather factors on yield forma-tion at different growth stages.This highlights the best suitability of ANNs for yield forecasting in rainfed conditions and for the study on relative impacts of weather factors on yield.Thus,the study aims to provide valuable insights to support stakeholders in planning effective crop management strategies and formulating relevant policies.展开更多
BACKGROUND Ischemic heart disease(IHD)impacts the quality of life and has the highest mortality rate of cardiovascular diseases globally.AIM To compare variations in the parameters of the single-lead electrocardiogram...BACKGROUND Ischemic heart disease(IHD)impacts the quality of life and has the highest mortality rate of cardiovascular diseases globally.AIM To compare variations in the parameters of the single-lead electrocardiogram(ECG)during resting conditions and physical exertion in individuals diagnosed with IHD and those without the condition using vasodilator-induced stress computed tomography(CT)myocardial perfusion imaging as the diagnostic reference standard.METHODS This single center observational study included 80 participants.The participants were aged≥40 years and given an informed written consent to participate in the study.Both groups,G1(n=31)with and G2(n=49)without post stress induced myocardial perfusion defect,passed cardiologist consultation,anthropometric measurements,blood pressure and pulse rate measurement,echocardiography,cardio-ankle vascular index,bicycle ergometry,recording 3-min single-lead ECG(Cardio-Qvark)before and just after bicycle ergometry followed by performing CT myocardial perfusion.The LASSO regression with nested cross-validation was used to find the association between Cardio-Qvark parameters and the existence of the perfusion defect.Statistical processing was performed with the R programming language v4.2,Python v.3.10[^R],and Statistica 12 program.RESULTS Bicycle ergometry yielded an area under the receiver operating characteristic curve of 50.7%[95%confidence interval(CI):0.388-0.625],specificity of 53.1%(95%CI:0.392-0.673),and sensitivity of 48.4%(95%CI:0.306-0.657).In contrast,the Cardio-Qvark test performed notably better with an area under the receiver operating characteristic curve of 67%(95%CI:0.530-0.801),specificity of 75.5%(95%CI:0.628-0.88),and sensitivity of 51.6%(95%CI:0.333-0.695).CONCLUSION The single-lead ECG has a relatively higher diagnostic accuracy compared with bicycle ergometry by using machine learning models,but the difference was not statistically significant.However,further investigations are required to uncover the hidden capabilities of single-lead ECG in IHD diagnosis.展开更多
Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of th...Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of these data has not been well stored,managed and mined.With the development of cloud computing technology,it provides a rare development opportunity for logging big data private cloud.The traditional petrophysical evaluation and interpretation model has encountered great challenges in the face of new evaluation objects.The solution research of logging big data distributed storage,processing and learning functions integrated in logging big data private cloud has not been carried out yet.To establish a distributed logging big-data private cloud platform centered on a unifi ed learning model,which achieves the distributed storage and processing of logging big data and facilitates the learning of novel knowledge patterns via the unifi ed logging learning model integrating physical simulation and data models in a large-scale functional space,thus resolving the geo-engineering evaluation problem of geothermal fi elds.Based on the research idea of“logging big data cloud platform-unifi ed logging learning model-large function space-knowledge learning&discovery-application”,the theoretical foundation of unified learning model,cloud platform architecture,data storage and learning algorithm,arithmetic power allocation and platform monitoring,platform stability,data security,etc.have been carried on analysis.The designed logging big data cloud platform realizes parallel distributed storage and processing of data and learning algorithms.The feasibility of constructing a well logging big data cloud platform based on a unifi ed learning model of physics and data is analyzed in terms of the structure,ecology,management and security of the cloud platform.The case study shows that the logging big data cloud platform has obvious technical advantages over traditional logging evaluation methods in terms of knowledge discovery method,data software and results sharing,accuracy,speed and complexity.展开更多
Sporadic E(Es)layers in the ionosphere are characterized by intense plasma irregularities in the E region at altitudes of 90-130 km.Because they can significantly influence radio communications and navigation systems,...Sporadic E(Es)layers in the ionosphere are characterized by intense plasma irregularities in the E region at altitudes of 90-130 km.Because they can significantly influence radio communications and navigation systems,accurate forecasting of Es layers is crucial for ensuring the precision and dependability of navigation satellite systems.In this study,we present Es predictions made by an empirical model and by a deep learning model,and analyze their differences comprehensively by comparing the model predictions to satellite RO measurements and ground-based ionosonde observations.The deep learning model exhibited significantly better performance,as indicated by its high coefficient of correlation(r=0.87)with RO observations and predictions,than did the empirical model(r=0.53).This study highlights the importance of integrating artificial intelligence technology into ionosphere modelling generally,and into predicting Es layer occurrences and characteristics,in particular.展开更多
The investigation by Zhu et al on the assessment of cellular proliferation markers to assist clinical decision-making in patients with hepatocellular carcinoma(HCC)using a machine learning model-based approach is a sc...The investigation by Zhu et al on the assessment of cellular proliferation markers to assist clinical decision-making in patients with hepatocellular carcinoma(HCC)using a machine learning model-based approach is a scientific approach.This study looked into the possibilities of using a Ki-67(a marker for cell proliferation)expression-based machine learning model to help doctors make decisions about treatment options for patients with HCC before surgery.The study used reconstructed tomography images of 164 patients with confirmed HCC from the intratumoral and peritumoral regions.The features were chosen using various statistical methods,including least absolute shrinkage and selection operator regression.Also,a nomogram was made using Radscore and clinical risk factors.It was tested for its ability to predict receiver operating characteristic curves and calibration curves,and its clinical benefits were found using decision curve analysis.The calibration curve demonstrated excellent consistency between predicted and actual probability,and the decision curve confirmed its clinical benefit.The proposed model is helpful for treating patients with HCC because the predicted and actual probabilities are very close to each other,as shown by the decision curve analysis.Further prospective studies are required,incorporating a multicenter and large sample size design,additional relevant exclusion criteria,information on tumors(size,number,and grade),and cancer stage to strengthen the clinical benefit in patients with HCC.展开更多
The backwater effect caused by tributary inflow can significantly elevate the water level profile upstream of a confluence point.However,the influence of mainstream and confluence discharges on the backwater effect in...The backwater effect caused by tributary inflow can significantly elevate the water level profile upstream of a confluence point.However,the influence of mainstream and confluence discharges on the backwater effect in a river reach remains unclear.In this study,various hydrological data collected from the Jingjiang Reach of the Yangtze River in China were statistically analyzed to determine the backwater degree and range with three representative mainstream discharges.The results indicated that the backwater degree increased with mainstream discharge,and a positive relationship was observed between the runoff ratio and backwater degree at specific representative mainstream discharges.Following the operation of the Three Gorges Project,the backwater effect in the Jingjiang Reach diminished.For instance,mean backwater degrees for low,moderate,and high mainstream discharges were recorded as 0.83 m,1.61 m,and 2.41 m during the period from 1990 to 2002,whereas these values decreased to 0.30 m,0.95 m,and 2.08 m from 2009 to 2020.The backwater range extended upstream as mainstream discharge increased from 7000 m3/s to 30000 m3/s.Moreover,a random forest-based machine learning model was used to quantify the backwater effect with varying mainstream and confluence discharges,accounting for the impacts of mainstream discharge,confluence discharge,and channel degradation in the Jingjiang Reach.At the Jianli Hydrological Station,a decrease in mainstream discharge during flood seasons resulted in a 7%–15%increase in monthly mean backwater degree,while an increase in mainstream discharge during dry seasons led to a 1%–15%decrease in monthly mean backwater degree.Furthermore,increasing confluence discharge from Dongting Lake during June to July and September to November resulted in an 11%–42%increase in monthly mean backwater degree.Continuous channel degradation in the Jingjiang Reach contributed to a 6%–19%decrease in monthly mean backwater degree.Under the influence of these factors,the monthly mean backwater degree in 2017 varied from a decrease of 53%to an increase of 37%compared to corresponding values in 1991.展开更多
Understanding spatial heterogeneity in groundwater responses to multiple factors is critical for water resource management in coastal cities.Daily groundwater depth(GWD)data from 43 wells(2018-2022)were collected in t...Understanding spatial heterogeneity in groundwater responses to multiple factors is critical for water resource management in coastal cities.Daily groundwater depth(GWD)data from 43 wells(2018-2022)were collected in three coastal cities in Jiangsu Province,China.Seasonal and Trend decomposition using Loess(STL)together with wavelet analysis and empirical mode decomposition were applied to identify tide-influenced wells while remaining wells were grouped by hierarchical clustering analysis(HCA).Machine learning models were developed to predict GWD,then their response to natural conditions and human activities was assessed by the Shapley Additive exPlanations(SHAP)method.Results showed that eXtreme Gradient Boosting(XGB)was superior to other models in terms of prediction performance and computational efficiency(R^(2)>0.95).GWD in Yancheng and southern Lianyungang were greater than those in Nantong,exhibiting larger fluctuations.Groundwater within 5 km of the coastline was affected by tides,with more pronounced effects in agricultural areas compared to urban areas.Shallow groundwater(3-7 m depth)responded immediately(0-1 day)to rainfall,primarily influenced by farmland and topography(slope and distance from rivers).Rainfall recharge to groundwater peaked at 50%farmland coverage,but this effect was suppressed by high temperatures(>30℃)which intensified as distance from rivers increased,especially in forest and grassland.Deep groundwater(>10 m)showed delayed responses to rainfall(1-4 days)and temperature(10-15 days),with GDP as the primary influence,followed by agricultural irrigation and population density.Farmland helped to maintain stable GWD in low population density regions,while excessive farmland coverage(>90%)led to overexploitation.In the early stages of GDP development,increased industrial and agricultural water demand led to GWD decline,but as GDP levels significantly improved,groundwater consumption pressure gradually eased.This methodological framework is applicable not only to coastal cities in China but also could be extended to coastal regions worldwide.展开更多
With the rapid development of economy,air pollution caused by industrial expansion has caused serious harm to human health and social development.Therefore,establishing an effective air pollution concentration predict...With the rapid development of economy,air pollution caused by industrial expansion has caused serious harm to human health and social development.Therefore,establishing an effective air pollution concentration prediction system is of great scientific and practical significance for accurate and reliable predictions.This paper proposes a combination of pointinterval prediction system for pollutant concentration prediction by leveraging neural network,meta-heuristic optimization algorithm,and fuzzy theory.Fuzzy information granulation technology is used in data preprocessing to transform numerical sequences into fuzzy particles for comprehensive feature extraction.The golden Jackal optimization algorithm is employed in the optimization stage to fine-tune model hyperparameters.In the prediction stage,an ensemble learning method combines training results frommultiplemodels to obtain final point predictions while also utilizing quantile regression and kernel density estimation methods for interval predictions on the test set.Experimental results demonstrate that the combined model achieves a high goodness of fit coefficient of determination(R^(2))at 99.3% and a maximum difference between prediction accuracy mean absolute percentage error(MAPE)and benchmark model at 12.6%.This suggests that the integrated learning system proposed in this paper can provide more accurate deterministic predictions as well as reliable uncertainty analysis compared to traditionalmodels,offering practical reference for air quality early warning.展开更多
It is fundamental and useful to investigate how deep learning forecasting models(DLMs)perform compared to operational oceanography forecast systems(OFSs).However,few studies have intercompared their performances using...It is fundamental and useful to investigate how deep learning forecasting models(DLMs)perform compared to operational oceanography forecast systems(OFSs).However,few studies have intercompared their performances using an identical reference.In this study,three physically reasonable DLMs are implemented for the forecasting of the sea surface temperature(SST),sea level anomaly(SLA),and sea surface velocity in the South China Sea.The DLMs are validated against both the testing dataset and the“OceanPredict”Class 4 dataset.Results show that the DLMs'RMSEs against the latter increase by 44%,245%,302%,and 109%for SST,SLA,current speed,and direction,respectively,compared to those against the former.Therefore,different references have significant influences on the validation,and it is necessary to use an identical and independent reference to intercompare the DLMs and OFSs.Against the Class 4 dataset,the DLMs present significantly better performance for SLA than the OFSs,and slightly better performances for other variables.The error patterns of the DLMs and OFSs show a high degree of similarity,which is reasonable from the viewpoint of predictability,facilitating further applications of the DLMs.For extreme events,the DLMs and OFSs both present large but similar forecast errors for SLA and current speed,while the DLMs are likely to give larger errors for SST and current direction.This study provides an evaluation of the forecast skills of commonly used DLMs and provides an example to objectively intercompare different DLMs.展开更多
This research investigates the influence of indoor and outdoor factors on photovoltaic(PV)power generation at Utrecht University to accurately predict PV system performance by identifying critical impact factors and i...This research investigates the influence of indoor and outdoor factors on photovoltaic(PV)power generation at Utrecht University to accurately predict PV system performance by identifying critical impact factors and improving renewable energy efficiency.To predict plant efficiency,nineteen variables are analyzed,consisting of nine indoor photovoltaic panel characteristics(Open Circuit Voltage(Voc),Short Circuit Current(Isc),Maximum Power(Pmpp),Maximum Voltage(Umpp),Maximum Current(Impp),Filling Factor(FF),Parallel Resistance(Rp),Series Resistance(Rs),Module Temperature)and ten environmental factors(Air Temperature,Air Humidity,Dew Point,Air Pressure,Irradiation,Irradiation Propagation,Wind Speed,Wind Speed Propagation,Wind Direction,Wind Direction Propagation).This study provides a new perspective not previously addressed in the literature.In this study,different machine learning methods such as Multilayer Perceptron(MLP),Multivariate Adaptive Regression Spline(MARS),Multiple Linear Regression(MLR),and Random Forest(RF)models are used to predict power values using data from installed PVpanels.Panel values obtained under real field conditions were used to train the models,and the results were compared.The Multilayer Perceptron(MLP)model was achieved with the highest classification accuracy of 0.990%.The machine learning models used for solar energy forecasting show high performance and produce results close to actual values.Models like Multi-Layer Perceptron(MLP)and Random Forest(RF)can be used in diverse locations based on load demand.展开更多
Software systems are present all around us and playing their vital roles in our daily life.The correct functioning of these systems is of prime concern.In addition to classical testing techniques,formal techniques lik...Software systems are present all around us and playing their vital roles in our daily life.The correct functioning of these systems is of prime concern.In addition to classical testing techniques,formal techniques like model checking are used to reinforce the quality and reliability of software systems.However,obtaining of behavior model,which is essential for model-based techniques,of unknown software systems is a challenging task.To mitigate this problem,an emerging black-box analysis technique,called Model Learning,can be applied.It complements existing model-based testing and verification approaches by providing behavior models of blackbox systems fully automatically.This paper surveys the model learning technique,which recently has attracted much attention from researchers,especially from the domains of testing and verification.First,we review the background and foundations of model learning,which form the basis of subsequent sections.Second,we present some well-known model learning tools and provide their merits and shortcomings in the form of a comparison table.Third,we describe the successful applications of model learning in multidisciplinary fields,current challenges along with possible future works,and concluding remarks.展开更多
Natural language processing(NLP)is a subfield of artificial intelligence that focuses on enabling computers to understand and process human languages.In the last five years,we have witnessed the rapid development of N...Natural language processing(NLP)is a subfield of artificial intelligence that focuses on enabling computers to understand and process human languages.In the last five years,we have witnessed the rapid development of NLP in tasks such as machine translation,question-answering,and machine reading comprehension based on deep learning and an enormous volume of annotated and unannotated data.In this paper,we will review the latest progress in the neural network-based NLP framework(neural NLP)from three perspectives:modeling,learning,and reasoning.In the modeling section,we will describe several fundamental neural network-based modeling paradigms,such as word embedding,sentence embedding,and sequence-to-sequence modeling,which are widely used in modern NLP engines.In the learning section,we will introduce widely used learning methods for NLP models,including supervised,semi-supervised,and unsupervised learning;multitask learning;transfer learning;and active learning.We view reasoning as a new and exciting direction for neural NLP,but it has yet to be well addressed.In the reasoning section,we will review reasoning mechanisms,including the knowledge,existing non-neural inference methods,and new neural inference methods.We emphasize the importance of reasoning in this paper because it is important for building interpretable and knowledgedriven neural NLP models to handle complex tasks.At the end of this paper,we will briefly outline our thoughts on the future directions of neural NLP.展开更多
Fibroblast growth factor (FGF)/fibroblast growth factor receptor (FGFR) signaling plays essential roles in bone development and diseases. Missense mutations in FGFs and FGFRs in humans can cause various congenital...Fibroblast growth factor (FGF)/fibroblast growth factor receptor (FGFR) signaling plays essential roles in bone development and diseases. Missense mutations in FGFs and FGFRs in humans can cause various congenital bone diseases, including chondrodysplasia syndromes, craniosynostosis syndromes and syndromes with dysregulated phosphate metabolism. FGF/FGFR signaling is also an important pathway involved in the maintenance of adult bone homeostasis. Multiple kinds of mouse models, mimicking human skeleton diseases caused by missense mutations in FGFs and FGFRs, have been established by knock-m/out and transgenic technologies. These genetically modified mice provide good models for studying the role of FGF/FGFR signaling in skeleton development and homeostasis. In this review, we summarize the mouse models of FGF signaling-related skeleton diseases and recent progresses regarding the molecular mechanisms, underlying the role of FGFs/FGFRs in the regulation of bone development and homeostasis. This review also provides a perspective view on future works to explore the roles of FGF signaling in skeletal development and homeostasis.展开更多
To perform landslide susceptibility prediction(LSP),it is important to select appropriate mapping unit and landslide-related conditioning factors.The efficient and automatic multi-scale segmentation(MSS)method propose...To perform landslide susceptibility prediction(LSP),it is important to select appropriate mapping unit and landslide-related conditioning factors.The efficient and automatic multi-scale segmentation(MSS)method proposed by the authors promotes the application of slope units.However,LSP modeling based on these slope units has not been performed.Moreover,the heterogeneity of conditioning factors in slope units is neglected,leading to incomplete input variables of LSP modeling.In this study,the slope units extracted by the MSS method are used to construct LSP modeling,and the heterogeneity of conditioning factors is represented by the internal variations of conditioning factors within slope unit using the descriptive statistics features of mean,standard deviation and range.Thus,slope units-based machine learning models considering internal variations of conditioning factors(variant slope-machine learning)are proposed.The Chongyi County is selected as the case study and is divided into 53,055 slope units.Fifteen original slope unit-based conditioning factors are expanded to 38 slope unit-based conditioning factors through considering their internal variations.Random forest(RF)and multi-layer perceptron(MLP)machine learning models are used to construct variant Slope-RF and Slope-MLP models.Meanwhile,the Slope-RF and Slope-MLP models without considering the internal variations of conditioning factors,and conventional grid units-based machine learning(Grid-RF and MLP)models are built for comparisons through the LSP performance assessments.Results show that the variant Slopemachine learning models have higher LSP performances than Slope-machine learning models;LSP results of variant Slope-machine learning models have stronger directivity and practical application than Grid-machine learning models.It is concluded that slope units extracted by MSS method can be appropriate for LSP modeling,and the heterogeneity of conditioning factors within slope units can more comprehensively reflect the relationships between conditioning factors and landslides.The research results have important reference significance for land use and landslide prevention.展开更多
The prediction of fundus fluorescein angiography(FFA)images from fundus structural images is a cutting-edge research topic in ophthalmological image processing.Prediction comprises estimating FFA from fundus camera im...The prediction of fundus fluorescein angiography(FFA)images from fundus structural images is a cutting-edge research topic in ophthalmological image processing.Prediction comprises estimating FFA from fundus camera imaging,single-phase FFA from scanning laser ophthalmoscopy(SLO),and three-phase FFA also from SLO.Although many deep learning models are available,a single model can only perform one or two of these prediction tasks.To accomplish three prediction tasks using a unified method,we propose a unified deep learning model for predicting FFA images from fundus structure images using a supervised generative adversarial network.The three prediction tasks are processed as follows:data preparation,network training under FFA supervision,and FFA image prediction from fundus structure images on a test set.By comparing the FFA images predicted by our model,pix2pix,and CycleGAN,we demonstrate the remarkable progress achieved by our proposal.The high performance of our model is validated in terms of the peak signal-to-noise ratio,structural similarity index,and mean squared error.展开更多
BACKGROUND Bleeding is one of the major complications after endoscopic submucosal dissection(ESD)in early gastric cancer(EGC)patients.There are limited studies on estimating the bleeding risk after ESD using an artifi...BACKGROUND Bleeding is one of the major complications after endoscopic submucosal dissection(ESD)in early gastric cancer(EGC)patients.There are limited studies on estimating the bleeding risk after ESD using an artificial intelligence system.AIM To derivate and verify the performance of the deep learning model and the clinical model for predicting bleeding risk after ESD in EGC patients.METHODS Patients with EGC who underwent ESD between January 2010 and June 2020 at the Samsung Medical Center were enrolled,and post-ESD bleeding(PEB)was investigated retrospectively.We split the entire cohort into a development set(80%)and a validation set(20%).The deep learning and clinical model were built on the development set and tested in the validation set.The performance of the deep learning model and the clinical model were compared using the area under the curve and the stratification of bleeding risk after ESD.RESULTS A total of 5629 patients were included,and PEB occurred in 325 patients.The area under the curve for predicting PEB was 0.71(95%confidence interval:0.63-0.78)in the deep learning model and 0.70(95%confidence interval:0.62-0.77)in the clinical model,without significant difference(P=0.730).The patients expected to the low-(<5%),intermediate-(≥5%,<9%),and high-risk(≥9%)categories were observed with actual bleeding rate of 2.2%,3.9%,and 11.6%,respectively,in the deep learning model;4.0%,8.8%,and 18.2%,respectively,in the clinical model.CONCLUSION A deep learning model can predict and stratify the bleeding risk after ESD in patients with EGC.展开更多
The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper ...The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper makes an attempt to assess landslide susceptibility in Shimla district of the northwest Indian Himalayan region.It examined the effectiveness of random forest(RF),multilayer perceptron(MLP),sequential minimal optimization regression(SMOreg)and bagging ensemble(B-RF,BSMOreg,B-MLP)models.A landslide inventory map comprising 1052 locations of past landslide occurrences was classified into training(70%)and testing(30%)datasets.The site-specific influencing factors were selected by employing a multicollinearity test.The relationship between past landslide occurrences and influencing factors was established using the frequency ratio method.The effectiveness of machine learning models was verified through performance assessors.The landslide susceptibility maps were validated by the area under the receiver operating characteristic curves(ROC-AUC),accuracy,precision,recall and F1-score.The key performance metrics and map validation demonstrated that the BRF model(correlation coefficient:0.988,mean absolute error:0.010,root mean square error:0.058,relative absolute error:2.964,ROC-AUC:0.947,accuracy:0.778,precision:0.819,recall:0.917 and F-1 score:0.865)outperformed the single classifiers and other bagging ensemble models for landslide susceptibility.The results show that the largest area was found under the very high susceptibility zone(33.87%),followed by the low(27.30%),high(20.68%)and moderate(18.16%)susceptibility zones.The factors,namely average annual rainfall,slope,lithology,soil texture and earthquake magnitude have been identified as the influencing factors for very high landslide susceptibility.Soil texture,lineament density and elevation have been attributed to high and moderate susceptibility.Thus,the study calls for devising suitable landslide mitigation measures in the study area.Structural measures,an immediate response system,community participation and coordination among stakeholders may help lessen the detrimental impact of landslides.The findings from this study could aid decision-makers in mitigating future catastrophes and devising suitable strategies in other geographical regions with similar geological characteristics.展开更多
BACKGROUND Colorectal cancer significantly impacts global health,with unplanned reoperations post-surgery being key determinants of patient outcomes.Existing predictive models for these reoperations lack precision in ...BACKGROUND Colorectal cancer significantly impacts global health,with unplanned reoperations post-surgery being key determinants of patient outcomes.Existing predictive models for these reoperations lack precision in integrating complex clinical data.AIM To develop and validate a machine learning model for predicting unplanned reoperation risk in colorectal cancer patients.METHODS Data of patients treated for colorectal cancer(n=2044)at the First Affiliated Hospital of Wenzhou Medical University and Wenzhou Central Hospital from March 2020 to March 2022 were retrospectively collected.Patients were divided into an experimental group(n=60)and a control group(n=1984)according to unplanned reoperation occurrence.Patients were also divided into a training group and a validation group(7:3 ratio).We used three different machine learning methods to screen characteristic variables.A nomogram was created based on multifactor logistic regression,and the model performance was assessed using receiver operating characteristic curve,calibration curve,Hosmer-Lemeshow test,and decision curve analysis.The risk scores of the two groups were calculated and compared to validate the model.RESULTS More patients in the experimental group were≥60 years old,male,and had a history of hypertension,laparotomy,and hypoproteinemia,compared to the control group.Multiple logistic regression analysis confirmed the following as independent risk factors for unplanned reoperation(P<0.05):Prognostic Nutritional Index value,history of laparotomy,hypertension,or stroke,hypoproteinemia,age,tumor-node-metastasis staging,surgical time,gender,and American Society of Anesthesiologists classification.Receiver operating characteristic curve analysis showed that the model had good discrimination and clinical utility.CONCLUSION This study used a machine learning approach to build a model that accurately predicts the risk of postoperative unplanned reoperation in patients with colorectal cancer,which can improve treatment decisions and prognosis.展开更多
Fine-grained weather forecasting data,i.e.,the grid data with high-resolution,have attracted increasing attention in recent years,especially for some specific applications such as the Winter Olympic Games.Although Eur...Fine-grained weather forecasting data,i.e.,the grid data with high-resolution,have attracted increasing attention in recent years,especially for some specific applications such as the Winter Olympic Games.Although European Centre for Medium-Range Weather Forecasts(ECMWF)provides grid prediction up to 240 hours,the coarse data are unable to meet high requirements of these major events.In this paper,we propose a method,called model residual machine learning(MRML),to generate grid prediction with high-resolution based on high-precision stations forecasting.MRML applies model output machine learning(MOML)for stations forecasting.Subsequently,MRML utilizes these forecasts to improve the quality of the grid data by fitting a machine learning(ML)model to the residuals.We demonstrate that MRML achieves high capability at diverse meteorological elements,specifically,temperature,relative humidity,and wind speed.In addition,MRML could be easily extended to other post-processing methods by invoking different techniques.In our experiments,MRML outperforms the traditional downscaling methods such as piecewise linear interpolation(PLI)on the testing data.展开更多
基金Supported by the National Natural Science Foundation of China(61103136,61370156,61503074)Open Research Foundation of Science and Technology on Aerospace Flight Dynamics Laboratory(2014afdl002)
文摘Action model learning has become a hot topic in knowledge engineering for automated planning.A key problem for learning action models is to analyze state changes before and after action executions from observed"plan traces".To support such an analysis,a new approach is proposed to partition propositions of plan traces into states.First,vector representations of propositions and actions are obtained by training a neural network called Skip-Gram borrowed from the area of natural language processing(NLP).Then,a type of semantic distance among propositions and actions is defined based on their similarity measures in the vector space.Finally,k-means and k-nearest neighbor(kNN)algorithms are exploited to map propositions to states.This approach is called state partition by word vector(SPWV),which is implemented on top of a recent action model learning framework by Rao et al.Experimental results on the benchmark domains show that SPWV leads to a lower error rate of the learnt action model,compared to the probability based approach for state partition that was developed by Rao et al.
基金funded through India Meteorological Department,New Delhi,India under the Forecasting Agricultural output using Space,Agrometeorol ogy and Land based observations(FASAL)project and fund number:No.ASC/FASAL/KT-11/01/HQ-2010.
文摘Background Cotton is one of the most important commercial crops after food crops,especially in countries like India,where it’s grown extensively under rainfed conditions.Because of its usage in multiple industries,such as textile,medicine,and automobile industries,it has greater commercial importance.The crop’s performance is greatly influenced by prevailing weather dynamics.As climate changes,assessing how weather changes affect crop performance is essential.Among various techniques that are available,crop models are the most effective and widely used tools for predicting yields.Results This study compares statistical and machine learning models to assess their ability to predict cotton yield across major producing districts of Karnataka,India,utilizing a long-term dataset spanning from 1990 to 2023 that includes yield and weather factors.The artificial neural networks(ANNs)performed superiorly with acceptable yield deviations ranging within±10%during both vegetative stage(F1)and mid stage(F2)for cotton.The model evaluation metrics such as root mean square error(RMSE),normalized root mean square error(nRMSE),and modelling efficiency(EF)were also within the acceptance limits in most districts.Furthermore,the tested ANN model was used to assess the importance of the dominant weather factors influencing crop yield in each district.Specifically,the use of morning relative humidity as an individual parameter and its interaction with maximum and minimum tempera-ture had a major influence on cotton yield in most of the yield predicted districts.These differences highlighted the differential interactions of weather factors in each district for cotton yield formation,highlighting individual response of each weather factor under different soils and management conditions over the major cotton growing districts of Karnataka.Conclusions Compared with statistical models,machine learning models such as ANNs proved higher efficiency in forecasting the cotton yield due to their ability to consider the interactive effects of weather factors on yield forma-tion at different growth stages.This highlights the best suitability of ANNs for yield forecasting in rainfed conditions and for the study on relative impacts of weather factors on yield.Thus,the study aims to provide valuable insights to support stakeholders in planning effective crop management strategies and formulating relevant policies.
基金Supported by Government Assignment,No.1023022600020-6RSF Grant,No.24-15-00549Ministry of Science and Higher Education of the Russian Federation within the Framework of State Support for the Creation and Development of World-Class Research Center,No.075-15-2022-304.
文摘BACKGROUND Ischemic heart disease(IHD)impacts the quality of life and has the highest mortality rate of cardiovascular diseases globally.AIM To compare variations in the parameters of the single-lead electrocardiogram(ECG)during resting conditions and physical exertion in individuals diagnosed with IHD and those without the condition using vasodilator-induced stress computed tomography(CT)myocardial perfusion imaging as the diagnostic reference standard.METHODS This single center observational study included 80 participants.The participants were aged≥40 years and given an informed written consent to participate in the study.Both groups,G1(n=31)with and G2(n=49)without post stress induced myocardial perfusion defect,passed cardiologist consultation,anthropometric measurements,blood pressure and pulse rate measurement,echocardiography,cardio-ankle vascular index,bicycle ergometry,recording 3-min single-lead ECG(Cardio-Qvark)before and just after bicycle ergometry followed by performing CT myocardial perfusion.The LASSO regression with nested cross-validation was used to find the association between Cardio-Qvark parameters and the existence of the perfusion defect.Statistical processing was performed with the R programming language v4.2,Python v.3.10[^R],and Statistica 12 program.RESULTS Bicycle ergometry yielded an area under the receiver operating characteristic curve of 50.7%[95%confidence interval(CI):0.388-0.625],specificity of 53.1%(95%CI:0.392-0.673),and sensitivity of 48.4%(95%CI:0.306-0.657).In contrast,the Cardio-Qvark test performed notably better with an area under the receiver operating characteristic curve of 67%(95%CI:0.530-0.801),specificity of 75.5%(95%CI:0.628-0.88),and sensitivity of 51.6%(95%CI:0.333-0.695).CONCLUSION The single-lead ECG has a relatively higher diagnostic accuracy compared with bicycle ergometry by using machine learning models,but the difference was not statistically significant.However,further investigations are required to uncover the hidden capabilities of single-lead ECG in IHD diagnosis.
基金supported By Grant (PLN2022-14) of State Key Laboratory of Oil and Gas Reservoir Geology and Exploitation (Southwest Petroleum University)。
文摘Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of these data has not been well stored,managed and mined.With the development of cloud computing technology,it provides a rare development opportunity for logging big data private cloud.The traditional petrophysical evaluation and interpretation model has encountered great challenges in the face of new evaluation objects.The solution research of logging big data distributed storage,processing and learning functions integrated in logging big data private cloud has not been carried out yet.To establish a distributed logging big-data private cloud platform centered on a unifi ed learning model,which achieves the distributed storage and processing of logging big data and facilitates the learning of novel knowledge patterns via the unifi ed logging learning model integrating physical simulation and data models in a large-scale functional space,thus resolving the geo-engineering evaluation problem of geothermal fi elds.Based on the research idea of“logging big data cloud platform-unifi ed logging learning model-large function space-knowledge learning&discovery-application”,the theoretical foundation of unified learning model,cloud platform architecture,data storage and learning algorithm,arithmetic power allocation and platform monitoring,platform stability,data security,etc.have been carried on analysis.The designed logging big data cloud platform realizes parallel distributed storage and processing of data and learning algorithms.The feasibility of constructing a well logging big data cloud platform based on a unifi ed learning model of physics and data is analyzed in terms of the structure,ecology,management and security of the cloud platform.The case study shows that the logging big data cloud platform has obvious technical advantages over traditional logging evaluation methods in terms of knowledge discovery method,data software and results sharing,accuracy,speed and complexity.
基金supported by the Project of Stable Support for Youth Team in Basic Research Field,CAS(grant No.YSBR-018)the National Natural Science Foundation of China(grant Nos.42188101,42130204)+4 种基金the B-type Strategic Priority Program of CAS(grant no.XDB41000000)the National Natural Science Foundation of China(NSFC)Distinguished Overseas Young Talents Program,Innovation Program for Quantum Science and Technology(2021ZD0300301)the Open Research Project of Large Research Infrastructures of CAS-“Study on the interaction between low/mid-latitude atmosphere and ionosphere based on the Chinese Meridian Project”.The project was supported also by the National Key Laboratory of Deep Space Exploration(Grant No.NKLDSE2023A002)the Open Fund of Anhui Provincial Key Laboratory of Intelligent Underground Detection(Grant No.APKLIUD23KF01)the China National Space Administration(CNSA)pre-research Project on Civil Aerospace Technologies No.D010305,D010301.
文摘Sporadic E(Es)layers in the ionosphere are characterized by intense plasma irregularities in the E region at altitudes of 90-130 km.Because they can significantly influence radio communications and navigation systems,accurate forecasting of Es layers is crucial for ensuring the precision and dependability of navigation satellite systems.In this study,we present Es predictions made by an empirical model and by a deep learning model,and analyze their differences comprehensively by comparing the model predictions to satellite RO measurements and ground-based ionosonde observations.The deep learning model exhibited significantly better performance,as indicated by its high coefficient of correlation(r=0.87)with RO observations and predictions,than did the empirical model(r=0.53).This study highlights the importance of integrating artificial intelligence technology into ionosphere modelling generally,and into predicting Es layer occurrences and characteristics,in particular.
文摘The investigation by Zhu et al on the assessment of cellular proliferation markers to assist clinical decision-making in patients with hepatocellular carcinoma(HCC)using a machine learning model-based approach is a scientific approach.This study looked into the possibilities of using a Ki-67(a marker for cell proliferation)expression-based machine learning model to help doctors make decisions about treatment options for patients with HCC before surgery.The study used reconstructed tomography images of 164 patients with confirmed HCC from the intratumoral and peritumoral regions.The features were chosen using various statistical methods,including least absolute shrinkage and selection operator regression.Also,a nomogram was made using Radscore and clinical risk factors.It was tested for its ability to predict receiver operating characteristic curves and calibration curves,and its clinical benefits were found using decision curve analysis.The calibration curve demonstrated excellent consistency between predicted and actual probability,and the decision curve confirmed its clinical benefit.The proposed model is helpful for treating patients with HCC because the predicted and actual probabilities are very close to each other,as shown by the decision curve analysis.Further prospective studies are required,incorporating a multicenter and large sample size design,additional relevant exclusion criteria,information on tumors(size,number,and grade),and cancer stage to strengthen the clinical benefit in patients with HCC.
基金supported by the National Key Research and Development Program of China(Grant No.2023YFC3209504)the National Natural Science Foundation of China(Grants No.U2040215 and 52479075)the Natural Science Foundation of Hubei Province(Grant No.2021CFA029).
文摘The backwater effect caused by tributary inflow can significantly elevate the water level profile upstream of a confluence point.However,the influence of mainstream and confluence discharges on the backwater effect in a river reach remains unclear.In this study,various hydrological data collected from the Jingjiang Reach of the Yangtze River in China were statistically analyzed to determine the backwater degree and range with three representative mainstream discharges.The results indicated that the backwater degree increased with mainstream discharge,and a positive relationship was observed between the runoff ratio and backwater degree at specific representative mainstream discharges.Following the operation of the Three Gorges Project,the backwater effect in the Jingjiang Reach diminished.For instance,mean backwater degrees for low,moderate,and high mainstream discharges were recorded as 0.83 m,1.61 m,and 2.41 m during the period from 1990 to 2002,whereas these values decreased to 0.30 m,0.95 m,and 2.08 m from 2009 to 2020.The backwater range extended upstream as mainstream discharge increased from 7000 m3/s to 30000 m3/s.Moreover,a random forest-based machine learning model was used to quantify the backwater effect with varying mainstream and confluence discharges,accounting for the impacts of mainstream discharge,confluence discharge,and channel degradation in the Jingjiang Reach.At the Jianli Hydrological Station,a decrease in mainstream discharge during flood seasons resulted in a 7%–15%increase in monthly mean backwater degree,while an increase in mainstream discharge during dry seasons led to a 1%–15%decrease in monthly mean backwater degree.Furthermore,increasing confluence discharge from Dongting Lake during June to July and September to November resulted in an 11%–42%increase in monthly mean backwater degree.Continuous channel degradation in the Jingjiang Reach contributed to a 6%–19%decrease in monthly mean backwater degree.Under the influence of these factors,the monthly mean backwater degree in 2017 varied from a decrease of 53%to an increase of 37%compared to corresponding values in 1991.
基金supported by the Natural Science Foundation of Jiangsu province,China(BK20240937)the Belt and Road Special Foundation of the National Key Laboratory of Water Disaster Prevention(2022491411,2021491811)the Basal Research Fund of Central Public Welfare Scientific Institution of Nanjing Hydraulic Research Institute(Y223006).
文摘Understanding spatial heterogeneity in groundwater responses to multiple factors is critical for water resource management in coastal cities.Daily groundwater depth(GWD)data from 43 wells(2018-2022)were collected in three coastal cities in Jiangsu Province,China.Seasonal and Trend decomposition using Loess(STL)together with wavelet analysis and empirical mode decomposition were applied to identify tide-influenced wells while remaining wells were grouped by hierarchical clustering analysis(HCA).Machine learning models were developed to predict GWD,then their response to natural conditions and human activities was assessed by the Shapley Additive exPlanations(SHAP)method.Results showed that eXtreme Gradient Boosting(XGB)was superior to other models in terms of prediction performance and computational efficiency(R^(2)>0.95).GWD in Yancheng and southern Lianyungang were greater than those in Nantong,exhibiting larger fluctuations.Groundwater within 5 km of the coastline was affected by tides,with more pronounced effects in agricultural areas compared to urban areas.Shallow groundwater(3-7 m depth)responded immediately(0-1 day)to rainfall,primarily influenced by farmland and topography(slope and distance from rivers).Rainfall recharge to groundwater peaked at 50%farmland coverage,but this effect was suppressed by high temperatures(>30℃)which intensified as distance from rivers increased,especially in forest and grassland.Deep groundwater(>10 m)showed delayed responses to rainfall(1-4 days)and temperature(10-15 days),with GDP as the primary influence,followed by agricultural irrigation and population density.Farmland helped to maintain stable GWD in low population density regions,while excessive farmland coverage(>90%)led to overexploitation.In the early stages of GDP development,increased industrial and agricultural water demand led to GWD decline,but as GDP levels significantly improved,groundwater consumption pressure gradually eased.This methodological framework is applicable not only to coastal cities in China but also could be extended to coastal regions worldwide.
基金supported by General Scientific Research Funding of the Science and Technology Development Fund(FDCT)in Macao(No.0150/2022/A)the Faculty Research Grants of Macao University of Science and Technology(No.FRG-22-074-FIE).
文摘With the rapid development of economy,air pollution caused by industrial expansion has caused serious harm to human health and social development.Therefore,establishing an effective air pollution concentration prediction system is of great scientific and practical significance for accurate and reliable predictions.This paper proposes a combination of pointinterval prediction system for pollutant concentration prediction by leveraging neural network,meta-heuristic optimization algorithm,and fuzzy theory.Fuzzy information granulation technology is used in data preprocessing to transform numerical sequences into fuzzy particles for comprehensive feature extraction.The golden Jackal optimization algorithm is employed in the optimization stage to fine-tune model hyperparameters.In the prediction stage,an ensemble learning method combines training results frommultiplemodels to obtain final point predictions while also utilizing quantile regression and kernel density estimation methods for interval predictions on the test set.Experimental results demonstrate that the combined model achieves a high goodness of fit coefficient of determination(R^(2))at 99.3% and a maximum difference between prediction accuracy mean absolute percentage error(MAPE)and benchmark model at 12.6%.This suggests that the integrated learning system proposed in this paper can provide more accurate deterministic predictions as well as reliable uncertainty analysis compared to traditionalmodels,offering practical reference for air quality early warning.
基金supported by the National Natural Science Foundation of China(Grant Nos.42375062 and 42275158)the National Key Scientific and Technological Infrastructure project“Earth System Science Numerical Simulator Facility”(EarthLab)the Natural Science Foundation of Gansu Province(Grant No.22JR5RF1080)。
文摘It is fundamental and useful to investigate how deep learning forecasting models(DLMs)perform compared to operational oceanography forecast systems(OFSs).However,few studies have intercompared their performances using an identical reference.In this study,three physically reasonable DLMs are implemented for the forecasting of the sea surface temperature(SST),sea level anomaly(SLA),and sea surface velocity in the South China Sea.The DLMs are validated against both the testing dataset and the“OceanPredict”Class 4 dataset.Results show that the DLMs'RMSEs against the latter increase by 44%,245%,302%,and 109%for SST,SLA,current speed,and direction,respectively,compared to those against the former.Therefore,different references have significant influences on the validation,and it is necessary to use an identical and independent reference to intercompare the DLMs and OFSs.Against the Class 4 dataset,the DLMs present significantly better performance for SLA than the OFSs,and slightly better performances for other variables.The error patterns of the DLMs and OFSs show a high degree of similarity,which is reasonable from the viewpoint of predictability,facilitating further applications of the DLMs.For extreme events,the DLMs and OFSs both present large but similar forecast errors for SLA and current speed,while the DLMs are likely to give larger errors for SST and current direction.This study provides an evaluation of the forecast skills of commonly used DLMs and provides an example to objectively intercompare different DLMs.
文摘This research investigates the influence of indoor and outdoor factors on photovoltaic(PV)power generation at Utrecht University to accurately predict PV system performance by identifying critical impact factors and improving renewable energy efficiency.To predict plant efficiency,nineteen variables are analyzed,consisting of nine indoor photovoltaic panel characteristics(Open Circuit Voltage(Voc),Short Circuit Current(Isc),Maximum Power(Pmpp),Maximum Voltage(Umpp),Maximum Current(Impp),Filling Factor(FF),Parallel Resistance(Rp),Series Resistance(Rs),Module Temperature)and ten environmental factors(Air Temperature,Air Humidity,Dew Point,Air Pressure,Irradiation,Irradiation Propagation,Wind Speed,Wind Speed Propagation,Wind Direction,Wind Direction Propagation).This study provides a new perspective not previously addressed in the literature.In this study,different machine learning methods such as Multilayer Perceptron(MLP),Multivariate Adaptive Regression Spline(MARS),Multiple Linear Regression(MLR),and Random Forest(RF)models are used to predict power values using data from installed PVpanels.Panel values obtained under real field conditions were used to train the models,and the results were compared.The Multilayer Perceptron(MLP)model was achieved with the highest classification accuracy of 0.990%.The machine learning models used for solar energy forecasting show high performance and produce results close to actual values.Models like Multi-Layer Perceptron(MLP)and Random Forest(RF)can be used in diverse locations based on load demand.
基金the National Natural Science Foundation of China(NSFC)(Grant Nos.61872016,61932007 and 61972013).
文摘Software systems are present all around us and playing their vital roles in our daily life.The correct functioning of these systems is of prime concern.In addition to classical testing techniques,formal techniques like model checking are used to reinforce the quality and reliability of software systems.However,obtaining of behavior model,which is essential for model-based techniques,of unknown software systems is a challenging task.To mitigate this problem,an emerging black-box analysis technique,called Model Learning,can be applied.It complements existing model-based testing and verification approaches by providing behavior models of blackbox systems fully automatically.This paper surveys the model learning technique,which recently has attracted much attention from researchers,especially from the domains of testing and verification.First,we review the background and foundations of model learning,which form the basis of subsequent sections.Second,we present some well-known model learning tools and provide their merits and shortcomings in the form of a comparison table.Third,we describe the successful applications of model learning in multidisciplinary fields,current challenges along with possible future works,and concluding remarks.
文摘Natural language processing(NLP)is a subfield of artificial intelligence that focuses on enabling computers to understand and process human languages.In the last five years,we have witnessed the rapid development of NLP in tasks such as machine translation,question-answering,and machine reading comprehension based on deep learning and an enormous volume of annotated and unannotated data.In this paper,we will review the latest progress in the neural network-based NLP framework(neural NLP)from three perspectives:modeling,learning,and reasoning.In the modeling section,we will describe several fundamental neural network-based modeling paradigms,such as word embedding,sentence embedding,and sequence-to-sequence modeling,which are widely used in modern NLP engines.In the learning section,we will introduce widely used learning methods for NLP models,including supervised,semi-supervised,and unsupervised learning;multitask learning;transfer learning;and active learning.We view reasoning as a new and exciting direction for neural NLP,but it has yet to be well addressed.In the reasoning section,we will review reasoning mechanisms,including the knowledge,existing non-neural inference methods,and new neural inference methods.We emphasize the importance of reasoning in this paper because it is important for building interpretable and knowledgedriven neural NLP models to handle complex tasks.At the end of this paper,we will briefly outline our thoughts on the future directions of neural NLP.
基金supported by grants from the National Natural Science Foundation of China (81030036, 81270012, 81170809)Special Funds for Major State Basic Research Program of China (973 Program) (2014CB942904) +1 种基金the Committee of Science and Technology of Chongqing (CSTC 2011jjA1468)the foundation from national key laboratory (SKLZZ201017)
文摘Fibroblast growth factor (FGF)/fibroblast growth factor receptor (FGFR) signaling plays essential roles in bone development and diseases. Missense mutations in FGFs and FGFRs in humans can cause various congenital bone diseases, including chondrodysplasia syndromes, craniosynostosis syndromes and syndromes with dysregulated phosphate metabolism. FGF/FGFR signaling is also an important pathway involved in the maintenance of adult bone homeostasis. Multiple kinds of mouse models, mimicking human skeleton diseases caused by missense mutations in FGFs and FGFRs, have been established by knock-m/out and transgenic technologies. These genetically modified mice provide good models for studying the role of FGF/FGFR signaling in skeleton development and homeostasis. In this review, we summarize the mouse models of FGF signaling-related skeleton diseases and recent progresses regarding the molecular mechanisms, underlying the role of FGFs/FGFRs in the regulation of bone development and homeostasis. This review also provides a perspective view on future works to explore the roles of FGF signaling in skeletal development and homeostasis.
基金funded by the Natural Science Foundation of China(Grant Nos.41807285,41972280 and 52179103).
文摘To perform landslide susceptibility prediction(LSP),it is important to select appropriate mapping unit and landslide-related conditioning factors.The efficient and automatic multi-scale segmentation(MSS)method proposed by the authors promotes the application of slope units.However,LSP modeling based on these slope units has not been performed.Moreover,the heterogeneity of conditioning factors in slope units is neglected,leading to incomplete input variables of LSP modeling.In this study,the slope units extracted by the MSS method are used to construct LSP modeling,and the heterogeneity of conditioning factors is represented by the internal variations of conditioning factors within slope unit using the descriptive statistics features of mean,standard deviation and range.Thus,slope units-based machine learning models considering internal variations of conditioning factors(variant slope-machine learning)are proposed.The Chongyi County is selected as the case study and is divided into 53,055 slope units.Fifteen original slope unit-based conditioning factors are expanded to 38 slope unit-based conditioning factors through considering their internal variations.Random forest(RF)and multi-layer perceptron(MLP)machine learning models are used to construct variant Slope-RF and Slope-MLP models.Meanwhile,the Slope-RF and Slope-MLP models without considering the internal variations of conditioning factors,and conventional grid units-based machine learning(Grid-RF and MLP)models are built for comparisons through the LSP performance assessments.Results show that the variant Slopemachine learning models have higher LSP performances than Slope-machine learning models;LSP results of variant Slope-machine learning models have stronger directivity and practical application than Grid-machine learning models.It is concluded that slope units extracted by MSS method can be appropriate for LSP modeling,and the heterogeneity of conditioning factors within slope units can more comprehensively reflect the relationships between conditioning factors and landslides.The research results have important reference significance for land use and landslide prevention.
基金supported in part by the Gusu Innovation and Entrepreneurship Leading Talents in Suzhou City,grant numbers ZXL2021425 and ZXL2022476Doctor of Innovation and Entrepreneurship Program in Jiangsu Province,grant number JSSCBS20211440+6 种基金Jiangsu Province Key R&D Program,grant number BE2019682Natural Science Foundation of Jiangsu Province,grant number BK20200214National Key R&D Program of China,grant number 2017YFB0403701National Natural Science Foundation of China,grant numbers 61605210,61675226,and 62075235Youth Innovation Promotion Association of Chinese Academy of Sciences,grant number 2019320Frontier Science Research Project of the Chinese Academy of Sciences,grant number QYZDB-SSW-JSC03Strategic Priority Research Program of the Chinese Academy of Sciences,grant number XDB02060000.
文摘The prediction of fundus fluorescein angiography(FFA)images from fundus structural images is a cutting-edge research topic in ophthalmological image processing.Prediction comprises estimating FFA from fundus camera imaging,single-phase FFA from scanning laser ophthalmoscopy(SLO),and three-phase FFA also from SLO.Although many deep learning models are available,a single model can only perform one or two of these prediction tasks.To accomplish three prediction tasks using a unified method,we propose a unified deep learning model for predicting FFA images from fundus structure images using a supervised generative adversarial network.The three prediction tasks are processed as follows:data preparation,network training under FFA supervision,and FFA image prediction from fundus structure images on a test set.By comparing the FFA images predicted by our model,pix2pix,and CycleGAN,we demonstrate the remarkable progress achieved by our proposal.The high performance of our model is validated in terms of the peak signal-to-noise ratio,structural similarity index,and mean squared error.
文摘BACKGROUND Bleeding is one of the major complications after endoscopic submucosal dissection(ESD)in early gastric cancer(EGC)patients.There are limited studies on estimating the bleeding risk after ESD using an artificial intelligence system.AIM To derivate and verify the performance of the deep learning model and the clinical model for predicting bleeding risk after ESD in EGC patients.METHODS Patients with EGC who underwent ESD between January 2010 and June 2020 at the Samsung Medical Center were enrolled,and post-ESD bleeding(PEB)was investigated retrospectively.We split the entire cohort into a development set(80%)and a validation set(20%).The deep learning and clinical model were built on the development set and tested in the validation set.The performance of the deep learning model and the clinical model were compared using the area under the curve and the stratification of bleeding risk after ESD.RESULTS A total of 5629 patients were included,and PEB occurred in 325 patients.The area under the curve for predicting PEB was 0.71(95%confidence interval:0.63-0.78)in the deep learning model and 0.70(95%confidence interval:0.62-0.77)in the clinical model,without significant difference(P=0.730).The patients expected to the low-(<5%),intermediate-(≥5%,<9%),and high-risk(≥9%)categories were observed with actual bleeding rate of 2.2%,3.9%,and 11.6%,respectively,in the deep learning model;4.0%,8.8%,and 18.2%,respectively,in the clinical model.CONCLUSION A deep learning model can predict and stratify the bleeding risk after ESD in patients with EGC.
文摘The Indian Himalayan region is frequently experiencing climate change-induced landslides.Thus,landslide susceptibility assessment assumes greater significance for lessening the impact of a landslide hazard.This paper makes an attempt to assess landslide susceptibility in Shimla district of the northwest Indian Himalayan region.It examined the effectiveness of random forest(RF),multilayer perceptron(MLP),sequential minimal optimization regression(SMOreg)and bagging ensemble(B-RF,BSMOreg,B-MLP)models.A landslide inventory map comprising 1052 locations of past landslide occurrences was classified into training(70%)and testing(30%)datasets.The site-specific influencing factors were selected by employing a multicollinearity test.The relationship between past landslide occurrences and influencing factors was established using the frequency ratio method.The effectiveness of machine learning models was verified through performance assessors.The landslide susceptibility maps were validated by the area under the receiver operating characteristic curves(ROC-AUC),accuracy,precision,recall and F1-score.The key performance metrics and map validation demonstrated that the BRF model(correlation coefficient:0.988,mean absolute error:0.010,root mean square error:0.058,relative absolute error:2.964,ROC-AUC:0.947,accuracy:0.778,precision:0.819,recall:0.917 and F-1 score:0.865)outperformed the single classifiers and other bagging ensemble models for landslide susceptibility.The results show that the largest area was found under the very high susceptibility zone(33.87%),followed by the low(27.30%),high(20.68%)and moderate(18.16%)susceptibility zones.The factors,namely average annual rainfall,slope,lithology,soil texture and earthquake magnitude have been identified as the influencing factors for very high landslide susceptibility.Soil texture,lineament density and elevation have been attributed to high and moderate susceptibility.Thus,the study calls for devising suitable landslide mitigation measures in the study area.Structural measures,an immediate response system,community participation and coordination among stakeholders may help lessen the detrimental impact of landslides.The findings from this study could aid decision-makers in mitigating future catastrophes and devising suitable strategies in other geographical regions with similar geological characteristics.
基金This study has been reviewed and approved by the Clinical Research Ethics Committee of Wenzhou Central Hospital and the First Hospital Affiliated to Wenzhou Medical University,No.KY2024-R016.
文摘BACKGROUND Colorectal cancer significantly impacts global health,with unplanned reoperations post-surgery being key determinants of patient outcomes.Existing predictive models for these reoperations lack precision in integrating complex clinical data.AIM To develop and validate a machine learning model for predicting unplanned reoperation risk in colorectal cancer patients.METHODS Data of patients treated for colorectal cancer(n=2044)at the First Affiliated Hospital of Wenzhou Medical University and Wenzhou Central Hospital from March 2020 to March 2022 were retrospectively collected.Patients were divided into an experimental group(n=60)and a control group(n=1984)according to unplanned reoperation occurrence.Patients were also divided into a training group and a validation group(7:3 ratio).We used three different machine learning methods to screen characteristic variables.A nomogram was created based on multifactor logistic regression,and the model performance was assessed using receiver operating characteristic curve,calibration curve,Hosmer-Lemeshow test,and decision curve analysis.The risk scores of the two groups were calculated and compared to validate the model.RESULTS More patients in the experimental group were≥60 years old,male,and had a history of hypertension,laparotomy,and hypoproteinemia,compared to the control group.Multiple logistic regression analysis confirmed the following as independent risk factors for unplanned reoperation(P<0.05):Prognostic Nutritional Index value,history of laparotomy,hypertension,or stroke,hypoproteinemia,age,tumor-node-metastasis staging,surgical time,gender,and American Society of Anesthesiologists classification.Receiver operating characteristic curve analysis showed that the model had good discrimination and clinical utility.CONCLUSION This study used a machine learning approach to build a model that accurately predicts the risk of postoperative unplanned reoperation in patients with colorectal cancer,which can improve treatment decisions and prognosis.
基金Project supported by the National Natural Science Foundation of China(Nos.12101072 and 11421101)the National Key Research and Development Program of China(No.2018YFF0300104)+1 种基金the Beijing Municipal Science and Technology Project(No.Z201100005820002)the Open Research Fund of Shenzhen Research Institute of Big Data(No.2019ORF01001)。
文摘Fine-grained weather forecasting data,i.e.,the grid data with high-resolution,have attracted increasing attention in recent years,especially for some specific applications such as the Winter Olympic Games.Although European Centre for Medium-Range Weather Forecasts(ECMWF)provides grid prediction up to 240 hours,the coarse data are unable to meet high requirements of these major events.In this paper,we propose a method,called model residual machine learning(MRML),to generate grid prediction with high-resolution based on high-precision stations forecasting.MRML applies model output machine learning(MOML)for stations forecasting.Subsequently,MRML utilizes these forecasts to improve the quality of the grid data by fitting a machine learning(ML)model to the residuals.We demonstrate that MRML achieves high capability at diverse meteorological elements,specifically,temperature,relative humidity,and wind speed.In addition,MRML could be easily extended to other post-processing methods by invoking different techniques.In our experiments,MRML outperforms the traditional downscaling methods such as piecewise linear interpolation(PLI)on the testing data.