Engineering tests can yield inaccurate data due to instrument errors,human factors,and environmental interference,introducing uncertainty in numerical model updating.This study employs the probability-box(p-box)method...Engineering tests can yield inaccurate data due to instrument errors,human factors,and environmental interference,introducing uncertainty in numerical model updating.This study employs the probability-box(p-box)method for representing observational uncertainty and develops a two-step approximate Bayesian computation(ABC)framework using time-series data.Within the ABC framework,Euclidean and Bhattacharyya distances are employed as uncertainty quantification metrics to delineate approximate likelihood functions in the initial and subsequent steps,respectively.A novel variational Bayesian Monte Carlo method is introduced to efficiently apply the ABC framework amidst observational uncertainty,resulting in rapid convergence and accurate parameter estimation with minimal iterations.The efficacy of the proposed updating strategy is validated by its application to a shear frame model excited by seismic wave and an aviation pump force sensor for thermal output analysis.The results affirm the efficiency,robustness,and practical applicability of the proposed method.展开更多
Assessing the stability of slopes is one of the crucial tasks of geotechnical engineering for assessing and managing risks related to natural hazards,directly affecting safety and sustainable development.This study pr...Assessing the stability of slopes is one of the crucial tasks of geotechnical engineering for assessing and managing risks related to natural hazards,directly affecting safety and sustainable development.This study primarily focuses on developing robust and practical hybrid models to predict the slope stability status of circular failure mode.For this purpose,three robust models were developed using a database including 627 case histories of slope stability status.The models were developed using the random forest(RF),support vector machine(SVM),and extreme gradient boosting(XGB)techniques,employing 5-fold cross validation approach.To enhance the performance of models,this study employs Bayesian optimizer(BO)to fine-tuning their hyperparameters.The results indicate that the performance order of the three developed models is RF-BO>SVM-BO>XGB-BO.Furthermore,comparing the developed models with previous models,it was found that the RF-BO model can effectively determine the slope stability status with outstanding performance.This implies that the RF-BO model could serve as a dependable tool for project managers,assisting in the evaluation of slope stability during both the design and operational phases of projects,despite the inherent challenges in this domain.The results regarding the importance of influencing parameters indicate that cohesion,friction angle,and slope height exert the most significant impact on slope stability status.This suggests that concentrating on these parameters and employing the RF-BO model can effectively mitigate the severity of geohazards in the short-term and contribute to the attainment of long-term sustainable development objectives.展开更多
This study investigated forest recovery in the Atlantic Rainforest and Rupestrian Grassland of Brazil using the diffusive-logistic growth(DLG)model.This model simulates vegetation growth in the two mountain biomes con...This study investigated forest recovery in the Atlantic Rainforest and Rupestrian Grassland of Brazil using the diffusive-logistic growth(DLG)model.This model simulates vegetation growth in the two mountain biomes considering spatial location,time,and two key parameters:diffusion rate and growth rate.A Bayesian framework is employed to analyze the model's parameters and assess prediction uncertainties.Satellite imagery from 1992 and 2022 was used for model calibration and validation.By solving the DLG model using the finite difference method,we predicted a 6.6%–51.1%increase in vegetation density for the Atlantic Rainforest and a 5.3%–99.9%increase for the Rupestrian Grassland over 30 years,with the latter showing slower recovery but achieving a better model fit(lower RMSE)compared to the Atlantic Rainforest.The Bayesian approach revealed well-defined parameter distributions and lower parameter values for the Rupestrian Grassland,supporting the slower recovery prediction.Importantly,the model achieved good agreement with observed vegetation patterns in unseen validation data for both biomes.While there were minor spatial variations in accuracy,the overall distributions of predicted and observed vegetation density were comparable.Furthermore,this study highlights the importance of considering uncertainty in model predictions.Bayesian inference allowed us to quantify this uncertainty,demonstrating that the model's performance can vary across locations.Our approach provides valuable insights into forest regeneration process uncertainties,enabling comparisons of modeled scenarios at different recovery stages for better decision-making in these critical mountain biomes.展开更多
In this paper,we propose a hybrid decode-and-forward and soft information relaying(HDFSIR)strategy to mitigate error propagation in coded cooperative communications.In the HDFSIR approach,the relay operates in decode-...In this paper,we propose a hybrid decode-and-forward and soft information relaying(HDFSIR)strategy to mitigate error propagation in coded cooperative communications.In the HDFSIR approach,the relay operates in decode-and-forward(DF)mode when it successfully decodes the received message;otherwise,it switches to soft information relaying(SIR)mode.The benefits of the DF and SIR forwarding strategies are combined to achieve better performance than deploying the DF or SIR strategy alone.Closed-form expressions for the outage probability and symbol error rate(SER)are derived for coded cooperative communication with HDFSIR and energy-harvesting relays.Additionally,we introduce a novel normalized log-likelihood-ratio based soft estimation symbol(NL-SES)mapping technique,which enhances soft symbol accuracy for higher-order modulation,and propose a model characterizing the relationship between the estimated complex soft symbol and the actual high-order modulated symbol.Further-more,the hybrid DF-SIR strategy is extended to a distributed Alamouti space-time-coded cooperative network.To evaluate the~performance of the proposed HDFSIR strategy,we implement extensive Monte Carlo simulations under varying channel conditions.Results demonstrate significant improvements with the hybrid technique outperforming individual DF and SIR strategies in both conventional and distributed Alamouti space-time coded cooperative networks.Moreover,at a SER of 10^(-3),the proposed NL-SES mapping demonstrated a 3.5 dB performance gain over the conventional averaging one,highlighting its superior accuracy in estimating soft symbols for quadrature phase-shift keying modulation.展开更多
Challenges in stratigraphic modeling arise from underground uncertainty.While borehole exploration is reliable,it remains sparse due to economic and site constraints.Electrical resistivity tomography(ERT)as a cost-eff...Challenges in stratigraphic modeling arise from underground uncertainty.While borehole exploration is reliable,it remains sparse due to economic and site constraints.Electrical resistivity tomography(ERT)as a cost-effective geophysical technique can acquire high-density data;however,uncertainty and nonuniqueness inherent in ERT impede its usage for stratigraphy identification.This paper integrates ERT and onsite observations for the first time to propose a novel method for characterizing stratigraphic profiles.The method consists of two steps:(1)ERT for prior knowledge:ERT data are processed by soft clustering using the Gaussian mixture model,followed by probability smoothing to quantify its depthdependent uncertainty;and(2)Observations for calibration:a spatial sequential Bayesian updating(SSBU)algorithm is developed to update the prior knowledge based on likelihoods derived from onsite observations,namely topsoil and boreholes.The effectiveness of the proposed method is validated through its application to a real slope site in Foshan,China.Comparative analysis with advanced borehole-driven methods highlights the superiority of incorporating ERT data in stratigraphic modeling,in terms of prediction accuracy at borehole locations and sensitivity to borehole data.Informed by ERT,reduced sensitivity to boreholes provides a fundamental solution to the longstanding challenge of sparse measurements.The paper further discusses the impact of ERT uncertainty on the proposed model using time-lapse measurements,the impact of model resolution,and applicability in engineering projects.This study,as a breakthrough in stratigraphic modeling,bridges gaps in combining geophysical and geotechnical data to address measurement sparsity and paves the way for more economical geotechnical exploration.展开更多
To ensure agreement between theoretical calculations and experimental data,parameters to selected nuclear physics models are perturbed and fine-tuned in nuclear data evaluations.This approach assumes that the chosen s...To ensure agreement between theoretical calculations and experimental data,parameters to selected nuclear physics models are perturbed and fine-tuned in nuclear data evaluations.This approach assumes that the chosen set of models accurately represents the‘true’distribution of considered observables.Furthermore,the models are chosen globally,indicating their applicability across the entire energy range of interest.However,this approach overlooks uncertainties inherent in the models themselves.In this work,we propose that instead of selecting globally a winning model set and proceeding with it as if it was the‘true’model set,we,instead,take a weighted average over multiple models within a Bayesian model averaging(BMA)framework,each weighted by its posterior probability.The method involves executing a set of TALYS calculations by randomly varying multiple nuclear physics models and their parameters to yield a vector of calculated observables.Next,computed likelihood function values at each incident energy point were then combined with the prior distributions to obtain updated posterior distributions for selected cross sections and the elastic angular distributions.As the cross sections and elastic angular distributions were updated locally on a per-energy-point basis,the approach typically results in discontinuities or“kinks”in the cross section curves,and these were addressed using spline interpolation.The proposed BMA method was applied to the evaluation of proton-induced reactions on ^(58)Ni between 1 and 100 MeV.The results demonstrated a favorable comparison with experimental data as well as with the TENDL-2023 evaluation.展开更多
BACKGROUND Portal hypertension(PHT),primarily induced by cirrhosis,manifests severe symptoms impacting patient survival.Although transjugular intrahepatic portosystemic shunt(TIPS)is a critical intervention for managi...BACKGROUND Portal hypertension(PHT),primarily induced by cirrhosis,manifests severe symptoms impacting patient survival.Although transjugular intrahepatic portosystemic shunt(TIPS)is a critical intervention for managing PHT,it carries risks like hepatic encephalopathy,thus affecting patient survival prognosis.To our knowledge,existing prognostic models for post-TIPS survival in patients with PHT fail to account for the interplay among and collective impact of various prognostic factors on outcomes.Consequently,the development of an innovative modeling approach is essential to address this limitation.AIM To develop and validate a Bayesian network(BN)-based survival prediction model for patients with cirrhosis-induced PHT having undergone TIPS.METHODS The clinical data of 393 patients with cirrhosis-induced PHT who underwent TIPS surgery at the Second Affiliated Hospital of Chongqing Medical University between January 2015 and May 2022 were retrospectively analyzed.Variables were selected using Cox and least absolute shrinkage and selection operator regression methods,and a BN-based model was established and evaluated to predict survival in patients having undergone TIPS surgery for PHT.RESULTS Variable selection revealed the following as key factors impacting survival:age,ascites,hypertension,indications for TIPS,postoperative portal vein pressure(post-PVP),aspartate aminotransferase,alkaline phosphatase,total bilirubin,prealbumin,the Child-Pugh grade,and the model for end-stage liver disease(MELD)score.Based on the above-mentioned variables,a BN-based 2-year survival prognostic prediction model was constructed,which identified the following factors to be directly linked to the survival time:age,ascites,indications for TIPS,concurrent hypertension,post-PVP,the Child-Pugh grade,and the MELD score.The Bayesian information criterion was 3589.04,and 10-fold cross-validation indicated an average log-likelihood loss of 5.55 with a standard deviation of 0.16.The model’s accuracy,precision,recall,and F1 score were 0.90,0.92,0.97,and 0.95 respectively,with the area under the receiver operating characteristic curve being 0.72.CONCLUSION This study successfully developed a BN-based survival prediction model with good predictive capabilities.It offers valuable insights for treatment strategies and prognostic evaluations in patients having undergone TIPS surgery for PHT.展开更多
The emerging virtual coupling technology aims to operate multiple train units in a Virtually Coupled Train Set(VCTS)at a minimal but safe distance.To guarantee collision avoidance,the safety distance should be calcula...The emerging virtual coupling technology aims to operate multiple train units in a Virtually Coupled Train Set(VCTS)at a minimal but safe distance.To guarantee collision avoidance,the safety distance should be calculated using the state-of-the-art space-time separation principle that separates the Emergency Braking(EB)trajectories of two successive units during the whole EB process.In this case,the minimal safety distance is usually numerically calculated without an analytic formulation.Thus,the constrained VCTS control problem is hard to address with space-time separation,which is still a gap in the existing literature.To solve this problem,we propose a Distributed Economic Model Predictive Control(DEMPC)approach with computation efficiency and theoretical guarantee.Specifically,to alleviate the computation burden,we transform implicit safety constraints into explicitly linear ones,such that the optimal control problem in DEMPC is a quadratic programming problem that can be solved efficiently.For theoretical analysis,sufficient conditions are derived to guarantee the recursive feasibility and stability of DEMPC,employing compatibility constraints,tube techniques and terminal ingredient tuning.Moreover,we extend our approach with globally optimal and distributed online EB configuration methods to shorten the minimal distance among VCTS.Finally,experimental results demonstrate the performance and advantages of the proposed approaches.展开更多
Stable water isotopes are natural tracers quantifying the contribution of moisture recycling to local precipitation,i.e.,the moisture recycling ratio,but various isotope-based models usually lead to different results,...Stable water isotopes are natural tracers quantifying the contribution of moisture recycling to local precipitation,i.e.,the moisture recycling ratio,but various isotope-based models usually lead to different results,which affects the accuracy of local moisture recycling.In this study,a total of 18 stations from four typical areas in China were selected to compare the performance of isotope-based linear and Bayesian mixing models and to determine local moisture recycling ratio.Among the three vapor sources including advection,transpiration,and surface evaporation,the advection vapor usually played a dominant role,and the contribution of surface evaporation was less than that of transpiration.When the abnormal values were ignored,the arithmetic averages of differences between isotope-based linear and the Bayesian mixing models were 0.9%for transpiration,0.2%for surface evaporation,and–1.1%for advection,respectively,and the medians were 0.5%,0.2%,and–0.8%,respectively.The importance of transpiration was slightly less for most cases when the Bayesian mixing model was applied,and the contribution of advection was relatively larger.The Bayesian mixing model was found to perform better in determining an efficient solution since linear model sometimes resulted in negative contribution ratios.Sensitivity test with two isotope scenarios indicated that the Bayesian model had a relatively low sensitivity to the changes in isotope input,and it was important to accurately estimate the isotopes in precipitation vapor.Generally,the Bayesian mixing model should be recommended instead of a linear model.The findings are useful for understanding the performance of isotope-based linear and Bayesian mixing models under various climate backgrounds.展开更多
We apply stochastic seismic inversion and Bayesian facies classification for porosity modeling and igneous rock identification in the presalt interval of the Santos Basin. This integration of seismic and well-derived ...We apply stochastic seismic inversion and Bayesian facies classification for porosity modeling and igneous rock identification in the presalt interval of the Santos Basin. This integration of seismic and well-derived information enhances reservoir characterization. Stochastic inversion and Bayesian classification are powerful tools because they permit addressing the uncertainties in the model. We used the ES-MDA algorithm to achieve the realizations equivalent to the percentiles P10, P50, and P90 of acoustic impedance, a novel method for acoustic inversion in presalt. The facies were divided into five: reservoir 1,reservoir 2, tight carbonates, clayey rocks, and igneous rocks. To deal with the overlaps in acoustic impedance values of facies, we included geological information using a priori probability, indicating that structural highs are reservoir-dominated. To illustrate our approach, we conducted porosity modeling using facies-related rock-physics models for rock-physics inversion in an area with a well drilled in a coquina bank and evaluated the thickness and extension of an igneous intrusion near the carbonate-salt interface. The modeled porosity and the classified seismic facies are in good agreement with the ones observed in the wells. Notably, the coquinas bank presents an improvement in the porosity towards the top. The a priori probability model was crucial for limiting the clayey rocks to the structural lows. In Well B, the hit rate of the igneous rock in the three scenarios is higher than 60%, showing an excellent thickness-prediction capability.展开更多
Xinjiang Uygur Autonomous Region is a typical inland arid area in China with a sparse and uneven distribution of meteorological stations,limited access to precipitation data,and significant water scarcity.Evaluating a...Xinjiang Uygur Autonomous Region is a typical inland arid area in China with a sparse and uneven distribution of meteorological stations,limited access to precipitation data,and significant water scarcity.Evaluating and integrating precipitation datasets from different sources to accurately characterize precipitation patterns has become a challenge to provide more accurate and alternative precipitation information for the region,which can even improve the performance of hydrological modelling.This study evaluated the applicability of widely used five satellite-based precipitation products(Climate Hazards Group InfraRed Precipitation with Station(CHIRPS),China Meteorological Forcing Dataset(CMFD),Climate Prediction Center morphing method(CMORPH),Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks-Climate Data Record(PERSIANN-CDR),and Tropical Rainfall Measuring Mission Multi-satellite Precipitation Analysis(TMPA))and a reanalysis precipitation dataset(ECMWF Reanalysis v5-Land Dataset(ERA5-Land))in Xinjiang using ground-based observational precipitation data from a limited number of meteorological stations.Based on this assessment,we proposed a framework that integrated different precipitation datasets with varying spatial resolutions using a dynamic Bayesian model averaging(DBMA)approach,the expectation-maximization method,and the ordinary Kriging interpolation method.The daily precipitation data merged using the DBMA approach exhibited distinct spatiotemporal variability,with an outstanding performance,as indicated by low root mean square error(RMSE=1.40 mm/d)and high Person's correlation coefficient(CC=0.67).Compared with the traditional simple model averaging(SMA)and individual product data,although the DBMA-fused precipitation data were slightly lower than the best precipitation product(CMFD),the overall performance of DBMA was more robust.The error analysis between DBMA-fused precipitation dataset and the more advanced Integrated Multi-satellite Retrievals for Global Precipitation Measurement Final(IMERG-F)precipitation product,as well as hydrological simulations in the Ebinur Lake Basin,further demonstrated the superior performance of DBMA-fused precipitation dataset in the entire Xinjiang region.The proposed framework for solving the fusion problem of multi-source precipitation data with different spatial resolutions is feasible for application in inland arid areas,and aids in obtaining more accurate regional hydrological information and improving regional water resources management capabilities and meteorological research in these regions.展开更多
Statistical biases may be introduced by imprecisely quantifying background radiation reference levels. It is, therefore, imperative to devise a simple, adaptable approach for precisely describing the reference backgro...Statistical biases may be introduced by imprecisely quantifying background radiation reference levels. It is, therefore, imperative to devise a simple, adaptable approach for precisely describing the reference background levels of naturally occurring radionuclides (NOR) in mining sites. As a substitute statistical method, we suggest using Bayesian modeling in this work to examine the spatial distribution of NOR. For naturally occurring gamma-induced radionuclides like 232Th, 40K, and 238U, statistical parameters are inferred using the Markov Chain Monte Carlo (MCMC) method. After obtaining an accurate subsample using bootstrapping, we exclude any possible outliers that fall outside of the Highest Density Interval (HDI). We use MCMC to build a Bayesian model with the resampled data and make predictions about the posterior distribution of radionuclides produced by gamma irradiation. This method offers a strong and dependable way to describe NOR reference background values, which is important for managing and evaluating radiation risks in mining contexts.展开更多
To enhance the efficiency of vaccine manufacturing,this study focuses on optimizing the microfluidic conditions and lipid mix ratios of messenger RNA-lipid nanoparticles(mRNA-LNP).Different mRNA-LNP formulations(n=24)...To enhance the efficiency of vaccine manufacturing,this study focuses on optimizing the microfluidic conditions and lipid mix ratios of messenger RNA-lipid nanoparticles(mRNA-LNP).Different mRNA-LNP formulations(n=24)were developed using an I-optimal design,where machine learning tools(XGBoost/Bayesian optimization and self-validated ensemble(SVEM))were used to optimize the process and predict lipid mix ratio.The investigation included material attributes,their respective ratios,and process attributes.The critical responses like particle size(PS),polydispersity index(PDI),Zeta potential,pKa,heat trend cycle,encapsulation efficiency(EE),recovery ratio,and encapsulated mRNA were evaluated.Overall prediction of SVEM(>97%)was comparably better than that of XGBoost/Bayesian optimization(>94%).Moreover,in actual experimental outcomes,SVEM prediction is close to the actual data as confirmed by the experimental PS(94-96 nm)is close to the predicted one(95-97 nm).The other parameters including PDI and EE were also close to the actual experimental data.展开更多
In recent years,moving target detection methods based on low-rank and sparse matrix decomposition have been developed,and they have achieved good results.However,there is not enough interpretation to support the assum...In recent years,moving target detection methods based on low-rank and sparse matrix decomposition have been developed,and they have achieved good results.However,there is not enough interpretation to support the assumption that there is a high correlation among the reverberations after each transmitting pulse.In order to explain the correlation of reverberations,a new reverberation model is proposed from the perspective of scattering cells in this paper.The scattering cells are the subarea divided from the detection area.The energy fluctuation of a scattering cell with time and the influence of the neighboring cells are considered.Key parameters of the model were analyzed by numerical analysis,and the applicability of the model was verified by experimental analysis.The results showed that the model can be used for several simulations to evaluate the performance of moving target detection methods.展开更多
The Bayesian structural equation model integrates the principles of Bayesian statistics, providing a more flexible and comprehensive modeling framework. In exploring complex relationships between variables, handling u...The Bayesian structural equation model integrates the principles of Bayesian statistics, providing a more flexible and comprehensive modeling framework. In exploring complex relationships between variables, handling uncertainty, and dealing with missing data, the Bayesian structural equation model demonstrates unique advantages. Therefore, Bayesian methods are used in this paper to establish a structural equation model of innovative talent cognition, with the measurement of college students’ cognition of innovative talent being studied. An in-depth analysis is conducted on the effects of innovative self-efficacy, social resources, innovative personality traits, and school education, aiming to explore the factors influencing college students’ innovative talent. The results indicate that innovative self-efficacy plays a key role in perception, social resources are significantly positively correlated with the perception of innovative talents, innovative personality tendencies and school education are positively correlated with the perception of innovative talents, but the impact is not significant.展开更多
We consider the two-point,two-time(space-time)correlation of passive scalar R(r,τ)in the Kraichnan model under the assumption of homogeneity and isotropy.Using the fine-gird PDF method,we find that R(r,τ)satisfies a...We consider the two-point,two-time(space-time)correlation of passive scalar R(r,τ)in the Kraichnan model under the assumption of homogeneity and isotropy.Using the fine-gird PDF method,we find that R(r,τ)satisfies a diffusion equation with constant diffusion coefficient determined by velocity variance and molecular diffusion.Itssolution can be expressed in terms of the two-point,one time correlation of passive scalar,i.e.,R(r,0).Moreover,the decorrelation o R(k,τ),which is the Fourier transform of R(r,τ),is determined byR(k,0)and a diffusion kernal.展开更多
In this paper,the forecasting equations of a 2nd-order space-time differential remainder are deduced from the Navier-Stokes primitive equations and Eulerian operator by Taylor-series expansion.Here we introduce a cubi...In this paper,the forecasting equations of a 2nd-order space-time differential remainder are deduced from the Navier-Stokes primitive equations and Eulerian operator by Taylor-series expansion.Here we introduce a cubic spline numerical model(Spline Model for short),which is with a quasi-Lagrangian time-split integration scheme of fitting cubic spline/bicubic surface to all physical variable fields in the atmospheric equations on spherical discrete latitude-longitude mesh.A new algorithm of"fitting cubic spline—time step integration—fitting cubic spline—……"is developed to determine their first-and2nd-order derivatives and their upstream points for time discrete integral to the governing equations in Spline Model.And the cubic spline function and its mathematical polarities are also discussed to understand the Spline Model’s mathematical foundation of numerical analysis.It is pointed out that the Spline Model has mathematical laws of"convergence"of the cubic spline functions contracting to the original functions as well as its 1st-order and 2nd-order derivatives.The"optimality"of the 2nd-order derivative of the cubic spline functions is optimal approximation to that of the original functions.In addition,a Hermite bicubic patch is equivalent to operate on a grid for a 2nd-order derivative variable field.Besides,the slopes and curvatures of a central difference are identified respectively,with a smoothing coefficient of 1/3,three-point smoothing of that of a cubic spline.Then the slopes and curvatures of a central difference are calculated from the smoothing coefficient 1/3 and three-point smoothing of that of a cubic spline,respectively.Furthermore,a global simulation case of adiabatic,non-frictional and"incompressible"model atmosphere is shown with the quasi-Lagrangian time integration by using a global Spline Model,whose initial condition comes from the NCEP reanalysis data,along with quasi-uniform latitude-longitude grids and the so-called"shallow atmosphere"Navier-Stokes primitive equations in the spherical coordinates.The Spline Model,which adopted the Navier-Stokes primitive equations and quasi-Lagrangian time-split integration scheme,provides an initial ideal case of global atmospheric circulation.In addition,considering the essentially non-linear atmospheric motions,the Spline Model could judge reasonably well simple points of any smoothed variable field according to its fitting spline curvatures that must conform to its physical interpretation.展开更多
Objective To investigate the spatiotemporal patterns and socioeconomic factors influencing the incidence of tuberculosis(TB)in the Guangdong Province between 2010 and 2019.Method Spatial and temporal variations in TB ...Objective To investigate the spatiotemporal patterns and socioeconomic factors influencing the incidence of tuberculosis(TB)in the Guangdong Province between 2010 and 2019.Method Spatial and temporal variations in TB incidence were mapped using heat maps and hierarchical clustering.Socioenvironmental influencing factors were evaluated using a Bayesian spatiotemporal conditional autoregressive(ST-CAR)model.Results Annual incidence of TB in Guangdong decreased from 91.85/100,000 in 2010 to 53.06/100,000in 2019.Spatial hotspots were found in northeastern Guangdong,particularly in Heyuan,Shanwei,and Shantou,while Shenzhen,Dongguan,and Foshan had the lowest rates in the Pearl River Delta.The STCAR model showed that the TB risk was lower with higher per capita Gross Domestic Product(GDP)[Relative Risk(RR),0.91;95%Confidence Interval(CI):0.86–0.98],more the ratio of licensed physicians and physician(RR,0.94;95%CI:0.90-0.98),and higher per capita public expenditure(RR,0.94;95%CI:0.90–0.97),with a marginal effect of population density(RR,0.86;95%CI:0.86–1.00).Conclusion The incidence of TB in Guangdong varies spatially and temporally.Areas with poor economic conditions and insufficient healthcare resources are at an increased risk of TB infection.Strategies focusing on equitable health resource distribution and economic development are the key to TB control.展开更多
Traffic emissions have become the major air pollution source in urban areas.Therefore,understanding the highly non-stational and complex impact of traffic factors on air quality is very important for building air qual...Traffic emissions have become the major air pollution source in urban areas.Therefore,understanding the highly non-stational and complex impact of traffic factors on air quality is very important for building air quality prediction models.Using real-world air pollutant data from Taipei City,this study integrates diverse factors,including traffic flow,speed,rainfall patterns,andmeteorological factors.We constructed a Bayesian network probabilitymodel based on rainfall events as a big data analysis framework to investigate understand traffic factor causality relationships and condition probabilities for meteorological factors and air pollutant concentrations.Generalized Additive Model(GAM)verified non-linear relationships between traffic factors and air pollutants.Consequently,we propose a long short term memory(LSTM)model to predict airborne pollutant concentrations.This study propose a new approach of air pollutants and meteorological variable analysis procedure by considering both rainfall amount and patterns.Results indicate improved air quality when controlling vehicle speed above 40 km/h and maintaining an average vehicle flow<1200 vehicles per hour.This study also classified rainfall events into four types depending on its characteristic.Wet deposition from varied rainfall types significantly affects air quality,with TypeⅠrainfall events(long-duration heavy rain)having the most pronounced impact.An LSTM model incorporating GAM and Bayesian network outcomes yields excellent performance,achieving correlation R^(2)>0.9 and 0.8 for first and second order air pollutants,i.e.,CO,NO,NO_(2),and NO_(x);and O_(3),PM_(10),and PM_(2.5),respectively.展开更多
基金supported by the National Natural Science Foundation of China(Grant No.U23B20105).
文摘Engineering tests can yield inaccurate data due to instrument errors,human factors,and environmental interference,introducing uncertainty in numerical model updating.This study employs the probability-box(p-box)method for representing observational uncertainty and develops a two-step approximate Bayesian computation(ABC)framework using time-series data.Within the ABC framework,Euclidean and Bhattacharyya distances are employed as uncertainty quantification metrics to delineate approximate likelihood functions in the initial and subsequent steps,respectively.A novel variational Bayesian Monte Carlo method is introduced to efficiently apply the ABC framework amidst observational uncertainty,resulting in rapid convergence and accurate parameter estimation with minimal iterations.The efficacy of the proposed updating strategy is validated by its application to a shear frame model excited by seismic wave and an aviation pump force sensor for thermal output analysis.The results affirm the efficiency,robustness,and practical applicability of the proposed method.
文摘Assessing the stability of slopes is one of the crucial tasks of geotechnical engineering for assessing and managing risks related to natural hazards,directly affecting safety and sustainable development.This study primarily focuses on developing robust and practical hybrid models to predict the slope stability status of circular failure mode.For this purpose,three robust models were developed using a database including 627 case histories of slope stability status.The models were developed using the random forest(RF),support vector machine(SVM),and extreme gradient boosting(XGB)techniques,employing 5-fold cross validation approach.To enhance the performance of models,this study employs Bayesian optimizer(BO)to fine-tuning their hyperparameters.The results indicate that the performance order of the three developed models is RF-BO>SVM-BO>XGB-BO.Furthermore,comparing the developed models with previous models,it was found that the RF-BO model can effectively determine the slope stability status with outstanding performance.This implies that the RF-BO model could serve as a dependable tool for project managers,assisting in the evaluation of slope stability during both the design and operational phases of projects,despite the inherent challenges in this domain.The results regarding the importance of influencing parameters indicate that cohesion,friction angle,and slope height exert the most significant impact on slope stability status.This suggests that concentrating on these parameters and employing the RF-BO model can effectively mitigate the severity of geohazards in the short-term and contribute to the attainment of long-term sustainable development objectives.
基金financial support from the Brazilian National Council for Scientific and Technological Development(CNPq)and the Federal University of Ouro PretoFinancial support from the Minas Gerais Research Foundation(FAPEMIG)under grant number APQ-06559-24 is also gratefully acknowledged。
文摘This study investigated forest recovery in the Atlantic Rainforest and Rupestrian Grassland of Brazil using the diffusive-logistic growth(DLG)model.This model simulates vegetation growth in the two mountain biomes considering spatial location,time,and two key parameters:diffusion rate and growth rate.A Bayesian framework is employed to analyze the model's parameters and assess prediction uncertainties.Satellite imagery from 1992 and 2022 was used for model calibration and validation.By solving the DLG model using the finite difference method,we predicted a 6.6%–51.1%increase in vegetation density for the Atlantic Rainforest and a 5.3%–99.9%increase for the Rupestrian Grassland over 30 years,with the latter showing slower recovery but achieving a better model fit(lower RMSE)compared to the Atlantic Rainforest.The Bayesian approach revealed well-defined parameter distributions and lower parameter values for the Rupestrian Grassland,supporting the slower recovery prediction.Importantly,the model achieved good agreement with observed vegetation patterns in unseen validation data for both biomes.While there were minor spatial variations in accuracy,the overall distributions of predicted and observed vegetation density were comparable.Furthermore,this study highlights the importance of considering uncertainty in model predictions.Bayesian inference allowed us to quantify this uncertainty,demonstrating that the model's performance can vary across locations.Our approach provides valuable insights into forest regeneration process uncertainties,enabling comparisons of modeled scenarios at different recovery stages for better decision-making in these critical mountain biomes.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2024-02-02160).
文摘In this paper,we propose a hybrid decode-and-forward and soft information relaying(HDFSIR)strategy to mitigate error propagation in coded cooperative communications.In the HDFSIR approach,the relay operates in decode-and-forward(DF)mode when it successfully decodes the received message;otherwise,it switches to soft information relaying(SIR)mode.The benefits of the DF and SIR forwarding strategies are combined to achieve better performance than deploying the DF or SIR strategy alone.Closed-form expressions for the outage probability and symbol error rate(SER)are derived for coded cooperative communication with HDFSIR and energy-harvesting relays.Additionally,we introduce a novel normalized log-likelihood-ratio based soft estimation symbol(NL-SES)mapping technique,which enhances soft symbol accuracy for higher-order modulation,and propose a model characterizing the relationship between the estimated complex soft symbol and the actual high-order modulated symbol.Further-more,the hybrid DF-SIR strategy is extended to a distributed Alamouti space-time-coded cooperative network.To evaluate the~performance of the proposed HDFSIR strategy,we implement extensive Monte Carlo simulations under varying channel conditions.Results demonstrate significant improvements with the hybrid technique outperforming individual DF and SIR strategies in both conventional and distributed Alamouti space-time coded cooperative networks.Moreover,at a SER of 10^(-3),the proposed NL-SES mapping demonstrated a 3.5 dB performance gain over the conventional averaging one,highlighting its superior accuracy in estimating soft symbols for quadrature phase-shift keying modulation.
基金the financial support from the National Key R&D Program of China(Grant No.2021YFC3001003)Science and Technology Development Fund,Macao SAR(File No.0056/2023/RIB2)Guangdong Provincial Department of Science and Technology(Grant No.2022A0505030019).
文摘Challenges in stratigraphic modeling arise from underground uncertainty.While borehole exploration is reliable,it remains sparse due to economic and site constraints.Electrical resistivity tomography(ERT)as a cost-effective geophysical technique can acquire high-density data;however,uncertainty and nonuniqueness inherent in ERT impede its usage for stratigraphy identification.This paper integrates ERT and onsite observations for the first time to propose a novel method for characterizing stratigraphic profiles.The method consists of two steps:(1)ERT for prior knowledge:ERT data are processed by soft clustering using the Gaussian mixture model,followed by probability smoothing to quantify its depthdependent uncertainty;and(2)Observations for calibration:a spatial sequential Bayesian updating(SSBU)algorithm is developed to update the prior knowledge based on likelihoods derived from onsite observations,namely topsoil and boreholes.The effectiveness of the proposed method is validated through its application to a real slope site in Foshan,China.Comparative analysis with advanced borehole-driven methods highlights the superiority of incorporating ERT data in stratigraphic modeling,in terms of prediction accuracy at borehole locations and sensitivity to borehole data.Informed by ERT,reduced sensitivity to boreholes provides a fundamental solution to the longstanding challenge of sparse measurements.The paper further discusses the impact of ERT uncertainty on the proposed model using time-lapse measurements,the impact of model resolution,and applicability in engineering projects.This study,as a breakthrough in stratigraphic modeling,bridges gaps in combining geophysical and geotechnical data to address measurement sparsity and paves the way for more economical geotechnical exploration.
基金funding from the Paul ScherrerInstitute,Switzerland through the NES/GFA-ABE Cross Project。
文摘To ensure agreement between theoretical calculations and experimental data,parameters to selected nuclear physics models are perturbed and fine-tuned in nuclear data evaluations.This approach assumes that the chosen set of models accurately represents the‘true’distribution of considered observables.Furthermore,the models are chosen globally,indicating their applicability across the entire energy range of interest.However,this approach overlooks uncertainties inherent in the models themselves.In this work,we propose that instead of selecting globally a winning model set and proceeding with it as if it was the‘true’model set,we,instead,take a weighted average over multiple models within a Bayesian model averaging(BMA)framework,each weighted by its posterior probability.The method involves executing a set of TALYS calculations by randomly varying multiple nuclear physics models and their parameters to yield a vector of calculated observables.Next,computed likelihood function values at each incident energy point were then combined with the prior distributions to obtain updated posterior distributions for selected cross sections and the elastic angular distributions.As the cross sections and elastic angular distributions were updated locally on a per-energy-point basis,the approach typically results in discontinuities or“kinks”in the cross section curves,and these were addressed using spline interpolation.The proposed BMA method was applied to the evaluation of proton-induced reactions on ^(58)Ni between 1 and 100 MeV.The results demonstrated a favorable comparison with experimental data as well as with the TENDL-2023 evaluation.
基金Supported by the Chinese Nursing Association,No.ZHKY202111Scientific Research Program of School of Nursing,Chongqing Medical University,No.20230307Chongqing Science and Health Joint Medical Research Program,No.2024MSXM063.
文摘BACKGROUND Portal hypertension(PHT),primarily induced by cirrhosis,manifests severe symptoms impacting patient survival.Although transjugular intrahepatic portosystemic shunt(TIPS)is a critical intervention for managing PHT,it carries risks like hepatic encephalopathy,thus affecting patient survival prognosis.To our knowledge,existing prognostic models for post-TIPS survival in patients with PHT fail to account for the interplay among and collective impact of various prognostic factors on outcomes.Consequently,the development of an innovative modeling approach is essential to address this limitation.AIM To develop and validate a Bayesian network(BN)-based survival prediction model for patients with cirrhosis-induced PHT having undergone TIPS.METHODS The clinical data of 393 patients with cirrhosis-induced PHT who underwent TIPS surgery at the Second Affiliated Hospital of Chongqing Medical University between January 2015 and May 2022 were retrospectively analyzed.Variables were selected using Cox and least absolute shrinkage and selection operator regression methods,and a BN-based model was established and evaluated to predict survival in patients having undergone TIPS surgery for PHT.RESULTS Variable selection revealed the following as key factors impacting survival:age,ascites,hypertension,indications for TIPS,postoperative portal vein pressure(post-PVP),aspartate aminotransferase,alkaline phosphatase,total bilirubin,prealbumin,the Child-Pugh grade,and the model for end-stage liver disease(MELD)score.Based on the above-mentioned variables,a BN-based 2-year survival prognostic prediction model was constructed,which identified the following factors to be directly linked to the survival time:age,ascites,indications for TIPS,concurrent hypertension,post-PVP,the Child-Pugh grade,and the MELD score.The Bayesian information criterion was 3589.04,and 10-fold cross-validation indicated an average log-likelihood loss of 5.55 with a standard deviation of 0.16.The model’s accuracy,precision,recall,and F1 score were 0.90,0.92,0.97,and 0.95 respectively,with the area under the receiver operating characteristic curve being 0.72.CONCLUSION This study successfully developed a BN-based survival prediction model with good predictive capabilities.It offers valuable insights for treatment strategies and prognostic evaluations in patients having undergone TIPS surgery for PHT.
基金supported by the National Natural Science Foundation of China(52372310)the State Key Laboratory of Advanced Rail Autonomous Operation(RAO2023ZZ001)+1 种基金the Fundamental Research Funds for the Central Universities(2022JBQY001)Beijing Laboratory of Urban Rail Transit.
文摘The emerging virtual coupling technology aims to operate multiple train units in a Virtually Coupled Train Set(VCTS)at a minimal but safe distance.To guarantee collision avoidance,the safety distance should be calculated using the state-of-the-art space-time separation principle that separates the Emergency Braking(EB)trajectories of two successive units during the whole EB process.In this case,the minimal safety distance is usually numerically calculated without an analytic formulation.Thus,the constrained VCTS control problem is hard to address with space-time separation,which is still a gap in the existing literature.To solve this problem,we propose a Distributed Economic Model Predictive Control(DEMPC)approach with computation efficiency and theoretical guarantee.Specifically,to alleviate the computation burden,we transform implicit safety constraints into explicitly linear ones,such that the optimal control problem in DEMPC is a quadratic programming problem that can be solved efficiently.For theoretical analysis,sufficient conditions are derived to guarantee the recursive feasibility and stability of DEMPC,employing compatibility constraints,tube techniques and terminal ingredient tuning.Moreover,we extend our approach with globally optimal and distributed online EB configuration methods to shorten the minimal distance among VCTS.Finally,experimental results demonstrate the performance and advantages of the proposed approaches.
基金This study was supported by the National Natural Science Foundation of China(42261008,41971034)the Natural Science Foundation of Gansu Province,China(22JR5RA074).
文摘Stable water isotopes are natural tracers quantifying the contribution of moisture recycling to local precipitation,i.e.,the moisture recycling ratio,but various isotope-based models usually lead to different results,which affects the accuracy of local moisture recycling.In this study,a total of 18 stations from four typical areas in China were selected to compare the performance of isotope-based linear and Bayesian mixing models and to determine local moisture recycling ratio.Among the three vapor sources including advection,transpiration,and surface evaporation,the advection vapor usually played a dominant role,and the contribution of surface evaporation was less than that of transpiration.When the abnormal values were ignored,the arithmetic averages of differences between isotope-based linear and the Bayesian mixing models were 0.9%for transpiration,0.2%for surface evaporation,and–1.1%for advection,respectively,and the medians were 0.5%,0.2%,and–0.8%,respectively.The importance of transpiration was slightly less for most cases when the Bayesian mixing model was applied,and the contribution of advection was relatively larger.The Bayesian mixing model was found to perform better in determining an efficient solution since linear model sometimes resulted in negative contribution ratios.Sensitivity test with two isotope scenarios indicated that the Bayesian model had a relatively low sensitivity to the changes in isotope input,and it was important to accurately estimate the isotopes in precipitation vapor.Generally,the Bayesian mixing model should be recommended instead of a linear model.The findings are useful for understanding the performance of isotope-based linear and Bayesian mixing models under various climate backgrounds.
基金Equinor for financing the R&D projectthe Institute of Science and Technology of Petroleum Geophysics of Brazil for supporting this research。
文摘We apply stochastic seismic inversion and Bayesian facies classification for porosity modeling and igneous rock identification in the presalt interval of the Santos Basin. This integration of seismic and well-derived information enhances reservoir characterization. Stochastic inversion and Bayesian classification are powerful tools because they permit addressing the uncertainties in the model. We used the ES-MDA algorithm to achieve the realizations equivalent to the percentiles P10, P50, and P90 of acoustic impedance, a novel method for acoustic inversion in presalt. The facies were divided into five: reservoir 1,reservoir 2, tight carbonates, clayey rocks, and igneous rocks. To deal with the overlaps in acoustic impedance values of facies, we included geological information using a priori probability, indicating that structural highs are reservoir-dominated. To illustrate our approach, we conducted porosity modeling using facies-related rock-physics models for rock-physics inversion in an area with a well drilled in a coquina bank and evaluated the thickness and extension of an igneous intrusion near the carbonate-salt interface. The modeled porosity and the classified seismic facies are in good agreement with the ones observed in the wells. Notably, the coquinas bank presents an improvement in the porosity towards the top. The a priori probability model was crucial for limiting the clayey rocks to the structural lows. In Well B, the hit rate of the igneous rock in the three scenarios is higher than 60%, showing an excellent thickness-prediction capability.
基金supported by The Technology Innovation Team(Tianshan Innovation Team),Innovative Team for Efficient Utilization of Water Resources in Arid Regions(2022TSYCTD0001)the National Natural Science Foundation of China(42171269)the Xinjiang Academician Workstation Cooperative Research Project(2020.B-001).
文摘Xinjiang Uygur Autonomous Region is a typical inland arid area in China with a sparse and uneven distribution of meteorological stations,limited access to precipitation data,and significant water scarcity.Evaluating and integrating precipitation datasets from different sources to accurately characterize precipitation patterns has become a challenge to provide more accurate and alternative precipitation information for the region,which can even improve the performance of hydrological modelling.This study evaluated the applicability of widely used five satellite-based precipitation products(Climate Hazards Group InfraRed Precipitation with Station(CHIRPS),China Meteorological Forcing Dataset(CMFD),Climate Prediction Center morphing method(CMORPH),Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks-Climate Data Record(PERSIANN-CDR),and Tropical Rainfall Measuring Mission Multi-satellite Precipitation Analysis(TMPA))and a reanalysis precipitation dataset(ECMWF Reanalysis v5-Land Dataset(ERA5-Land))in Xinjiang using ground-based observational precipitation data from a limited number of meteorological stations.Based on this assessment,we proposed a framework that integrated different precipitation datasets with varying spatial resolutions using a dynamic Bayesian model averaging(DBMA)approach,the expectation-maximization method,and the ordinary Kriging interpolation method.The daily precipitation data merged using the DBMA approach exhibited distinct spatiotemporal variability,with an outstanding performance,as indicated by low root mean square error(RMSE=1.40 mm/d)and high Person's correlation coefficient(CC=0.67).Compared with the traditional simple model averaging(SMA)and individual product data,although the DBMA-fused precipitation data were slightly lower than the best precipitation product(CMFD),the overall performance of DBMA was more robust.The error analysis between DBMA-fused precipitation dataset and the more advanced Integrated Multi-satellite Retrievals for Global Precipitation Measurement Final(IMERG-F)precipitation product,as well as hydrological simulations in the Ebinur Lake Basin,further demonstrated the superior performance of DBMA-fused precipitation dataset in the entire Xinjiang region.The proposed framework for solving the fusion problem of multi-source precipitation data with different spatial resolutions is feasible for application in inland arid areas,and aids in obtaining more accurate regional hydrological information and improving regional water resources management capabilities and meteorological research in these regions.
文摘Statistical biases may be introduced by imprecisely quantifying background radiation reference levels. It is, therefore, imperative to devise a simple, adaptable approach for precisely describing the reference background levels of naturally occurring radionuclides (NOR) in mining sites. As a substitute statistical method, we suggest using Bayesian modeling in this work to examine the spatial distribution of NOR. For naturally occurring gamma-induced radionuclides like 232Th, 40K, and 238U, statistical parameters are inferred using the Markov Chain Monte Carlo (MCMC) method. After obtaining an accurate subsample using bootstrapping, we exclude any possible outliers that fall outside of the Highest Density Interval (HDI). We use MCMC to build a Bayesian model with the resampled data and make predictions about the posterior distribution of radionuclides produced by gamma irradiation. This method offers a strong and dependable way to describe NOR reference background values, which is important for managing and evaluating radiation risks in mining contexts.
基金supported by the Advance Production of Vaccine Raw Materials(Grant Nos.:20022404 and 20018168)funded by the Ministry of Trade,Industry&Energy(MOTIE,Korea)supported by the National Research Foundation of Korea(NRF)grant funded by the Korean government(MSIT)(Grant No.:NRF-2018R1A5A2023127)Dongguk University Research Fund of 2023(Grant No.:S-2023-G0001-00099)。
文摘To enhance the efficiency of vaccine manufacturing,this study focuses on optimizing the microfluidic conditions and lipid mix ratios of messenger RNA-lipid nanoparticles(mRNA-LNP).Different mRNA-LNP formulations(n=24)were developed using an I-optimal design,where machine learning tools(XGBoost/Bayesian optimization and self-validated ensemble(SVEM))were used to optimize the process and predict lipid mix ratio.The investigation included material attributes,their respective ratios,and process attributes.The critical responses like particle size(PS),polydispersity index(PDI),Zeta potential,pKa,heat trend cycle,encapsulation efficiency(EE),recovery ratio,and encapsulated mRNA were evaluated.Overall prediction of SVEM(>97%)was comparably better than that of XGBoost/Bayesian optimization(>94%).Moreover,in actual experimental outcomes,SVEM prediction is close to the actual data as confirmed by the experimental PS(94-96 nm)is close to the predicted one(95-97 nm).The other parameters including PDI and EE were also close to the actual experimental data.
基金supported by the National Natural Science Foundation of China(Grant Nos.61631008,61471137,50509059,and No.51779061)the Fok Ying-Tong Education Foundation,China(Grant No.151007)the Heilongjiang Province Outstanding Youth Science Fund(JC2017017)
文摘In recent years,moving target detection methods based on low-rank and sparse matrix decomposition have been developed,and they have achieved good results.However,there is not enough interpretation to support the assumption that there is a high correlation among the reverberations after each transmitting pulse.In order to explain the correlation of reverberations,a new reverberation model is proposed from the perspective of scattering cells in this paper.The scattering cells are the subarea divided from the detection area.The energy fluctuation of a scattering cell with time and the influence of the neighboring cells are considered.Key parameters of the model were analyzed by numerical analysis,and the applicability of the model was verified by experimental analysis.The results showed that the model can be used for several simulations to evaluate the performance of moving target detection methods.
文摘The Bayesian structural equation model integrates the principles of Bayesian statistics, providing a more flexible and comprehensive modeling framework. In exploring complex relationships between variables, handling uncertainty, and dealing with missing data, the Bayesian structural equation model demonstrates unique advantages. Therefore, Bayesian methods are used in this paper to establish a structural equation model of innovative talent cognition, with the measurement of college students’ cognition of innovative talent being studied. An in-depth analysis is conducted on the effects of innovative self-efficacy, social resources, innovative personality traits, and school education, aiming to explore the factors influencing college students’ innovative talent. The results indicate that innovative self-efficacy plays a key role in perception, social resources are significantly positively correlated with the perception of innovative talents, innovative personality tendencies and school education are positively correlated with the perception of innovative talents, but the impact is not significant.
基金supported by the National Natural Science Foun-dation of China(NSFC)Basic Science Center Program for“Multiscale Problems in Nonlinear Mechanics”(Grant No.11988102).
文摘We consider the two-point,two-time(space-time)correlation of passive scalar R(r,τ)in the Kraichnan model under the assumption of homogeneity and isotropy.Using the fine-gird PDF method,we find that R(r,τ)satisfies a diffusion equation with constant diffusion coefficient determined by velocity variance and molecular diffusion.Itssolution can be expressed in terms of the two-point,one time correlation of passive scalar,i.e.,R(r,0).Moreover,the decorrelation o R(k,τ),which is the Fourier transform of R(r,τ),is determined byR(k,0)and a diffusion kernal.
文摘In this paper,the forecasting equations of a 2nd-order space-time differential remainder are deduced from the Navier-Stokes primitive equations and Eulerian operator by Taylor-series expansion.Here we introduce a cubic spline numerical model(Spline Model for short),which is with a quasi-Lagrangian time-split integration scheme of fitting cubic spline/bicubic surface to all physical variable fields in the atmospheric equations on spherical discrete latitude-longitude mesh.A new algorithm of"fitting cubic spline—time step integration—fitting cubic spline—……"is developed to determine their first-and2nd-order derivatives and their upstream points for time discrete integral to the governing equations in Spline Model.And the cubic spline function and its mathematical polarities are also discussed to understand the Spline Model’s mathematical foundation of numerical analysis.It is pointed out that the Spline Model has mathematical laws of"convergence"of the cubic spline functions contracting to the original functions as well as its 1st-order and 2nd-order derivatives.The"optimality"of the 2nd-order derivative of the cubic spline functions is optimal approximation to that of the original functions.In addition,a Hermite bicubic patch is equivalent to operate on a grid for a 2nd-order derivative variable field.Besides,the slopes and curvatures of a central difference are identified respectively,with a smoothing coefficient of 1/3,three-point smoothing of that of a cubic spline.Then the slopes and curvatures of a central difference are calculated from the smoothing coefficient 1/3 and three-point smoothing of that of a cubic spline,respectively.Furthermore,a global simulation case of adiabatic,non-frictional and"incompressible"model atmosphere is shown with the quasi-Lagrangian time integration by using a global Spline Model,whose initial condition comes from the NCEP reanalysis data,along with quasi-uniform latitude-longitude grids and the so-called"shallow atmosphere"Navier-Stokes primitive equations in the spherical coordinates.The Spline Model,which adopted the Navier-Stokes primitive equations and quasi-Lagrangian time-split integration scheme,provides an initial ideal case of global atmospheric circulation.In addition,considering the essentially non-linear atmospheric motions,the Spline Model could judge reasonably well simple points of any smoothed variable field according to its fitting spline curvatures that must conform to its physical interpretation.
基金supported by the Guangdong Provincial Clinical Research Center for Tuberculosis(No.2020B1111170014)。
文摘Objective To investigate the spatiotemporal patterns and socioeconomic factors influencing the incidence of tuberculosis(TB)in the Guangdong Province between 2010 and 2019.Method Spatial and temporal variations in TB incidence were mapped using heat maps and hierarchical clustering.Socioenvironmental influencing factors were evaluated using a Bayesian spatiotemporal conditional autoregressive(ST-CAR)model.Results Annual incidence of TB in Guangdong decreased from 91.85/100,000 in 2010 to 53.06/100,000in 2019.Spatial hotspots were found in northeastern Guangdong,particularly in Heyuan,Shanwei,and Shantou,while Shenzhen,Dongguan,and Foshan had the lowest rates in the Pearl River Delta.The STCAR model showed that the TB risk was lower with higher per capita Gross Domestic Product(GDP)[Relative Risk(RR),0.91;95%Confidence Interval(CI):0.86–0.98],more the ratio of licensed physicians and physician(RR,0.94;95%CI:0.90-0.98),and higher per capita public expenditure(RR,0.94;95%CI:0.90–0.97),with a marginal effect of population density(RR,0.86;95%CI:0.86–1.00).Conclusion The incidence of TB in Guangdong varies spatially and temporally.Areas with poor economic conditions and insufficient healthcare resources are at an increased risk of TB infection.Strategies focusing on equitable health resource distribution and economic development are the key to TB control.
基金supported by the Ministry of Environment(Environmental Protection Administration),Taiwan(Projects EPA-106-L103-02-A022,EPA-106-L102-02-A142)the"National"Science and Technology Council(Ministry of Science and Technology),Taiwan(Nos.108-2625-M-008-002,108-2119-M-008-003,108-2636-E-008-004,109-2636-E-008-008,110-2636-E-008-006,111-2636-E-008-014,and 112-2636-E-008-005(Young Scholar Fellowship Program),112-2119-M-008-010,and 108-2638-E-008-001-MY2(Shackleton Program Grant)).
文摘Traffic emissions have become the major air pollution source in urban areas.Therefore,understanding the highly non-stational and complex impact of traffic factors on air quality is very important for building air quality prediction models.Using real-world air pollutant data from Taipei City,this study integrates diverse factors,including traffic flow,speed,rainfall patterns,andmeteorological factors.We constructed a Bayesian network probabilitymodel based on rainfall events as a big data analysis framework to investigate understand traffic factor causality relationships and condition probabilities for meteorological factors and air pollutant concentrations.Generalized Additive Model(GAM)verified non-linear relationships between traffic factors and air pollutants.Consequently,we propose a long short term memory(LSTM)model to predict airborne pollutant concentrations.This study propose a new approach of air pollutants and meteorological variable analysis procedure by considering both rainfall amount and patterns.Results indicate improved air quality when controlling vehicle speed above 40 km/h and maintaining an average vehicle flow<1200 vehicles per hour.This study also classified rainfall events into four types depending on its characteristic.Wet deposition from varied rainfall types significantly affects air quality,with TypeⅠrainfall events(long-duration heavy rain)having the most pronounced impact.An LSTM model incorporating GAM and Bayesian network outcomes yields excellent performance,achieving correlation R^(2)>0.9 and 0.8 for first and second order air pollutants,i.e.,CO,NO,NO_(2),and NO_(x);and O_(3),PM_(10),and PM_(2.5),respectively.