Aerodynamic performances of axial compressors are significantly affected by variation of Reynolds number in aero-engines.In the design and analysis of compressors,previous correction methods for cascades and stages ha...Aerodynamic performances of axial compressors are significantly affected by variation of Reynolds number in aero-engines.In the design and analysis of compressors,previous correction methods for cascades and stages have difficulties in predicting comprehensively Reynolds number effects on airfoils,matching and characteristics curves.This study proposes Re-correction models for loss,deviation angle and endwall blockage based on classical theories and cascade tests,and loss and deviation models show good agreement in test data of NACA65 and C4 cascades.Throughflow method considering Reynolds number effects is developed by integrating the correction models into a verified Streamline Curvature(SLC)tool.A three-stage axial compressor is investigated through SLC and CFD methods from design Reynolds number(Red=2106)to low Re=4104,and the numerical methods are validated with test data of characteristic curves and spanwise distributions at Red.With Re reduction,SLC method with correction models well predicts variation in overall performances compared with CFD calculations and Wassell's model.Streamwise and spanwise matching such as total pressure and loss distributions in SLC predictions are basically consistent with those in CFD results at near-stall points under design and low Reynolds numbers.SLC and CFD methods share similar detections of stall risks in the third stage(Stg3),and their analyses of diffusion processes deviate to some extent due to different predictions in separated endwall flow.The correction models can be adopted to consider Reynolds number effects in through-flow design and analysis of axial compressors.展开更多
The rapid melting of Arctic sea ice poses significant risks to the safety of shipping routes.Accurate remote sensing data on sea ice concentration(SIC)is crucial for effective route planning of ships and ensuring navi...The rapid melting of Arctic sea ice poses significant risks to the safety of shipping routes.Accurate remote sensing data on sea ice concentration(SIC)is crucial for effective route planning of ships and ensuring navigational safety.Despite the availability of numerous SIC products in China,these datasets still lag behind mainstream international products in terms of data accuracy,spatiotemporal resolution,and time span.To enhance the accuracy of China's domestic SIC remote sensing data,this study used the SIC data derived from the passive microwave remote sensing dataset provided by the University of Bremen(BRM-SIC)as a reference to conduct a comprehensive evaluation and analysis of two additional SIC datasets:the dataset derived from the microwave radiation imager(MWRI)aboard the FY-3D satellite,provided by the National Satellite Meteorological Center(FY-SIC),and the dataset obtained through the DT-ASI algorithm from the microwave imager of the FY-3D satellite,provided by Ocean University of China(OUC-SIC).Based on the evaluation results,a TransUnet fusion correction model was developed.The performance of this model was then compared against Ordinary Least Squares(OLS),Random Forest(RF),and UNet correction models,through spatial and temporal analyses.Results indicate that,compared to FY-SIC data,the RMSE of the OUC-SIC data and the standard data is reduced by24.245%,while the R is increased by 12.516%.Overall,the accuracy of OUC-SIC data is superior to that of FY-SIC data.During the research period(2020–2022),the standard deviation(SD)and coefficient of variation(CV)of OUC-SIC were 3.877%and 10.582%,respectively,while those for FY-SIC were 7.836%and 7.982%,respectively.In the study area,compared with OUC-SIC data,FYSIC data exhibited a larger standard deviation of deviation and a smaller coefficient of variation of deviation across most sea areas.These results indicate that the OUC-SIC data exhibit better temporal and spatial stability,whereas the FY-SIC data show stronger relative dimensionless stability.Among the four correction models,all showed improvements over the original,unfused corrected data.The fusion corrections using the OLS,RF,UNet,and TransUnet models reduced RMSE by 5.563%,14.601%,42.927%,and48.316%,respectively.Correspondingly,R increased by 0.463%,1.176%,3.951%,and 4.342%,respectively.Among these models,TransUnet performed the best,effectively integrating the advantages of FY-SIC and OUC-SIC data and notably improving the overall accuracy and spatiotemporal stability of SIC data.展开更多
Marine forecasting is critical for navigation safety and disaster prevention.However,traditional ocean numerical forecasting models are often limited by substantial errors and inadequate capture of temporal-spatial fe...Marine forecasting is critical for navigation safety and disaster prevention.However,traditional ocean numerical forecasting models are often limited by substantial errors and inadequate capture of temporal-spatial features.To address the limitations,the paper proposes a TimeXer-based numerical forecast correction model optimized by an exogenous-variable attention mechanism.The model treats target forecast values as internal variables,and incorporates historical temporal-spatial data and seven-day numerical forecast results from traditional models as external variables based on the embedding strategy of TimeXer.Using a self-attention structure,the model captures correlations between exogenous variables and target sequences,explores intrinsic multi-dimensional relationships,and subsequently corrects endogenous variables with the mined exogenous features.The model’s performance is evaluated using metrics including MSE(Mean Squared Error),MAE(Mean Absolute Error),RMSE(Root Mean Square Error),MAPE(Mean Absolute Percentage Error),MSPE(Mean Square Percentage Error),and computational time,with TimeXer and PatchTST models serving as benchmarks.Experiment results show that the proposed model achieves lower errors and higher correction accuracy for both one-day and seven-day forecasts.展开更多
The accuracy of genomic annotation is crucial for subsequent functional investigations;however,computational protocols used in high-throughput annotation of open reading frames(ORFs)can introduce inconsistencies.These...The accuracy of genomic annotation is crucial for subsequent functional investigations;however,computational protocols used in high-throughput annotation of open reading frames(ORFs)can introduce inconsistencies.These inconsistencies,which lead to non-uniform extension or truncation of sequence ends,pose challenges for downstream analyses.Existing strategies to rectify these inconsistencies are time-consuming and labor-intensive,lacking specific approaches.To address this gap,we developed to GC,a tool that integrates genomic annotation with RNA-seq datasets to rectify annotation inconsistencies.Using to GC,we achieved an accuracy of nearly 100%accuracy in correcting inconsistencies in published Phytophthora sojae ORFs.We applied this innovative pipeline to the GPCR-bigrams gene family,which was predicted to have 42 members in the P.sojae genome but lacked experimental validation.By employing to GC,we identified 32 GPCR-bigram ORFs with inconsistencies between previous annotations and to GC-corrected sequences.Notably,among these were 5 genes(GPCR-TKL9,GPCR-TKL15,GPCR-PDE3,GPCR-AC3,and GPCR-AC4)showed substantial inconsistencies.Experimental gene annotation confirmed the effectiveness of to GC,as sequences obtained through cloning matched those annotated by to GC.Importantly,we discovered two novel GPCRs(GPCR-AC3 and GPCR-AC4),which were previously mispredicted as a single gene.CRISPR/Cas9-mediated knockout experiments revealed the involvement of GPCR-AC4 but not GPCR-AC3 in oospore production,further confirming their status as two separate genes.In addition to P.sojae,the reliability of the to GC pipeline in Phytophthora capsici and Pythium ultimum further emphasizes the robustness of this pipeline.Our findings highlight the utility of to GC for reliable gene model correction,facilitating investigations into biological functions and offering potential applications in diverse species analyses.展开更多
The hot deformation behavior of as-extruded Ti-6554 alloy was investigated through isothermal compression at 700–950°C and 0.001–1 s^(−1).The temperature rise under different deformation conditions was calculat...The hot deformation behavior of as-extruded Ti-6554 alloy was investigated through isothermal compression at 700–950°C and 0.001–1 s^(−1).The temperature rise under different deformation conditions was calculated,and the curve was corrected.The strain compensation constitutive model of as-extruded Ti-6554 alloy based on temperature rise correction was established.The microstructure evolution under different conditions was analyzed,and the dynamic recrystallization(DRX)mechanism was revealed.The results show that the flow stress decreases with the increase in strain rate and the decrease in deformation temperature.The deformation temperature rise gradually increases with the increase in strain rate and the decrease in deformation temperature.At 700°C/1 s^(−1),the temperature rise reaches 100°C.The corrected curve value is higher than the measured value,and the strain compensation constitutive model has high prediction accuracy.The precipitation of theαphase occurs during deformation in the twophase region,which promotes DRX process of theβphase.At low strain rate,the volume fraction of dynamic recrystallization increases with the increase in deformation temperature.DRX mechanism includes continuous DRX and discontinuous DRX.展开更多
Systematic bias is a type of model error that can affect the accuracy of data assimilation and forecasting that must be addressed.An online bias correction scheme called the sequential bias correction scheme(SBCS),was...Systematic bias is a type of model error that can affect the accuracy of data assimilation and forecasting that must be addressed.An online bias correction scheme called the sequential bias correction scheme(SBCS),was developed using the6 h average bias to correct the systematic bias during model integration.The primary purpose of this study is to investigate the impact of the SBCS in the high-resolution China Meteorological Administration Meso-scale(CMA-MESO)numerical weather prediction(NWP)model to reduce the systematic bias and to improve the data assimilation and forecast results through this method.The SBCS is improved upon and applied to the CMA-MESO 3-km model in this study.Four-week sequential data assimilation and forecast experiments,driven by rapid update and cycling(RUC),were conducted for the period from 2–29 May 2022.In terms of the characteristics of systematic bias,both the background and analysis show diurnal bias,and these large biases are affected by complex underlying surfaces(e.g.,oceans,coasts,and mountains).After the application of the SBCS,the results of the data assimilation show that the SBCS can reduce the systematic bias of the background and yield a neutral to slightly positive result for the analysis fields.In addition,the SBCS can reduce forecast errors and improve forecast results,especially for surface variables.The above results indicate that this scheme has good prospects for high-resolution regional NWP models.展开更多
The correction of Light Detection and Ranging(LiDAR)intensity data is of great significance for enhancing its application value.However,traditional intensity correction methods based on Terrestrial Laser Scanning(TLS)...The correction of Light Detection and Ranging(LiDAR)intensity data is of great significance for enhancing its application value.However,traditional intensity correction methods based on Terrestrial Laser Scanning(TLS)technology rely on manual site setup to collect intensity training data at different distances and incidence angles,which is noisy and limited in sample quantity,restricting the improvement of model accuracy.To overcome this limitation,this study proposes a fine-grained intensity correction modeling method based on Mobile Laser Scanning(MLS)technology.The method utilizes the continuous scanning characteristics of MLS technology to obtain dense point cloud intensity data at various distances and incidence angles.Then,a fine-grained screening strategy is employed to accurately select distance-intensity and incidence angle-intensity modeling samples.Finally,based on these samples,a high-precision intensity correction model is established through polynomial fitting functions.To verify the effectiveness of the proposed method,comparative experiments were designed,and the MLS modeling method was validated against the traditional TLS modeling method on the same test set.The results show that on Test Set 1,where the distance values vary widely(i.e.,0.1–3 m),the intensity consistency after correction using the MLS modeling method reached 7.692 times the original intensity,while the traditional TLS modeling method only increased to 4.630 times the original intensity.On Test Set 2,where the incidence angle values vary widely(i.e.,0○–80○),the MLS modeling method,although with a relatively smaller advantage,still improved the intensity consistency to 3.937 times the original intensity,slightly better than the TLS modeling method’s 3.413 times.These results demonstrate the significant advantage of the modeling method proposed in this study in enhancing the accuracy of intensity correction models.展开更多
Following the publication of Zeng et al.(2023),an inadvertent error was recently identified in Figure 1B and Supplementary Figure S3.To ensure the accuracy and integrity of our published work,we formally request a cor...Following the publication of Zeng et al.(2023),an inadvertent error was recently identified in Figure 1B and Supplementary Figure S3.To ensure the accuracy and integrity of our published work,we formally request a correction to address this issue and apologize for any confusion this error may have caused.For details,please refer to the modified Supplementary Materials.展开更多
Monte Carlo(MC) simulations have been performed to refine the estimation of the correction-toscaling exponent ω in the 2D φ^(4)model,which belongs to one of the most fundamental universality classes.If corrections h...Monte Carlo(MC) simulations have been performed to refine the estimation of the correction-toscaling exponent ω in the 2D φ^(4)model,which belongs to one of the most fundamental universality classes.If corrections have the form ∝ L^(-ω),then we find ω=1.546(30) andω=1.509(14) as the best estimates.These are obtained from the finite-size scaling of the susceptibility data in the range of linear lattice sizes L ∈[128,2048] at the critical value of the Binder cumulant and from the scaling of the corresponding pseudocritical couplings within L∈[64,2048].These values agree with several other MC estimates at the assumption of the power-law corrections and are comparable with the known results of the ε-expansion.In addition,we have tested the consistency with the scaling corrections of the form ∝ L^(-4/3),∝L^(-4/3)In L and ∝L^(-4/3)/ln L,which might be expected from some considerations of the renormalization group and Coulomb gas model.The latter option is consistent with our MC data.Our MC results served as a basis for a critical reconsideration of some earlier theoretical conjectures and scaling assumptions.In particular,we have corrected and refined our previous analysis by grouping Feynman diagrams.The renewed analysis gives ω≈4-d-2η as some approximation for spatial dimensions d <4,or ω≈1.5 in two dimensions.展开更多
The performance of corporate social responsibility is conducive to the con- tinuous improvement of their profitability, and promotes the upgrading of corporation value. However, it is difficult to confirm, calculate a...The performance of corporate social responsibility is conducive to the con- tinuous improvement of their profitability, and promotes the upgrading of corporation value. However, it is difficult to confirm, calculate and check the costs and benefits brought by the implementation of corporate social responsibility under the current ac- counting theory system, so it is difficult to estimate whether the fulfillment of corpo- rate social responsibility has any effects on the corporation value assessment. Therefore, based on corporate social responsibility, the correction mode of corpora- tion value assessment is put forward.展开更多
In view of the learners Chinglish,this paper puts forward a new teaching model-the Self-correction Model and makes an analysis of it from the view of cognitive psychology.
Based on the theory of the post method pedagogy and the teaching practice,this paper proposes a new teaching model,the self-correction model,and explains it from the point of view of linking theory.
The Spalart-Allmaras (S-A) turbulence model, the shear-stress transport (SST) turbulence model and their compressibility corrections are revaluated for hypersonic compression comer flows by using high-order differ...The Spalart-Allmaras (S-A) turbulence model, the shear-stress transport (SST) turbulence model and their compressibility corrections are revaluated for hypersonic compression comer flows by using high-order difference schemes. The compressibility effect of density gradient, pressure dilatation and turbulent Mach number is accounted. In order to reduce confusions between model uncertainties and discretization errors, the formally fifth-order explicit weighted compact nonlinear scheme (WCNS-E-5) is adopted for convection terms, and a fourth-order staggered central difference scheme is applied for viscous terms. The 15° and 34° compression comers at Mach number 9.22 are investigated. Numerical results show that the original SST model is superior to the original S-A model in the resolution of separated regions and predictions of wall pressures and wall heat-flux rates. The capability of the S-A model can be largely improved by blending Catris' and Shur's compressibility corrections. Among the three corrections of the SST model listed in the present paper, Catris' modification brings the best results. However, the dissipation and pressure dilatation corrections result in much larger separated regions than that of the experiment, and are much worse than the original SST model as well as the other two corrections. The correction of turbulent Mach number makes the separated region slightly smaller than that of the original SST model. Some results of low-order schemes are also presented. When compared to the results of the high-order schemes, the separated regions are smaller, and the peak wall pressures and peak heat-flux rates are lower in the region of the reattachment points.展开更多
Errors inevitably exist in numerical weather prediction (NWP) due to imperfect numeric and physical parameterizations. To eliminate these errors, by considering NWP as an inverse problem, an unknown term in the pred...Errors inevitably exist in numerical weather prediction (NWP) due to imperfect numeric and physical parameterizations. To eliminate these errors, by considering NWP as an inverse problem, an unknown term in the prediction equations can be estimated inversely by using the past data, which are presumed to represent the imperfection of the NWP model (model error, denoted as ME). In this first paper of a two-part series, an iteration method for obtaining the MEs in past intervals is presented, and the results from testing its convergence in idealized experiments are reported. Moreover, two batches of iteration tests were applied in the global forecast system of the Global and Regional Assimilation and Prediction System (GRAPES-GFS) for July-August 2009 and January-February 2010. The datasets associated with the initial conditions and sea surface temperature (SST) were both based on NCEP (National Centers for Environmental Prediction) FNL (final) data. The results showed that 6th h forecast errors were reduced to 10% of their original value after a 20-step iteration. Then, off-line forecast error corrections were estimated linearly based on the 2-month mean MEs and compared with forecast errors. The estimated error corrections agreed well with the forecast errors, but the linear growth rate of the estimation was steeper than the forecast error. The advantage of this iteration method is that the MEs can provide the foundation for online correction. A larger proportion of the forecast errors can be expected to be canceled out by properly introducing the model error correction into GRAPES-GFS.展开更多
This study presents a simplified multivariate bias correction scheme that is sequentially implemented in the GEOS5 data assimilation system and compared against a control experiment without model bias correction. The ...This study presents a simplified multivariate bias correction scheme that is sequentially implemented in the GEOS5 data assimilation system and compared against a control experiment without model bias correction. The results show considerable improvement in terms of the mean biases of rawinsonde observation-minus-background (OmB) residuals for observed water vapor, wind and temperature variables. The time series spectral analysis shows whitening of bias-corrected OmB residuals, and mean biases for rawinsonde observation-minus-analysis (OmA) are also improved. Some wind and temperature biases in the control experiment near the equatorial tropopause nearly vanish from the bias-corrected experiment. Despite the analysis improvement, the bias correction scheme has only a moderate impact on forecast skill. Significant interaction is also found among quality-control, satellite observation bias correction, and background bias correction, and the latter positively impacts satellite bias correction.展开更多
Based on the theory of first-order reaction kinetics,a thermal reaction kinetic model in integral form has been derive.To make the model more applicable,the effects of time and the conversion degree on the reaction ra...Based on the theory of first-order reaction kinetics,a thermal reaction kinetic model in integral form has been derive.To make the model more applicable,the effects of time and the conversion degree on the reaction rate parameters were considered.Two types of undetermined functions were used to compensate for the intrinsic variation of the reaction rate,and two types of correction methods are provided.The model was explained and verified using published experimental data of different polymer thermal reaction systems,and its effectiveness and wide adaptability were confirmed.For the given kinetic model,only one parameter needs to be determined.The proposed empirical model is expected to be used in the numerical simulation of polymer thermal reaction process.展开更多
By using error correction model, I conduct co-integration analysis on the research of the relationship between the per capita practical consumption and per capita practical disposable income of urban residents in Huna...By using error correction model, I conduct co-integration analysis on the research of the relationship between the per capita practical consumption and per capita practical disposable income of urban residents in Hunan Province from 1978 to 2009. The results show that there is a co-integration relationship between the per capita practical consumption and the practical per capita disposable income of urban residents, and based on these, the corresponding error correction model is established. Finally, corresponding countermeasures and suggestions are put forward as follows: broaden the income channel of urban residents; create goods consuming environment; perfect socialist security system.展开更多
Gamma-aminobutyric acid(GABA)is a natural non-protein functio nal amino acid,which has potential for fermentation industrial production by Lactobacillus brevis.This work investigated the batch fermentation process and...Gamma-aminobutyric acid(GABA)is a natural non-protein functio nal amino acid,which has potential for fermentation industrial production by Lactobacillus brevis.This work investigated the batch fermentation process and developed a kinetic model based on substrate restrictive model established by experimental data from L25(5~6)orthogonal experiments.In this study,the OD600 value of fermentation broth was fixed to constant after reaching its maximum because the microorganism death showed no effect on the enzyme activity of glutamate decarboxylase(GAD).As pH is one of the key parameters in fermentation process,a pH-dependent kinetic model based on radial basis function was developed to enhance the practicality of the model.Furthermore,as to decrease the deviations between the simulated curves and the experimental data,the rolling correction strategy with OD600 values that was measured in real-time was introduced into this work to modify the model.Finally,the accu racy of the rolling corrected and pH-dependent model was validated by good fitness between the simulated curves and data of the initial batch fermentation(pH 5.2).As a result,this pH-dependent kinetic model revealed that the optimal pH for biomass growth is 5.6-5.7 and for GABA production is about 5,respectively.Therefore,the developed model is practical and convenient for the instruction of GABA fermentation production,and it has instructive significance for the industrial scale.展开更多
This letter presents a new chunking method based on Maximum Entropy (ME) model with N-fold template correction model.First two types of machine learning models are described.Based on the analysis of the two models,the...This letter presents a new chunking method based on Maximum Entropy (ME) model with N-fold template correction model.First two types of machine learning models are described.Based on the analysis of the two models,then the chunking model which combines the profits of conditional probability model and rule based model is proposed.The selection of features and rule templates in the chunking model is discussed.Experimental results for the CoNLL-2000 corpus show that this approach achieves impressive accuracy in terms of the F-score:92.93%.Compared with the ME model and ME Markov model,the new chunking model achieves better performance.展开更多
In order to understand the effect of hardening ductility parameters and softening ductility parameters of the concrete damage plastic model in LS-DYNA,a sensitivity and reliability analysis of these parameters through...In order to understand the effect of hardening ductility parameters and softening ductility parameters of the concrete damage plastic model in LS-DYNA,a sensitivity and reliability analysis of these parameters through a convenient cube unit test was conducted. The results showed that the peak strength strain was independent of the hardening ductility parameter DH,but affected by AH,BH,and CH. The softening ductility was mainly related to the softening ductility parameter AS,but not affected by the damage ductility exponent BS. In case that the model with default parameters failed to match the AS-controlled damage softening phase,an optimized model with an AS correction was developed. The corrected model with the AS value of 2 matched well with the code model,and exhibited good feasibility in predicting the stress-strain curve of different grades of concrete. Moreover,the practicability of the corrected model was further validated by the conventional triaxial test. The simulated curve exhibited favorable consistence with the trial curve. Therefore,the model with parameter correction could provide a prospective reference for predicting the mechanical properties of concrete.展开更多
基金supported by the National Science and Tech-nology Major Project of China(Nos.2017-II-0007-0021 and J2019-II-0017-0038)。
文摘Aerodynamic performances of axial compressors are significantly affected by variation of Reynolds number in aero-engines.In the design and analysis of compressors,previous correction methods for cascades and stages have difficulties in predicting comprehensively Reynolds number effects on airfoils,matching and characteristics curves.This study proposes Re-correction models for loss,deviation angle and endwall blockage based on classical theories and cascade tests,and loss and deviation models show good agreement in test data of NACA65 and C4 cascades.Throughflow method considering Reynolds number effects is developed by integrating the correction models into a verified Streamline Curvature(SLC)tool.A three-stage axial compressor is investigated through SLC and CFD methods from design Reynolds number(Red=2106)to low Re=4104,and the numerical methods are validated with test data of characteristic curves and spanwise distributions at Red.With Re reduction,SLC method with correction models well predicts variation in overall performances compared with CFD calculations and Wassell's model.Streamwise and spanwise matching such as total pressure and loss distributions in SLC predictions are basically consistent with those in CFD results at near-stall points under design and low Reynolds numbers.SLC and CFD methods share similar detections of stall risks in the third stage(Stg3),and their analyses of diffusion processes deviate to some extent due to different predictions in separated endwall flow.The correction models can be adopted to consider Reynolds number effects in through-flow design and analysis of axial compressors.
基金supported by the National Natural Science Foundation of China(No.41971339)the SDUST Research Fund(No.2019TDJH103)。
文摘The rapid melting of Arctic sea ice poses significant risks to the safety of shipping routes.Accurate remote sensing data on sea ice concentration(SIC)is crucial for effective route planning of ships and ensuring navigational safety.Despite the availability of numerous SIC products in China,these datasets still lag behind mainstream international products in terms of data accuracy,spatiotemporal resolution,and time span.To enhance the accuracy of China's domestic SIC remote sensing data,this study used the SIC data derived from the passive microwave remote sensing dataset provided by the University of Bremen(BRM-SIC)as a reference to conduct a comprehensive evaluation and analysis of two additional SIC datasets:the dataset derived from the microwave radiation imager(MWRI)aboard the FY-3D satellite,provided by the National Satellite Meteorological Center(FY-SIC),and the dataset obtained through the DT-ASI algorithm from the microwave imager of the FY-3D satellite,provided by Ocean University of China(OUC-SIC).Based on the evaluation results,a TransUnet fusion correction model was developed.The performance of this model was then compared against Ordinary Least Squares(OLS),Random Forest(RF),and UNet correction models,through spatial and temporal analyses.Results indicate that,compared to FY-SIC data,the RMSE of the OUC-SIC data and the standard data is reduced by24.245%,while the R is increased by 12.516%.Overall,the accuracy of OUC-SIC data is superior to that of FY-SIC data.During the research period(2020–2022),the standard deviation(SD)and coefficient of variation(CV)of OUC-SIC were 3.877%and 10.582%,respectively,while those for FY-SIC were 7.836%and 7.982%,respectively.In the study area,compared with OUC-SIC data,FYSIC data exhibited a larger standard deviation of deviation and a smaller coefficient of variation of deviation across most sea areas.These results indicate that the OUC-SIC data exhibit better temporal and spatial stability,whereas the FY-SIC data show stronger relative dimensionless stability.Among the four correction models,all showed improvements over the original,unfused corrected data.The fusion corrections using the OLS,RF,UNet,and TransUnet models reduced RMSE by 5.563%,14.601%,42.927%,and48.316%,respectively.Correspondingly,R increased by 0.463%,1.176%,3.951%,and 4.342%,respectively.Among these models,TransUnet performed the best,effectively integrating the advantages of FY-SIC and OUC-SIC data and notably improving the overall accuracy and spatiotemporal stability of SIC data.
基金supported by the National Key Research and Development Program Project(2023YFC3107804)Planning Fund Project of Humanities and Social Sciences Research of the Ministry of Education(24YJA880097)the Graduate Education Reform Project in North China University of Technology(217051360025XN095-17)。
文摘Marine forecasting is critical for navigation safety and disaster prevention.However,traditional ocean numerical forecasting models are often limited by substantial errors and inadequate capture of temporal-spatial features.To address the limitations,the paper proposes a TimeXer-based numerical forecast correction model optimized by an exogenous-variable attention mechanism.The model treats target forecast values as internal variables,and incorporates historical temporal-spatial data and seven-day numerical forecast results from traditional models as external variables based on the embedding strategy of TimeXer.Using a self-attention structure,the model captures correlations between exogenous variables and target sequences,explores intrinsic multi-dimensional relationships,and subsequently corrects endogenous variables with the mined exogenous features.The model’s performance is evaluated using metrics including MSE(Mean Squared Error),MAE(Mean Absolute Error),RMSE(Root Mean Square Error),MAPE(Mean Absolute Percentage Error),MSPE(Mean Square Percentage Error),and computational time,with TimeXer and PatchTST models serving as benchmarks.Experiment results show that the proposed model achieves lower errors and higher correction accuracy for both one-day and seven-day forecasts.
基金supported by the grants to Min Qiu and Ming Wang from the National Natural Science Foundation of China(32100160 and 32100044)the grants to Ming Wang from the Jiangsu“Innovative and Entrepreneurial Talent”Program,China(JSSCRC2021510)the grants to Yuanchao Wang from the Chinese Modern Agricultural Industry Technology System(CARS-004-PS14)。
文摘The accuracy of genomic annotation is crucial for subsequent functional investigations;however,computational protocols used in high-throughput annotation of open reading frames(ORFs)can introduce inconsistencies.These inconsistencies,which lead to non-uniform extension or truncation of sequence ends,pose challenges for downstream analyses.Existing strategies to rectify these inconsistencies are time-consuming and labor-intensive,lacking specific approaches.To address this gap,we developed to GC,a tool that integrates genomic annotation with RNA-seq datasets to rectify annotation inconsistencies.Using to GC,we achieved an accuracy of nearly 100%accuracy in correcting inconsistencies in published Phytophthora sojae ORFs.We applied this innovative pipeline to the GPCR-bigrams gene family,which was predicted to have 42 members in the P.sojae genome but lacked experimental validation.By employing to GC,we identified 32 GPCR-bigram ORFs with inconsistencies between previous annotations and to GC-corrected sequences.Notably,among these were 5 genes(GPCR-TKL9,GPCR-TKL15,GPCR-PDE3,GPCR-AC3,and GPCR-AC4)showed substantial inconsistencies.Experimental gene annotation confirmed the effectiveness of to GC,as sequences obtained through cloning matched those annotated by to GC.Importantly,we discovered two novel GPCRs(GPCR-AC3 and GPCR-AC4),which were previously mispredicted as a single gene.CRISPR/Cas9-mediated knockout experiments revealed the involvement of GPCR-AC4 but not GPCR-AC3 in oospore production,further confirming their status as two separate genes.In addition to P.sojae,the reliability of the to GC pipeline in Phytophthora capsici and Pythium ultimum further emphasizes the robustness of this pipeline.Our findings highlight the utility of to GC for reliable gene model correction,facilitating investigations into biological functions and offering potential applications in diverse species analyses.
基金National Key R&D Program of China(2022YFB3706901)National Natural Science Foundation of China(52274382)Key Research and Development Program of Hubei Province(2022BAA024)。
文摘The hot deformation behavior of as-extruded Ti-6554 alloy was investigated through isothermal compression at 700–950°C and 0.001–1 s^(−1).The temperature rise under different deformation conditions was calculated,and the curve was corrected.The strain compensation constitutive model of as-extruded Ti-6554 alloy based on temperature rise correction was established.The microstructure evolution under different conditions was analyzed,and the dynamic recrystallization(DRX)mechanism was revealed.The results show that the flow stress decreases with the increase in strain rate and the decrease in deformation temperature.The deformation temperature rise gradually increases with the increase in strain rate and the decrease in deformation temperature.At 700°C/1 s^(−1),the temperature rise reaches 100°C.The corrected curve value is higher than the measured value,and the strain compensation constitutive model has high prediction accuracy.The precipitation of theαphase occurs during deformation in the twophase region,which promotes DRX process of theβphase.At low strain rate,the volume fraction of dynamic recrystallization increases with the increase in deformation temperature.DRX mechanism includes continuous DRX and discontinuous DRX.
基金supported by the National Natural Science Foundation of China(Grant Nos.U2242213,U2142213,42305167,42175105)。
文摘Systematic bias is a type of model error that can affect the accuracy of data assimilation and forecasting that must be addressed.An online bias correction scheme called the sequential bias correction scheme(SBCS),was developed using the6 h average bias to correct the systematic bias during model integration.The primary purpose of this study is to investigate the impact of the SBCS in the high-resolution China Meteorological Administration Meso-scale(CMA-MESO)numerical weather prediction(NWP)model to reduce the systematic bias and to improve the data assimilation and forecast results through this method.The SBCS is improved upon and applied to the CMA-MESO 3-km model in this study.Four-week sequential data assimilation and forecast experiments,driven by rapid update and cycling(RUC),were conducted for the period from 2–29 May 2022.In terms of the characteristics of systematic bias,both the background and analysis show diurnal bias,and these large biases are affected by complex underlying surfaces(e.g.,oceans,coasts,and mountains).After the application of the SBCS,the results of the data assimilation show that the SBCS can reduce the systematic bias of the background and yield a neutral to slightly positive result for the analysis fields.In addition,the SBCS can reduce forecast errors and improve forecast results,especially for surface variables.The above results indicate that this scheme has good prospects for high-resolution regional NWP models.
基金supported in part by the National Natural Science Foundation of China under grant number 31901239funded by Researchers Supporting Project Number(RSPD2025R947),King Saud University,Riyadh,Saudi Arabia.
文摘The correction of Light Detection and Ranging(LiDAR)intensity data is of great significance for enhancing its application value.However,traditional intensity correction methods based on Terrestrial Laser Scanning(TLS)technology rely on manual site setup to collect intensity training data at different distances and incidence angles,which is noisy and limited in sample quantity,restricting the improvement of model accuracy.To overcome this limitation,this study proposes a fine-grained intensity correction modeling method based on Mobile Laser Scanning(MLS)technology.The method utilizes the continuous scanning characteristics of MLS technology to obtain dense point cloud intensity data at various distances and incidence angles.Then,a fine-grained screening strategy is employed to accurately select distance-intensity and incidence angle-intensity modeling samples.Finally,based on these samples,a high-precision intensity correction model is established through polynomial fitting functions.To verify the effectiveness of the proposed method,comparative experiments were designed,and the MLS modeling method was validated against the traditional TLS modeling method on the same test set.The results show that on Test Set 1,where the distance values vary widely(i.e.,0.1–3 m),the intensity consistency after correction using the MLS modeling method reached 7.692 times the original intensity,while the traditional TLS modeling method only increased to 4.630 times the original intensity.On Test Set 2,where the incidence angle values vary widely(i.e.,0○–80○),the MLS modeling method,although with a relatively smaller advantage,still improved the intensity consistency to 3.937 times the original intensity,slightly better than the TLS modeling method’s 3.413 times.These results demonstrate the significant advantage of the modeling method proposed in this study in enhancing the accuracy of intensity correction models.
文摘Following the publication of Zeng et al.(2023),an inadvertent error was recently identified in Figure 1B and Supplementary Figure S3.To ensure the accuracy and integrity of our published work,we formally request a correction to address this issue and apologize for any confusion this error may have caused.For details,please refer to the modified Supplementary Materials.
文摘Monte Carlo(MC) simulations have been performed to refine the estimation of the correction-toscaling exponent ω in the 2D φ^(4)model,which belongs to one of the most fundamental universality classes.If corrections have the form ∝ L^(-ω),then we find ω=1.546(30) andω=1.509(14) as the best estimates.These are obtained from the finite-size scaling of the susceptibility data in the range of linear lattice sizes L ∈[128,2048] at the critical value of the Binder cumulant and from the scaling of the corresponding pseudocritical couplings within L∈[64,2048].These values agree with several other MC estimates at the assumption of the power-law corrections and are comparable with the known results of the ε-expansion.In addition,we have tested the consistency with the scaling corrections of the form ∝ L^(-4/3),∝L^(-4/3)In L and ∝L^(-4/3)/ln L,which might be expected from some considerations of the renormalization group and Coulomb gas model.The latter option is consistent with our MC data.Our MC results served as a basis for a critical reconsideration of some earlier theoretical conjectures and scaling assumptions.In particular,we have corrected and refined our previous analysis by grouping Feynman diagrams.The renewed analysis gives ω≈4-d-2η as some approximation for spatial dimensions d <4,or ω≈1.5 in two dimensions.
文摘The performance of corporate social responsibility is conducive to the con- tinuous improvement of their profitability, and promotes the upgrading of corporation value. However, it is difficult to confirm, calculate and check the costs and benefits brought by the implementation of corporate social responsibility under the current ac- counting theory system, so it is difficult to estimate whether the fulfillment of corpo- rate social responsibility has any effects on the corporation value assessment. Therefore, based on corporate social responsibility, the correction mode of corpora- tion value assessment is put forward.
文摘In view of the learners Chinglish,this paper puts forward a new teaching model-the Self-correction Model and makes an analysis of it from the view of cognitive psychology.
文摘Based on the theory of the post method pedagogy and the teaching practice,this paper proposes a new teaching model,the self-correction model,and explains it from the point of view of linking theory.
基金Foundation items: National Basic Research Program of China (2009CB723801) National Natural Science Foundation of China (11072259)
文摘The Spalart-Allmaras (S-A) turbulence model, the shear-stress transport (SST) turbulence model and their compressibility corrections are revaluated for hypersonic compression comer flows by using high-order difference schemes. The compressibility effect of density gradient, pressure dilatation and turbulent Mach number is accounted. In order to reduce confusions between model uncertainties and discretization errors, the formally fifth-order explicit weighted compact nonlinear scheme (WCNS-E-5) is adopted for convection terms, and a fourth-order staggered central difference scheme is applied for viscous terms. The 15° and 34° compression comers at Mach number 9.22 are investigated. Numerical results show that the original SST model is superior to the original S-A model in the resolution of separated regions and predictions of wall pressures and wall heat-flux rates. The capability of the S-A model can be largely improved by blending Catris' and Shur's compressibility corrections. Among the three corrections of the SST model listed in the present paper, Catris' modification brings the best results. However, the dissipation and pressure dilatation corrections result in much larger separated regions than that of the experiment, and are much worse than the original SST model as well as the other two corrections. The correction of turbulent Mach number makes the separated region slightly smaller than that of the original SST model. Some results of low-order schemes are also presented. When compared to the results of the high-order schemes, the separated regions are smaller, and the peak wall pressures and peak heat-flux rates are lower in the region of the reattachment points.
基金funded by the National Natural Science Foundation Science Fund for Youth (Grant No.41405095)the Key Projects in the National Science and Technology Pillar Program during the Twelfth Fiveyear Plan Period (Grant No.2012BAC22B02)the National Natural Science Foundation Science Fund for Creative Research Groups (Grant No.41221064)
文摘Errors inevitably exist in numerical weather prediction (NWP) due to imperfect numeric and physical parameterizations. To eliminate these errors, by considering NWP as an inverse problem, an unknown term in the prediction equations can be estimated inversely by using the past data, which are presumed to represent the imperfection of the NWP model (model error, denoted as ME). In this first paper of a two-part series, an iteration method for obtaining the MEs in past intervals is presented, and the results from testing its convergence in idealized experiments are reported. Moreover, two batches of iteration tests were applied in the global forecast system of the Global and Regional Assimilation and Prediction System (GRAPES-GFS) for July-August 2009 and January-February 2010. The datasets associated with the initial conditions and sea surface temperature (SST) were both based on NCEP (National Centers for Environmental Prediction) FNL (final) data. The results showed that 6th h forecast errors were reduced to 10% of their original value after a 20-step iteration. Then, off-line forecast error corrections were estimated linearly based on the 2-month mean MEs and compared with forecast errors. The estimated error corrections agreed well with the forecast errors, but the linear growth rate of the estimation was steeper than the forecast error. The advantage of this iteration method is that the MEs can provide the foundation for online correction. A larger proportion of the forecast errors can be expected to be canceled out by properly introducing the model error correction into GRAPES-GFS.
文摘This study presents a simplified multivariate bias correction scheme that is sequentially implemented in the GEOS5 data assimilation system and compared against a control experiment without model bias correction. The results show considerable improvement in terms of the mean biases of rawinsonde observation-minus-background (OmB) residuals for observed water vapor, wind and temperature variables. The time series spectral analysis shows whitening of bias-corrected OmB residuals, and mean biases for rawinsonde observation-minus-analysis (OmA) are also improved. Some wind and temperature biases in the control experiment near the equatorial tropopause nearly vanish from the bias-corrected experiment. Despite the analysis improvement, the bias correction scheme has only a moderate impact on forecast skill. Significant interaction is also found among quality-control, satellite observation bias correction, and background bias correction, and the latter positively impacts satellite bias correction.
基金supported by the National Key Research and Development Program of China(Grant No.2018YFB2001002)。
文摘Based on the theory of first-order reaction kinetics,a thermal reaction kinetic model in integral form has been derive.To make the model more applicable,the effects of time and the conversion degree on the reaction rate parameters were considered.Two types of undetermined functions were used to compensate for the intrinsic variation of the reaction rate,and two types of correction methods are provided.The model was explained and verified using published experimental data of different polymer thermal reaction systems,and its effectiveness and wide adaptability were confirmed.For the given kinetic model,only one parameter needs to be determined.The proposed empirical model is expected to be used in the numerical simulation of polymer thermal reaction process.
基金Supported by the Scientific Research Subject of Department of Education in Hunan Province(10C0556)
文摘By using error correction model, I conduct co-integration analysis on the research of the relationship between the per capita practical consumption and per capita practical disposable income of urban residents in Hunan Province from 1978 to 2009. The results show that there is a co-integration relationship between the per capita practical consumption and the practical per capita disposable income of urban residents, and based on these, the corresponding error correction model is established. Finally, corresponding countermeasures and suggestions are put forward as follows: broaden the income channel of urban residents; create goods consuming environment; perfect socialist security system.
基金supported by the National Natural Science Foundation of China(21621004,22078239)the Beijing-Tianjin-Hebei Basic Research Cooperation Project(B2021210008)+1 种基金Tianjin Synthetic Biotechnology Innovation Capacity Improvement Project(TSBICIP-KJGG-004)the Tianjin Development Program for Innovation and Entrepreneurship(2018)。
文摘Gamma-aminobutyric acid(GABA)is a natural non-protein functio nal amino acid,which has potential for fermentation industrial production by Lactobacillus brevis.This work investigated the batch fermentation process and developed a kinetic model based on substrate restrictive model established by experimental data from L25(5~6)orthogonal experiments.In this study,the OD600 value of fermentation broth was fixed to constant after reaching its maximum because the microorganism death showed no effect on the enzyme activity of glutamate decarboxylase(GAD).As pH is one of the key parameters in fermentation process,a pH-dependent kinetic model based on radial basis function was developed to enhance the practicality of the model.Furthermore,as to decrease the deviations between the simulated curves and the experimental data,the rolling correction strategy with OD600 values that was measured in real-time was introduced into this work to modify the model.Finally,the accu racy of the rolling corrected and pH-dependent model was validated by good fitness between the simulated curves and data of the initial batch fermentation(pH 5.2).As a result,this pH-dependent kinetic model revealed that the optimal pH for biomass growth is 5.6-5.7 and for GABA production is about 5,respectively.Therefore,the developed model is practical and convenient for the instruction of GABA fermentation production,and it has instructive significance for the industrial scale.
基金Supported by National Natural Science Foundation of China (No.60504021).
文摘This letter presents a new chunking method based on Maximum Entropy (ME) model with N-fold template correction model.First two types of machine learning models are described.Based on the analysis of the two models,then the chunking model which combines the profits of conditional probability model and rule based model is proposed.The selection of features and rule templates in the chunking model is discussed.Experimental results for the CoNLL-2000 corpus show that this approach achieves impressive accuracy in terms of the F-score:92.93%.Compared with the ME model and ME Markov model,the new chunking model achieves better performance.
基金Supported by the National Natural Science Foundation of China(10272109)
文摘In order to understand the effect of hardening ductility parameters and softening ductility parameters of the concrete damage plastic model in LS-DYNA,a sensitivity and reliability analysis of these parameters through a convenient cube unit test was conducted. The results showed that the peak strength strain was independent of the hardening ductility parameter DH,but affected by AH,BH,and CH. The softening ductility was mainly related to the softening ductility parameter AS,but not affected by the damage ductility exponent BS. In case that the model with default parameters failed to match the AS-controlled damage softening phase,an optimized model with an AS correction was developed. The corrected model with the AS value of 2 matched well with the code model,and exhibited good feasibility in predicting the stress-strain curve of different grades of concrete. Moreover,the practicability of the corrected model was further validated by the conventional triaxial test. The simulated curve exhibited favorable consistence with the trial curve. Therefore,the model with parameter correction could provide a prospective reference for predicting the mechanical properties of concrete.