It is known that correlation does not imply causality.Some relationships identified in the analysis of data are coincidental or unknown,and some are produced by real-world causality of the situation,which is problemat...It is known that correlation does not imply causality.Some relationships identified in the analysis of data are coincidental or unknown,and some are produced by real-world causality of the situation,which is problematic,since there is a need to differentiate between these two scenarios.Until recently,the proper−semantic−causality of the relationship could have been determined only by human experts from the area of expertise of the studied data.This has changed with the advance of large language models,which are often utilized as surrogates for such human experts,making the process automated and readily available to all data analysts.This motivates the main objective of this work,which is to introduce the design and implementation of a large language model-based semantic causality evaluator based on correlation analysis,together with its visual analysis model called Causal heatmap.After the implementation itself,the model is evaluated from the point of view of the quality of the visual model,from the point of view of the quality of causal evaluation based on large language models,and from the point of view of comparative analysis,while the results reached in the study highlight the usability of large language models in the task and the potential of the proposed approach in the analysis of unknown datasets.The results of the experimental evaluation demonstrate the usefulness of the Causal heatmap method,supported by the evident highlighting of interesting relationships,while suppressing irrelevant ones.展开更多
Detecting geomagnetic anomalies preceding earthquakes is a challenging yet promising area of research that has gained increasing attention in recent years.This study introduces a novel reconstruction-based modeling ap...Detecting geomagnetic anomalies preceding earthquakes is a challenging yet promising area of research that has gained increasing attention in recent years.This study introduces a novel reconstruction-based modeling approach enhanced by negative learning,employing a Bidirectional Long Short-Term Memory(BiLSTM)network explicitly trained to accurately reconstruct non-seismic geomagnetic signals while intentionally amplifying reconstruction errors for seismic signals.By penalizing the model for accurately reconstructing seismic anomalies,the negative learning approach effectively magnifies the differences between normal and anomalous data.This strategic differentiation enhances the sensitivity of the BiLSTM network,enabling improved detection of subtle geomagnetic anomalies that may serve as earthquake precursors.Experimental validation clearly demonstrated statistically significant higher reconstruction errors for seismic signals compared to non-seismic signals,confirmed through the Mann-Whitney U test with a p-value of 0.0035 for Root Mean Square Error(RMSE).These results provide compelling evidence of the enhanced anomaly detection capability achieved through negative learning.Unlike traditional classification-based methods,negative learning explicitly encourages sensitivity to subtle precursor signals embedded within complex geomagnetic data,establishing a robust basis for further development of reliable earthquake prediction methods.展开更多
Interaction between the converter and the grid may lead to harmonic oscillations.The impedance-based method is an effective way to deal with the stability issue.In this study,the impedance-based method is used to inve...Interaction between the converter and the grid may lead to harmonic oscillations.The impedance-based method is an effective way to deal with the stability issue.In this study,the impedance-based method is used to investigate the small-signal stability of a cascaded 12-pulse line-commutated converter-based high-voltage direct current(LCC-HVDC)transmission system.In the modeling part,the impedance models of the single rectifier and inverter are established respectively with consideration to the effect of frequency coupling,which has improved the accuracy of the models.Based on the models,the AC impedance models of the cascaded LCC-HVDC transmission system are established both on the rectifier and inverter side.In the stability analysis part,the stability of the system is analyzed under different working conditions.The simulation results reveal that the established impedance model can properly represent the stability of this system.The findings of this study can provide a theoretical reference for the stability design and oscillation suppression strategy of LCC-HVDC transmission systems and LCC interconnected systems.展开更多
This paper introduces a novel fractional-order model based on the Caputo-Fabrizio(CF)derivative for analyzing computer virus propagation in networked environments.The model partitions the computer population into four...This paper introduces a novel fractional-order model based on the Caputo-Fabrizio(CF)derivative for analyzing computer virus propagation in networked environments.The model partitions the computer population into four compartments:susceptible,latently infected,breaking-out,and antivirus-capable systems.By employing the CF derivative—which uses a nonsingular exponential kernel—the framework effectively captures memory-dependent and nonlocal characteristics intrinsic to cyber systems,aspects inadequately represented by traditional integer-order models.Under Lipschitz continuity and boundedness assumptions,the existence and uniqueness of solutions are rigorously established via fixed-point theory.We develop a tailored two-step Adams-Bashforth numerical scheme for the CF framework and prove its second-order accuracy.Extensive numerical simulations across various fractional orders reveal that memory effects significantly influence virus transmission and control dynamics;smaller fractional orders produce more pronounced memory effects,delaying both infection spread and antivirus activation.Further theoretical analysis,including Hyers-Ulam stability and sensitivity assessments,reinforces the model’s robustness and identifies key parameters governing virus dynamics.The study also extends the framework to incorporate stochastic effects through a stochastic CF formulation.These results underscore fractional-order modeling as a powerful analytical tool for developing robust and effective cybersecurity strategies.展开更多
In-space cable-driven manipulators exhibit several advantages,such as a large range of motion,high dexterity,and lightweight structure.However,kinematic and dynamic analysis play an essential role in designing a cable...In-space cable-driven manipulators exhibit several advantages,such as a large range of motion,high dexterity,and lightweight structure.However,kinematic and dynamic analysis play an essential role in designing a cable-driven manipulator.In this paper,the kinematic analysis of a type of cable-driven manipulator is performed,and a motion planning scheme is conducted to actuate this manipulator.Moreover,a flexible multi-body dynamic model of a cable-driven manipulator considering the frictional contact between the cables and pulleys is established.To describe properties such as flexibility,vibration,and variable length of the cable,this paper utilizes reducedorder beam elements of the Absolute Nodal Coordinates Formulation(ANCF)in Arbitrary Lagrangian Eulerian(ALE)framework.Additionally,a virtual element is introduced to model the contact segment in the cable-pulley system.A tension decay factor is employed to account for the friction in the contact segment.To validate the proposed method,a semi-analytical model based on D'Alembert's principle is established.Cross-verification is performed to validate the accuracy of both models.The model is further applied to simulate the rotation of the cable-driven manipulator with different structural parameters and frictional factors.The results from the analyses provide valuable guidance for the design and motion control of the in-space cable-driven manipulator.Finally,a prototype of a single module is manufactured and tested.Ground experiments are carried out to verify the kinematic and dynamic models.展开更多
Predicting rock blasting outcomes in mining has been crucial since its inception.Blasting remains the most energy-and cost-efficient method for rock breaking and is often the only practical option.However,the mechanis...Predicting rock blasting outcomes in mining has been crucial since its inception.Blasting remains the most energy-and cost-efficient method for rock breaking and is often the only practical option.However,the mechanism is complex,influenced by various rock properties,explosives,and blast design parameters,making their effects difficult to quantify.Traditional stress-based models struggle with many parameters,such as stress and Poisson's ratio,which are challenging to measure in the field.Empirical models,though simpler,often oversimplify blast conditions.Both types of models are limited to simulating a few blastholes and cannot handle full-scale blasts involving hundreds of blastholes.However,modeling full-scale blasts with all blast design parameters is most required for modern mining applications.This paper presents a novel strain-based modeling approach for blasting and geomechanical applications,utilizing measurable variables such as particle velocity,strain,and displacement.By bypassing complex constitutive relations,strain-based models capture critical blasting trends and simulate full-scale blasts with full-blast design parameters with minimal calibration.The framework encompasses field strain measurements,model construction based on measurable variables,and laboratoryderived strain-failure criteria,each offering potential for future enhancement.Additionally,a standardized field test for site characterization is recommended.The approach is demonstrated through the Multiple Blasthole Fragmentation model,which simulates rock fragmentation and fragment strain during blasting,highlighting the practicality and effectiveness of strain-based modeling for multiple blasthole blasts.Moreover,this approach extends beyond blasting,with potential applications in highwall stability monitoring and other geomechanical applications.Strain-based modeling provides a simplified yet effective solution,avoiding the complexities of rock constitutive relations and field stress measurements while enabling full-blast design simulations for large-scale field blasts.展开更多
With the rapid proliferation of electric vehicles,their charging loads pose new challenges to power grid stability and operational efficiency.To address this,this study employs a Monte Carlo simulation model to analyz...With the rapid proliferation of electric vehicles,their charging loads pose new challenges to power grid stability and operational efficiency.To address this,this study employs a Monte Carlo simulation model to analyze the charging load characteristics of six battery electric vehicle categories in Hebei Province,leveraging multi-source probabilistic distribution data under typical operational scenarios.The findings reveal that electric vehicle charging loads are primarily concentrated during midday and nighttime periods,with significant load fluctuations exerting substantial pressure on the grid.In response,this paper proposes strategic interventions including optimized charging infrastructure planning,time-of-use electricity pricing mechanisms,and smart charging technologies to balance grid loads.The results provide a theoretical foundation for electric vehicle load forecasting,smart grid dispatching,and vehicle-grid integration,thereby enhancing grid operational efficiency and sustainability.展开更多
Activation pruning reduces neural network complexity by eliminating low-importance neuron activations,yet identifying the critical pruning threshold—beyond which accuracy rapidly deteriorates—remains computationally...Activation pruning reduces neural network complexity by eliminating low-importance neuron activations,yet identifying the critical pruning threshold—beyond which accuracy rapidly deteriorates—remains computationally expensive and typically requires exhaustive search.We introduce a thermodynamics-inspired framework that treats activation distributions as energy-filtered physical systems and employs the free energy of activations as a principled evaluation metric.Phase-transition-like phenomena in the free-energy profile—such as extrema,inflection points,and curvature changes—yield reliable estimates of the critical pruning threshold,providing a theoretically grounded means of predicting sharp accuracy degradation.To further enhance efficiency,we propose a renormalized free energy technique that approximates full-evaluation free energy using only the activation distribution of the unpruned network.This eliminates repeated forward passes,dramatically reducing computational overhead and achieving speedups of up to 550×for MLPs.Extensive experiments across diverse vision architectures(MLP,CNN,ResNet,MobileNet,Vision Transformer)and text models(LSTM,BERT,ELECTRA,T5,GPT-2)on multiple datasets validate the generality,robustness,and computational efficiency of our approach.Overall,this work establishes a theoretically grounded and practically effective framework for activation pruning,bridging the gap between analytical understanding and efficient deployment of sparse neural networks.展开更多
Thermal spalling in heterogeneous rocks under rapid heating poses critical risks to deep mining and geothermal operations.In this study,we develop a coupled thermal-mechanical-damage(TM D)model that explicitly incorpo...Thermal spalling in heterogeneous rocks under rapid heating poses critical risks to deep mining and geothermal operations.In this study,we develop a coupled thermal-mechanical-damage(TM D)model that explicitly incorporates Weibull distributed heterogeneity to a single fracture in rock,and validate it against ceramic quenching and granite acoustic emission experiments.Distance based generalized sensitivity analysis(DGSA)is applied to quantify the influence and interactions of key parameters,revealing the dominant controls on spalling onset,severity,and damage morphology.The results demonstrate that thermal stress dominates crack initiation and propagation,that lateral constraints can significantly delay and suppress spalling,and that material heterogeneity markedly influences peak stress and damage modes within a certain range of thermal expansion coefficient and has multiple effects on thermal spalling.This study provides a theoretical basis for quantitative assessment and parameter optimization of thermal spalling processes in rock masses.展开更多
The increasing electrification of urban transportation,i.e.,subways and electric vehicles(EV),brings more interactions between the power system and transportation system and further results in fault propagation across...The increasing electrification of urban transportation,i.e.,subways and electric vehicles(EV),brings more interactions between the power system and transportation system and further results in fault propagation across them.To analyze vulnerability of the coupling system under extreme events,this paper establishes a multi-layer urban electric-transportation interdependent network(ETIN)model.First,a weighted coupled metro-road traffic network(CTN)model and network path planning approach are proposed.A prospect theory-based failure load redistribution(FLR)method is further established to account for uncertainty of TN link capacity affected by power supply.Second,topology and emergency control strategy of power network(PN)are modeled,followed by formulation of multi-layer ETIN model.In particular,the inter-layer fault propagation from PN to TN is modeled based on power supply correlation strength,while from TN to PN is modeled based on traffic flow.A few indexes are then defined to quantify vulnerability of ETIN under deliberate attack.Finally,the proposed method is verified on an electric-transportation system to show influence of fault propagations within ETIN on its vulnerability under extreme events.展开更多
The northern segment of the North-South Seismic Belt is characterized by intense crustal deformation,well-developed active tectonics,and frequent occurrences of strong earthquakes.Therefore,conducting a Probabilistic ...The northern segment of the North-South Seismic Belt is characterized by intense crustal deformation,well-developed active tectonics,and frequent occurrences of strong earthquakes.Therefore,conducting a Probabilistic Seismic Hazard Analysis(PSHA)for this region is of significant importance for supporting seismic fortification in major engineering projects and formulating disaster prevention and mitigation policies.In this study,a composite seismic source model was constructed by integrating data on historical earthquakes,active faults,and paleoseismicity.Furthermore,a logic tree framework was employed to quantify epistemic uncertainties,enabling a systematic seismic hazard assessment of the region.To more accurately characterize the spatial heterogeneity of seismic activity,improvements were made to both the Circular Spatial Smoothing Model(CSSM)with a fixed radius and the Adaptive Spatial Smoothing Model(ASSM),with full consideration given to the spatiotemporal completeness of historical earthquake magnitudes.Regarding the CSSM,for scenarios involving small sample sizes in earthquake catalogs,the cross-validation method proposed in this study demonstrated higher robustness than the maximum likelihood method in determining the optimal correlation distance.Performance evaluation results indicate that while both models effectively characterize seismic activity,the ASSM exhibits superior overall predictive performance compared to the CSSM,owing to its ability to adaptively adjust the smoothing radius according to seismic density.Significant discrepancies were observed in the Peak Ground Acceleration(PGA)results calculated with a 10%probability of exceedance in 50 years across different combinations of seismic source models.The single spatially smoothed point-source model yielded a maximum PGA of approximately 0.52 g,with high-value areas concentrated near historical epicenters,thereby significantly underestimating the hazard associated with major fault zones.When combined with the simple fault-source model,the maximum PGA increased to 0.8 g,with high-value zones exhibiting a zonal distribution along faults;however,the risk remained underestimated for faults with low slip rates that are nevertheless approaching their recurrence cycles.Following the introduction of the time-dependent characteristic fault-source model,local PGA values for faults in the middle-to-late stages of their recurrence cycles increased by a factor of 2 to 7 compared to the single model.These results demonstrate that the characteristic fault-source model reasonably delineates the time-dependence of large earthquake recurrence,thereby providing a more accurate assessment of imminent seismic risks.By comprehensively applying the improved spatially smoothed pointsource model,the simple fault-source model,and the characteristic fault-source model,the following faults within the region were identified as having high seismic hazard:the Huangxianggou,Zhangxian,and Tianshui segments of the Xiqinling northern edge fault;the Maqin-Maqu segment of the Dongkunlun fault;the Longriqu fault;the Maoergai fault;the Elashan fault;the Riyueshan fault;the eastern segment of the Lenglongling fault;the Maxianshan segment of the Maxianshan northern Margin fault;and the Maomaoshan-Jinqianghe segment of the Laohushan-Maomaoshan fault.As these faults are located within seismic gaps or are approaching the recurrence periods of large earthquakes,they should be prioritized for current and future seismic monitoring as well as disaster prevention and mitigation efforts.展开更多
Cascading failures pose a serious threat to the survivability of underwater unmanned swarm networks(UUSNs),significantly limiting their service ability in collaborative missions such as military reconnaissance and env...Cascading failures pose a serious threat to the survivability of underwater unmanned swarm networks(UUSNs),significantly limiting their service ability in collaborative missions such as military reconnaissance and environmental monitoring.Existing failure models primarily focus on power grids and traffic systems,and don't address the unique challenges of weak-communication UUSNs.In UUSNs,cascading failure present a complex and dynamic process driven by the coupling of unstable acoustic channels,passive node drift,adversarial attacks,and network heterogeneity.To address these challenges,a directed weighted graph model of UUSNs is first developed,in which node positions are updated according to ocean-current-driven drift and link weights reflect the probability of successful acoustic transmission.Building on this UUSNs graph model,a cascading failure model is proposed that integrates a normal-failure-recovery state-cycle mechanism,multiple attack strategies,and routingbased load redistribution.Finally,under a five-level connectivity UUSNs scheme,simulations are conducted to analyze how dynamic topology,network load,node recovery delay,and attack modes jointly affect network survivability.The main findings are:(1)moderate node drift can improve survivability by activating weak links;(2)based-energy routing(BER)outperform based-depth routing(BDR)in harsh conditions;(3)node self-recovery time is critical to network survivability;(4)traditional degree-based critical node metrics are inadequate for weak-communication UUSNs.These results provide a theoretical foundation for designing robust survivability mechanisms in weak-communication UUSNs.展开更多
To assess the effectiveness of vaccination in contaminated environments,this study introduces a modeling framework that encompasses two transmission routes,namely direct human-to-human contact and indirect human-to-en...To assess the effectiveness of vaccination in contaminated environments,this study introduces a modeling framework that encompasses two transmission routes,namely direct human-to-human contact and indirect human-to-environment contact,as well as the implementation of new M72/AS01_(E)vaccine.Motivated by this,a coupled age-structured tuberculosis(TB)model is proposed.Its well-posedness requirement is verified using the integrated semigroup theory.Furthermore,this study presents a comprehensive analysis of threshold dynamics associated with the proposed model.Specifically,the global stability of the disease-free and positive steady states is demonstrated by employing Lyapunov functionals.Lastly,the effects of the vaccination with M72/AS01_(E)and contaminated environments on TB control are numerically simulated.Experimental results indicate that high concentrations of Mycobacterium tuberculosis in contaminated environments may somewhat impede TB control efforts,but that large-scale deployment of new vaccine could significantly reduce the prevalence of TB.展开更多
Owing to intensified globalization and informatization,the structures of the urban scale hierarchy and urban networks between cities have become increasingly intertwined,resulting in different spatial effects.Therefor...Owing to intensified globalization and informatization,the structures of the urban scale hierarchy and urban networks between cities have become increasingly intertwined,resulting in different spatial effects.Therefore,this paper analyzes the spatial interaction between urban scale hierarchy and urban networks in China from 2019 to 2023,drawing on Baidu migration data and employing a spatial simultaneous equation model.The results reveal a significant positive spatial correlation between cities with higher hierarchy and those with greater network centrality.Within a static framework,we identify a positive interaction between urban scale hierarchy and urban network centrality,while their spatial cross-effects manifest as negative neighborhood interactions based on geographical distance and positive cross-scale interactions shaped by network connections.Within a dynamic framework,changes in urban scale hierarchy and urban networks are mutually reinforcing,thereby widening disparities within the urban hierarchy.Furthermore,an increase in a city’s network centrality had a dampening effect on the population growth of neighboring cities and network-connected cities.This study enhances understanding of the spatial organisation of urban systems and offers insights for coordinated regional development.展开更多
Deep rock engineering is affected by coupled thermo-hydro-mechanical(THM)-dynamic fields,necessitating the elucidation of the dynamic mechanical behavior and failure mechanisms.This study utilized a Multi-field Couple...Deep rock engineering is affected by coupled thermo-hydro-mechanical(THM)-dynamic fields,necessitating the elucidation of the dynamic mechanical behavior and failure mechanisms.This study utilized a Multi-field Coupled Controlled Split Hopkinson Pressure Bar(MCC-SHPB)system to elucidate the cross-scale dynamic responses of rocks and the boundaries of failure modes under THM coupling.Impact tests were conducted on green sandstone under coupled conditions of temperature(25℃-80℃),confining pressure(0-15 MPa),and seepage water pressure(0-15 MPa).Scanning electron microscopy(SEM)microstructural characterization and COMSOL Multiphysics numerical simulations were conducted,and a dynamic constitutive theoretical framework and failure-prediction methodology were established.We investigated the impact toughness index(I_(t)),dynamic modulus(E_(d)),dynamic triaxial compressive strength(TCS_(d)),fragmentation degree(W),and failure modes of green sandstone under thermo-confining pressure-seepage-impact loading conditions.The key findings reveal that the(I_(t))reflects different energy regulation mechanisms across different confining pressure regimes.Thermal-microcrack interactions dominate at low pressure,and energy absorption prevails at high pressure.A triphasic dynamic modulus model captures stiffness evolution under energy-driven conditions,revealing cross-scale crack nucleation-propagation and fragment reorganization.The TCSd inflection point signifies energy dissipation shifts,causing nonlinear skeleton bearing-capacity degradation.A critical criterion based on the W was established to distinguish between the two failure modes and predict the unstable failure initiation.Numerical simulations were used to elucidate the effects of inertia-dominated crack propagation and stress wave interference,validating the critical criterion and the predictive accuracy of the theoretical model during cross-scale failure.This study provides a theoretical foundation for assessing the dynamic stability of rock masses subjected to multi-field coupling during deep resource exploitation.展开更多
Anti-jamming performance evaluation has recently received significant attention. For Link-16, the anti-jamming performance evaluation and selection of the optimal anti-jamming technologies are urgent problems to be so...Anti-jamming performance evaluation has recently received significant attention. For Link-16, the anti-jamming performance evaluation and selection of the optimal anti-jamming technologies are urgent problems to be solved. A comprehensive evaluation method is proposed, which combines grey relational analysis (GRA) and cloud model, to evaluate the anti-jamming performances of Link-16. Firstly, on the basis of establishing the anti-jamming performance evaluation indicator system of Link-16, the linear combination of analytic hierarchy process(AHP) and entropy weight method (EWM) are used to calculate the combined weight. Secondly, the qualitative and quantitative concept transformation model, i.e., the cloud model, is introduced to evaluate the anti-jamming abilities of Link-16 under each jamming scheme. In addition, GRA calculates the correlation degree between evaluation indicators and the anti-jamming performance of Link-16, and assesses the best anti-jamming technology. Finally, simulation results prove that the proposed evaluation model can achieve the objective of feasible and practical evaluation, which opens up a novel way for the research of anti-jamming performance evaluations of Link-16.展开更多
Shotcrete is one of the common solutions for shallow sliding.It works by forming a protective layer with high strength and cementing the loose soil particles on the slope surface to prevent shallow sliding.However,the...Shotcrete is one of the common solutions for shallow sliding.It works by forming a protective layer with high strength and cementing the loose soil particles on the slope surface to prevent shallow sliding.However,the solidification time of conventional cement paste is long when shotcrete is used to treat cohesionless soil landslide.The idea of reinforcing slope with polyurethane solidified soil(i.e.,mixture of polyurethane and sand)was proposed.Model tests and finite element analysis were carried out to study the effectiveness of the proposed new method on the emergency treatment of cohesionless soil landslide.Surcharge loading on the crest of the slope was applied step by step until landslide was triggered so as to test and compare the stability and bearing capacity of slope models with different conditions.The simulated slope displacements were relatively close to the measured results,and the simulated slope deformation characteristics were in good agreement with the observed phenomena,which verifies the accuracy of the numerical method.Under the condition of surcharge loading on the crest of the slope,the unreinforced slope slid when the surcharge loading exceeded 30 k Pa,which presented a failure mode of local instability and collapse at the shallow layer of slope top.The reinforced slope remained stable even when the surcharge loading reached 48 k Pa.The displacement of the reinforced slope was reduced by more than 95%.Overall,this study verifies the effectiveness of polyurethane in the emergency treatment of cohesionless soil landslide and should have broad application prospects in the field of geological disasters concerning the safety of people's live.展开更多
Large-scale Language Models(LLMs)have achieved significant breakthroughs in Natural Language Processing(NLP),driven by the pre-training and fine-tuning paradigm.While this approach allows models to specialize in speci...Large-scale Language Models(LLMs)have achieved significant breakthroughs in Natural Language Processing(NLP),driven by the pre-training and fine-tuning paradigm.While this approach allows models to specialize in specific tasks with reduced training costs,the substantial memory requirements during fine-tuning present a barrier to broader deployment.Parameter-Efficient Fine-Tuning(PEFT)techniques,such as Low-Rank Adaptation(LoRA),and parameter quantization methods have emerged as solutions to address these challenges by optimizing memory usage and computational efficiency.Among these,QLoRA,which combines PEFT and quantization,has demonstrated notable success in reducing memory footprints during fine-tuning,prompting the development of various QLoRA variants.Despite these advancements,the quantitative impact of key variables on the fine-tuning performance of quantized LLMs remains underexplored.This study presents a comprehensive analysis of these key variables,focusing on their influence across different layer types and depths within LLM architectures.Our investigation uncovers several critical findings:(1)Larger layers,such as MLP layers,can maintain performance despite reductions in adapter rank,while smaller layers,like self-attention layers,aremore sensitive to such changes;(2)The effectiveness of balancing factors depends more on specific values rather than layer type or depth;(3)In quantization-aware fine-tuning,larger layers can effectively utilize smaller adapters,whereas smaller layers struggle to do so.These insights suggest that layer type is a more significant determinant of fine-tuning success than layer depth when optimizing quantized LLMs.Moreover,for the same discount of trainable parameters,reducing the trainable parameters in a larger layer is more effective in preserving fine-tuning accuracy than in a smaller one.This study provides valuable guidance for more efficient fine-tuning strategies and opens avenues for further research into optimizing LLM fine-tuning in resource-constrained environments.展开更多
Background Cotton is one of the most important commercial crops after food crops,especially in countries like India,where it’s grown extensively under rainfed conditions.Because of its usage in multiple industries,su...Background Cotton is one of the most important commercial crops after food crops,especially in countries like India,where it’s grown extensively under rainfed conditions.Because of its usage in multiple industries,such as textile,medicine,and automobile industries,it has greater commercial importance.The crop’s performance is greatly influenced by prevailing weather dynamics.As climate changes,assessing how weather changes affect crop performance is essential.Among various techniques that are available,crop models are the most effective and widely used tools for predicting yields.Results This study compares statistical and machine learning models to assess their ability to predict cotton yield across major producing districts of Karnataka,India,utilizing a long-term dataset spanning from 1990 to 2023 that includes yield and weather factors.The artificial neural networks(ANNs)performed superiorly with acceptable yield deviations ranging within±10%during both vegetative stage(F1)and mid stage(F2)for cotton.The model evaluation metrics such as root mean square error(RMSE),normalized root mean square error(nRMSE),and modelling efficiency(EF)were also within the acceptance limits in most districts.Furthermore,the tested ANN model was used to assess the importance of the dominant weather factors influencing crop yield in each district.Specifically,the use of morning relative humidity as an individual parameter and its interaction with maximum and minimum tempera-ture had a major influence on cotton yield in most of the yield predicted districts.These differences highlighted the differential interactions of weather factors in each district for cotton yield formation,highlighting individual response of each weather factor under different soils and management conditions over the major cotton growing districts of Karnataka.Conclusions Compared with statistical models,machine learning models such as ANNs proved higher efficiency in forecasting the cotton yield due to their ability to consider the interactive effects of weather factors on yield forma-tion at different growth stages.This highlights the best suitability of ANNs for yield forecasting in rainfed conditions and for the study on relative impacts of weather factors on yield.Thus,the study aims to provide valuable insights to support stakeholders in planning effective crop management strategies and formulating relevant policies.展开更多
DNA microarray technology is an extremely effective technique for studying gene expression patterns in cells, and the main challenge currently faced by this technology is how to analyze the large amount of gene expres...DNA microarray technology is an extremely effective technique for studying gene expression patterns in cells, and the main challenge currently faced by this technology is how to analyze the large amount of gene expression data generated. To address this, this paper employs a mixed-effects model to analyze gene expression data. In terms of data selection, 1176 genes from the white mouse gene expression dataset under two experimental conditions were chosen, setting up two conditions: pneumococcal infection and no infection, and constructing a mixed-effects model. After preprocessing the gene chip information, the data were imported into the model, preliminary results were calculated, and permutation tests were performed to biologically validate the preliminary results using GSEA. The final dataset consists of 20 groups of gene expression data from pneumococcal infection, which categorizes functionally related genes based on the similarity of their expression profiles, facilitating the study of genes with unknown functions.展开更多
基金supported by University Grant Agency of Matej Bel University in Banská Bystrica project number UGA-14-PDS-2025.
文摘It is known that correlation does not imply causality.Some relationships identified in the analysis of data are coincidental or unknown,and some are produced by real-world causality of the situation,which is problematic,since there is a need to differentiate between these two scenarios.Until recently,the proper−semantic−causality of the relationship could have been determined only by human experts from the area of expertise of the studied data.This has changed with the advance of large language models,which are often utilized as surrogates for such human experts,making the process automated and readily available to all data analysts.This motivates the main objective of this work,which is to introduce the design and implementation of a large language model-based semantic causality evaluator based on correlation analysis,together with its visual analysis model called Causal heatmap.After the implementation itself,the model is evaluated from the point of view of the quality of the visual model,from the point of view of the quality of causal evaluation based on large language models,and from the point of view of comparative analysis,while the results reached in the study highlight the usability of large language models in the task and the potential of the proposed approach in the analysis of unknown datasets.The results of the experimental evaluation demonstrate the usefulness of the Causal heatmap method,supported by the evident highlighting of interesting relationships,while suppressing irrelevant ones.
基金funded by the Ministry of Higher Education through Universiti Putra Malaysia(UPM)under Grant FRGS/1/2023/STG07/UPM/02/4.
文摘Detecting geomagnetic anomalies preceding earthquakes is a challenging yet promising area of research that has gained increasing attention in recent years.This study introduces a novel reconstruction-based modeling approach enhanced by negative learning,employing a Bidirectional Long Short-Term Memory(BiLSTM)network explicitly trained to accurately reconstruct non-seismic geomagnetic signals while intentionally amplifying reconstruction errors for seismic signals.By penalizing the model for accurately reconstructing seismic anomalies,the negative learning approach effectively magnifies the differences between normal and anomalous data.This strategic differentiation enhances the sensitivity of the BiLSTM network,enabling improved detection of subtle geomagnetic anomalies that may serve as earthquake precursors.Experimental validation clearly demonstrated statistically significant higher reconstruction errors for seismic signals compared to non-seismic signals,confirmed through the Mann-Whitney U test with a p-value of 0.0035 for Root Mean Square Error(RMSE).These results provide compelling evidence of the enhanced anomaly detection capability achieved through negative learning.Unlike traditional classification-based methods,negative learning explicitly encourages sensitivity to subtle precursor signals embedded within complex geomagnetic data,establishing a robust basis for further development of reliable earthquake prediction methods.
基金supported in part by the National Natural Science Foundation of China under 52125704 and 51937001.
文摘Interaction between the converter and the grid may lead to harmonic oscillations.The impedance-based method is an effective way to deal with the stability issue.In this study,the impedance-based method is used to investigate the small-signal stability of a cascaded 12-pulse line-commutated converter-based high-voltage direct current(LCC-HVDC)transmission system.In the modeling part,the impedance models of the single rectifier and inverter are established respectively with consideration to the effect of frequency coupling,which has improved the accuracy of the models.Based on the models,the AC impedance models of the cascaded LCC-HVDC transmission system are established both on the rectifier and inverter side.In the stability analysis part,the stability of the system is analyzed under different working conditions.The simulation results reveal that the established impedance model can properly represent the stability of this system.The findings of this study can provide a theoretical reference for the stability design and oscillation suppression strategy of LCC-HVDC transmission systems and LCC interconnected systems.
基金supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University(IMSIU)(grant number IMSIU-DDRSP2601).
文摘This paper introduces a novel fractional-order model based on the Caputo-Fabrizio(CF)derivative for analyzing computer virus propagation in networked environments.The model partitions the computer population into four compartments:susceptible,latently infected,breaking-out,and antivirus-capable systems.By employing the CF derivative—which uses a nonsingular exponential kernel—the framework effectively captures memory-dependent and nonlocal characteristics intrinsic to cyber systems,aspects inadequately represented by traditional integer-order models.Under Lipschitz continuity and boundedness assumptions,the existence and uniqueness of solutions are rigorously established via fixed-point theory.We develop a tailored two-step Adams-Bashforth numerical scheme for the CF framework and prove its second-order accuracy.Extensive numerical simulations across various fractional orders reveal that memory effects significantly influence virus transmission and control dynamics;smaller fractional orders produce more pronounced memory effects,delaying both infection spread and antivirus activation.Further theoretical analysis,including Hyers-Ulam stability and sensitivity assessments,reinforces the model’s robustness and identifies key parameters governing virus dynamics.The study also extends the framework to incorporate stochastic effects through a stochastic CF formulation.These results underscore fractional-order modeling as a powerful analytical tool for developing robust and effective cybersecurity strategies.
基金co-supported by the National Natural Science Foundation of China(Nos.12102034 and 12125201)the Open Fund of State Key Laboratory of Robotics and Systems(HIT),China。
文摘In-space cable-driven manipulators exhibit several advantages,such as a large range of motion,high dexterity,and lightweight structure.However,kinematic and dynamic analysis play an essential role in designing a cable-driven manipulator.In this paper,the kinematic analysis of a type of cable-driven manipulator is performed,and a motion planning scheme is conducted to actuate this manipulator.Moreover,a flexible multi-body dynamic model of a cable-driven manipulator considering the frictional contact between the cables and pulleys is established.To describe properties such as flexibility,vibration,and variable length of the cable,this paper utilizes reducedorder beam elements of the Absolute Nodal Coordinates Formulation(ANCF)in Arbitrary Lagrangian Eulerian(ALE)framework.Additionally,a virtual element is introduced to model the contact segment in the cable-pulley system.A tension decay factor is employed to account for the friction in the contact segment.To validate the proposed method,a semi-analytical model based on D'Alembert's principle is established.Cross-verification is performed to validate the accuracy of both models.The model is further applied to simulate the rotation of the cable-driven manipulator with different structural parameters and frictional factors.The results from the analyses provide valuable guidance for the design and motion control of the in-space cable-driven manipulator.Finally,a prototype of a single module is manufactured and tested.Ground experiments are carried out to verify the kinematic and dynamic models.
文摘Predicting rock blasting outcomes in mining has been crucial since its inception.Blasting remains the most energy-and cost-efficient method for rock breaking and is often the only practical option.However,the mechanism is complex,influenced by various rock properties,explosives,and blast design parameters,making their effects difficult to quantify.Traditional stress-based models struggle with many parameters,such as stress and Poisson's ratio,which are challenging to measure in the field.Empirical models,though simpler,often oversimplify blast conditions.Both types of models are limited to simulating a few blastholes and cannot handle full-scale blasts involving hundreds of blastholes.However,modeling full-scale blasts with all blast design parameters is most required for modern mining applications.This paper presents a novel strain-based modeling approach for blasting and geomechanical applications,utilizing measurable variables such as particle velocity,strain,and displacement.By bypassing complex constitutive relations,strain-based models capture critical blasting trends and simulate full-scale blasts with full-blast design parameters with minimal calibration.The framework encompasses field strain measurements,model construction based on measurable variables,and laboratoryderived strain-failure criteria,each offering potential for future enhancement.Additionally,a standardized field test for site characterization is recommended.The approach is demonstrated through the Multiple Blasthole Fragmentation model,which simulates rock fragmentation and fragment strain during blasting,highlighting the practicality and effectiveness of strain-based modeling for multiple blasthole blasts.Moreover,this approach extends beyond blasting,with potential applications in highwall stability monitoring and other geomechanical applications.Strain-based modeling provides a simplified yet effective solution,avoiding the complexities of rock constitutive relations and field stress measurements while enabling full-blast design simulations for large-scale field blasts.
基金funded by Humanities and Social Sciences of Ministry of Education Planning Fund of China,grant number 21YJA790009National Natural Science Foundation of China,grant number 72140001.
文摘With the rapid proliferation of electric vehicles,their charging loads pose new challenges to power grid stability and operational efficiency.To address this,this study employs a Monte Carlo simulation model to analyze the charging load characteristics of six battery electric vehicle categories in Hebei Province,leveraging multi-source probabilistic distribution data under typical operational scenarios.The findings reveal that electric vehicle charging loads are primarily concentrated during midday and nighttime periods,with significant load fluctuations exerting substantial pressure on the grid.In response,this paper proposes strategic interventions including optimized charging infrastructure planning,time-of-use electricity pricing mechanisms,and smart charging technologies to balance grid loads.The results provide a theoretical foundation for electric vehicle load forecasting,smart grid dispatching,and vehicle-grid integration,thereby enhancing grid operational efficiency and sustainability.
基金output of a research project implemented as part of the Basic Research Program at HSE University。
文摘Activation pruning reduces neural network complexity by eliminating low-importance neuron activations,yet identifying the critical pruning threshold—beyond which accuracy rapidly deteriorates—remains computationally expensive and typically requires exhaustive search.We introduce a thermodynamics-inspired framework that treats activation distributions as energy-filtered physical systems and employs the free energy of activations as a principled evaluation metric.Phase-transition-like phenomena in the free-energy profile—such as extrema,inflection points,and curvature changes—yield reliable estimates of the critical pruning threshold,providing a theoretically grounded means of predicting sharp accuracy degradation.To further enhance efficiency,we propose a renormalized free energy technique that approximates full-evaluation free energy using only the activation distribution of the unpruned network.This eliminates repeated forward passes,dramatically reducing computational overhead and achieving speedups of up to 550×for MLPs.Extensive experiments across diverse vision architectures(MLP,CNN,ResNet,MobileNet,Vision Transformer)and text models(LSTM,BERT,ELECTRA,T5,GPT-2)on multiple datasets validate the generality,robustness,and computational efficiency of our approach.Overall,this work establishes a theoretically grounded and practically effective framework for activation pruning,bridging the gap between analytical understanding and efficient deployment of sparse neural networks.
基金funded by the National Natural Science Foundation of China(Nos.52574100,52574001,and 52311530070)the Major National Science and Technology Project for Deep Earth of China(No.2024ZD1003805)+1 种基金the Fundamental Research Funds for the Central Universities of China(No.FRF-IDRY-20-003,Interdisciplinary Research Project for Young Teachers of USTB)DE gratefully acknowledges support from the G.Albert Shoemaker endowment.
文摘Thermal spalling in heterogeneous rocks under rapid heating poses critical risks to deep mining and geothermal operations.In this study,we develop a coupled thermal-mechanical-damage(TM D)model that explicitly incorporates Weibull distributed heterogeneity to a single fracture in rock,and validate it against ceramic quenching and granite acoustic emission experiments.Distance based generalized sensitivity analysis(DGSA)is applied to quantify the influence and interactions of key parameters,revealing the dominant controls on spalling onset,severity,and damage morphology.The results demonstrate that thermal stress dominates crack initiation and propagation,that lateral constraints can significantly delay and suppress spalling,and that material heterogeneity markedly influences peak stress and damage modes within a certain range of thermal expansion coefficient and has multiple effects on thermal spalling.This study provides a theoretical basis for quantitative assessment and parameter optimization of thermal spalling processes in rock masses.
文摘The increasing electrification of urban transportation,i.e.,subways and electric vehicles(EV),brings more interactions between the power system and transportation system and further results in fault propagation across them.To analyze vulnerability of the coupling system under extreme events,this paper establishes a multi-layer urban electric-transportation interdependent network(ETIN)model.First,a weighted coupled metro-road traffic network(CTN)model and network path planning approach are proposed.A prospect theory-based failure load redistribution(FLR)method is further established to account for uncertainty of TN link capacity affected by power supply.Second,topology and emergency control strategy of power network(PN)are modeled,followed by formulation of multi-layer ETIN model.In particular,the inter-layer fault propagation from PN to TN is modeled based on power supply correlation strength,while from TN to PN is modeled based on traffic flow.A few indexes are then defined to quantify vulnerability of ETIN under deliberate attack.Finally,the proposed method is verified on an electric-transportation system to show influence of fault propagations within ETIN on its vulnerability under extreme events.
基金supported by the National Key R&D Program of China(No.2022YFC3003502).
文摘The northern segment of the North-South Seismic Belt is characterized by intense crustal deformation,well-developed active tectonics,and frequent occurrences of strong earthquakes.Therefore,conducting a Probabilistic Seismic Hazard Analysis(PSHA)for this region is of significant importance for supporting seismic fortification in major engineering projects and formulating disaster prevention and mitigation policies.In this study,a composite seismic source model was constructed by integrating data on historical earthquakes,active faults,and paleoseismicity.Furthermore,a logic tree framework was employed to quantify epistemic uncertainties,enabling a systematic seismic hazard assessment of the region.To more accurately characterize the spatial heterogeneity of seismic activity,improvements were made to both the Circular Spatial Smoothing Model(CSSM)with a fixed radius and the Adaptive Spatial Smoothing Model(ASSM),with full consideration given to the spatiotemporal completeness of historical earthquake magnitudes.Regarding the CSSM,for scenarios involving small sample sizes in earthquake catalogs,the cross-validation method proposed in this study demonstrated higher robustness than the maximum likelihood method in determining the optimal correlation distance.Performance evaluation results indicate that while both models effectively characterize seismic activity,the ASSM exhibits superior overall predictive performance compared to the CSSM,owing to its ability to adaptively adjust the smoothing radius according to seismic density.Significant discrepancies were observed in the Peak Ground Acceleration(PGA)results calculated with a 10%probability of exceedance in 50 years across different combinations of seismic source models.The single spatially smoothed point-source model yielded a maximum PGA of approximately 0.52 g,with high-value areas concentrated near historical epicenters,thereby significantly underestimating the hazard associated with major fault zones.When combined with the simple fault-source model,the maximum PGA increased to 0.8 g,with high-value zones exhibiting a zonal distribution along faults;however,the risk remained underestimated for faults with low slip rates that are nevertheless approaching their recurrence cycles.Following the introduction of the time-dependent characteristic fault-source model,local PGA values for faults in the middle-to-late stages of their recurrence cycles increased by a factor of 2 to 7 compared to the single model.These results demonstrate that the characteristic fault-source model reasonably delineates the time-dependence of large earthquake recurrence,thereby providing a more accurate assessment of imminent seismic risks.By comprehensively applying the improved spatially smoothed pointsource model,the simple fault-source model,and the characteristic fault-source model,the following faults within the region were identified as having high seismic hazard:the Huangxianggou,Zhangxian,and Tianshui segments of the Xiqinling northern edge fault;the Maqin-Maqu segment of the Dongkunlun fault;the Longriqu fault;the Maoergai fault;the Elashan fault;the Riyueshan fault;the eastern segment of the Lenglongling fault;the Maxianshan segment of the Maxianshan northern Margin fault;and the Maomaoshan-Jinqianghe segment of the Laohushan-Maomaoshan fault.As these faults are located within seismic gaps or are approaching the recurrence periods of large earthquakes,they should be prioritized for current and future seismic monitoring as well as disaster prevention and mitigation efforts.
基金supported in part by the National Natural Science Foundation of China(Key Program)under Grant No.62031021。
文摘Cascading failures pose a serious threat to the survivability of underwater unmanned swarm networks(UUSNs),significantly limiting their service ability in collaborative missions such as military reconnaissance and environmental monitoring.Existing failure models primarily focus on power grids and traffic systems,and don't address the unique challenges of weak-communication UUSNs.In UUSNs,cascading failure present a complex and dynamic process driven by the coupling of unstable acoustic channels,passive node drift,adversarial attacks,and network heterogeneity.To address these challenges,a directed weighted graph model of UUSNs is first developed,in which node positions are updated according to ocean-current-driven drift and link weights reflect the probability of successful acoustic transmission.Building on this UUSNs graph model,a cascading failure model is proposed that integrates a normal-failure-recovery state-cycle mechanism,multiple attack strategies,and routingbased load redistribution.Finally,under a five-level connectivity UUSNs scheme,simulations are conducted to analyze how dynamic topology,network load,node recovery delay,and attack modes jointly affect network survivability.The main findings are:(1)moderate node drift can improve survivability by activating weak links;(2)based-energy routing(BER)outperform based-depth routing(BDR)in harsh conditions;(3)node self-recovery time is critical to network survivability;(4)traditional degree-based critical node metrics are inadequate for weak-communication UUSNs.These results provide a theoretical foundation for designing robust survivability mechanisms in weak-communication UUSNs.
文摘To assess the effectiveness of vaccination in contaminated environments,this study introduces a modeling framework that encompasses two transmission routes,namely direct human-to-human contact and indirect human-to-environment contact,as well as the implementation of new M72/AS01_(E)vaccine.Motivated by this,a coupled age-structured tuberculosis(TB)model is proposed.Its well-posedness requirement is verified using the integrated semigroup theory.Furthermore,this study presents a comprehensive analysis of threshold dynamics associated with the proposed model.Specifically,the global stability of the disease-free and positive steady states is demonstrated by employing Lyapunov functionals.Lastly,the effects of the vaccination with M72/AS01_(E)and contaminated environments on TB control are numerically simulated.Experimental results indicate that high concentrations of Mycobacterium tuberculosis in contaminated environments may somewhat impede TB control efforts,but that large-scale deployment of new vaccine could significantly reduce the prevalence of TB.
基金Under the auspices of the National Natural Science Foundation of China(No.42371222,41971167)Fundamental Scientific Research Funds of Central China Normal University(No.CCNU24ZZ120)。
文摘Owing to intensified globalization and informatization,the structures of the urban scale hierarchy and urban networks between cities have become increasingly intertwined,resulting in different spatial effects.Therefore,this paper analyzes the spatial interaction between urban scale hierarchy and urban networks in China from 2019 to 2023,drawing on Baidu migration data and employing a spatial simultaneous equation model.The results reveal a significant positive spatial correlation between cities with higher hierarchy and those with greater network centrality.Within a static framework,we identify a positive interaction between urban scale hierarchy and urban network centrality,while their spatial cross-effects manifest as negative neighborhood interactions based on geographical distance and positive cross-scale interactions shaped by network connections.Within a dynamic framework,changes in urban scale hierarchy and urban networks are mutually reinforcing,thereby widening disparities within the urban hierarchy.Furthermore,an increase in a city’s network centrality had a dampening effect on the population growth of neighboring cities and network-connected cities.This study enhances understanding of the spatial organisation of urban systems and offers insights for coordinated regional development.
基金supported by the National Natural Science Foundation of China(Grant Nos.12272411 and 42007259).
文摘Deep rock engineering is affected by coupled thermo-hydro-mechanical(THM)-dynamic fields,necessitating the elucidation of the dynamic mechanical behavior and failure mechanisms.This study utilized a Multi-field Coupled Controlled Split Hopkinson Pressure Bar(MCC-SHPB)system to elucidate the cross-scale dynamic responses of rocks and the boundaries of failure modes under THM coupling.Impact tests were conducted on green sandstone under coupled conditions of temperature(25℃-80℃),confining pressure(0-15 MPa),and seepage water pressure(0-15 MPa).Scanning electron microscopy(SEM)microstructural characterization and COMSOL Multiphysics numerical simulations were conducted,and a dynamic constitutive theoretical framework and failure-prediction methodology were established.We investigated the impact toughness index(I_(t)),dynamic modulus(E_(d)),dynamic triaxial compressive strength(TCS_(d)),fragmentation degree(W),and failure modes of green sandstone under thermo-confining pressure-seepage-impact loading conditions.The key findings reveal that the(I_(t))reflects different energy regulation mechanisms across different confining pressure regimes.Thermal-microcrack interactions dominate at low pressure,and energy absorption prevails at high pressure.A triphasic dynamic modulus model captures stiffness evolution under energy-driven conditions,revealing cross-scale crack nucleation-propagation and fragment reorganization.The TCSd inflection point signifies energy dissipation shifts,causing nonlinear skeleton bearing-capacity degradation.A critical criterion based on the W was established to distinguish between the two failure modes and predict the unstable failure initiation.Numerical simulations were used to elucidate the effects of inertia-dominated crack propagation and stress wave interference,validating the critical criterion and the predictive accuracy of the theoretical model during cross-scale failure.This study provides a theoretical foundation for assessing the dynamic stability of rock masses subjected to multi-field coupling during deep resource exploitation.
基金Heilongjiang Provincial Natural Science Foundation of China (LH2021F009)。
文摘Anti-jamming performance evaluation has recently received significant attention. For Link-16, the anti-jamming performance evaluation and selection of the optimal anti-jamming technologies are urgent problems to be solved. A comprehensive evaluation method is proposed, which combines grey relational analysis (GRA) and cloud model, to evaluate the anti-jamming performances of Link-16. Firstly, on the basis of establishing the anti-jamming performance evaluation indicator system of Link-16, the linear combination of analytic hierarchy process(AHP) and entropy weight method (EWM) are used to calculate the combined weight. Secondly, the qualitative and quantitative concept transformation model, i.e., the cloud model, is introduced to evaluate the anti-jamming abilities of Link-16 under each jamming scheme. In addition, GRA calculates the correlation degree between evaluation indicators and the anti-jamming performance of Link-16, and assesses the best anti-jamming technology. Finally, simulation results prove that the proposed evaluation model can achieve the objective of feasible and practical evaluation, which opens up a novel way for the research of anti-jamming performance evaluations of Link-16.
基金the financial support from the Fujian Science Foundation for Outstanding Youth(2023J06039)the National Natural Science Foundation of China(Grant No.41977259,U2005205,41972268)the Independent Research Project of Technology Innovation Center for Monitoring and Restoration Engineering of Ecological Fragile Zone in Southeast China(KY-090000-04-2022-019)。
文摘Shotcrete is one of the common solutions for shallow sliding.It works by forming a protective layer with high strength and cementing the loose soil particles on the slope surface to prevent shallow sliding.However,the solidification time of conventional cement paste is long when shotcrete is used to treat cohesionless soil landslide.The idea of reinforcing slope with polyurethane solidified soil(i.e.,mixture of polyurethane and sand)was proposed.Model tests and finite element analysis were carried out to study the effectiveness of the proposed new method on the emergency treatment of cohesionless soil landslide.Surcharge loading on the crest of the slope was applied step by step until landslide was triggered so as to test and compare the stability and bearing capacity of slope models with different conditions.The simulated slope displacements were relatively close to the measured results,and the simulated slope deformation characteristics were in good agreement with the observed phenomena,which verifies the accuracy of the numerical method.Under the condition of surcharge loading on the crest of the slope,the unreinforced slope slid when the surcharge loading exceeded 30 k Pa,which presented a failure mode of local instability and collapse at the shallow layer of slope top.The reinforced slope remained stable even when the surcharge loading reached 48 k Pa.The displacement of the reinforced slope was reduced by more than 95%.Overall,this study verifies the effectiveness of polyurethane in the emergency treatment of cohesionless soil landslide and should have broad application prospects in the field of geological disasters concerning the safety of people's live.
基金supported by the National Key R&D Program of China(No.2021YFB0301200)National Natural Science Foundation of China(No.62025208).
文摘Large-scale Language Models(LLMs)have achieved significant breakthroughs in Natural Language Processing(NLP),driven by the pre-training and fine-tuning paradigm.While this approach allows models to specialize in specific tasks with reduced training costs,the substantial memory requirements during fine-tuning present a barrier to broader deployment.Parameter-Efficient Fine-Tuning(PEFT)techniques,such as Low-Rank Adaptation(LoRA),and parameter quantization methods have emerged as solutions to address these challenges by optimizing memory usage and computational efficiency.Among these,QLoRA,which combines PEFT and quantization,has demonstrated notable success in reducing memory footprints during fine-tuning,prompting the development of various QLoRA variants.Despite these advancements,the quantitative impact of key variables on the fine-tuning performance of quantized LLMs remains underexplored.This study presents a comprehensive analysis of these key variables,focusing on their influence across different layer types and depths within LLM architectures.Our investigation uncovers several critical findings:(1)Larger layers,such as MLP layers,can maintain performance despite reductions in adapter rank,while smaller layers,like self-attention layers,aremore sensitive to such changes;(2)The effectiveness of balancing factors depends more on specific values rather than layer type or depth;(3)In quantization-aware fine-tuning,larger layers can effectively utilize smaller adapters,whereas smaller layers struggle to do so.These insights suggest that layer type is a more significant determinant of fine-tuning success than layer depth when optimizing quantized LLMs.Moreover,for the same discount of trainable parameters,reducing the trainable parameters in a larger layer is more effective in preserving fine-tuning accuracy than in a smaller one.This study provides valuable guidance for more efficient fine-tuning strategies and opens avenues for further research into optimizing LLM fine-tuning in resource-constrained environments.
基金funded through India Meteorological Department,New Delhi,India under the Forecasting Agricultural output using Space,Agrometeorol ogy and Land based observations(FASAL)project and fund number:No.ASC/FASAL/KT-11/01/HQ-2010.
文摘Background Cotton is one of the most important commercial crops after food crops,especially in countries like India,where it’s grown extensively under rainfed conditions.Because of its usage in multiple industries,such as textile,medicine,and automobile industries,it has greater commercial importance.The crop’s performance is greatly influenced by prevailing weather dynamics.As climate changes,assessing how weather changes affect crop performance is essential.Among various techniques that are available,crop models are the most effective and widely used tools for predicting yields.Results This study compares statistical and machine learning models to assess their ability to predict cotton yield across major producing districts of Karnataka,India,utilizing a long-term dataset spanning from 1990 to 2023 that includes yield and weather factors.The artificial neural networks(ANNs)performed superiorly with acceptable yield deviations ranging within±10%during both vegetative stage(F1)and mid stage(F2)for cotton.The model evaluation metrics such as root mean square error(RMSE),normalized root mean square error(nRMSE),and modelling efficiency(EF)were also within the acceptance limits in most districts.Furthermore,the tested ANN model was used to assess the importance of the dominant weather factors influencing crop yield in each district.Specifically,the use of morning relative humidity as an individual parameter and its interaction with maximum and minimum tempera-ture had a major influence on cotton yield in most of the yield predicted districts.These differences highlighted the differential interactions of weather factors in each district for cotton yield formation,highlighting individual response of each weather factor under different soils and management conditions over the major cotton growing districts of Karnataka.Conclusions Compared with statistical models,machine learning models such as ANNs proved higher efficiency in forecasting the cotton yield due to their ability to consider the interactive effects of weather factors on yield forma-tion at different growth stages.This highlights the best suitability of ANNs for yield forecasting in rainfed conditions and for the study on relative impacts of weather factors on yield.Thus,the study aims to provide valuable insights to support stakeholders in planning effective crop management strategies and formulating relevant policies.
文摘DNA microarray technology is an extremely effective technique for studying gene expression patterns in cells, and the main challenge currently faced by this technology is how to analyze the large amount of gene expression data generated. To address this, this paper employs a mixed-effects model to analyze gene expression data. In terms of data selection, 1176 genes from the white mouse gene expression dataset under two experimental conditions were chosen, setting up two conditions: pneumococcal infection and no infection, and constructing a mixed-effects model. After preprocessing the gene chip information, the data were imported into the model, preliminary results were calculated, and permutation tests were performed to biologically validate the preliminary results using GSEA. The final dataset consists of 20 groups of gene expression data from pneumococcal infection, which categorizes functionally related genes based on the similarity of their expression profiles, facilitating the study of genes with unknown functions.