In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic h...In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic heterogeneous infrastructures,unstable links and non-uniform hardware capabilities create critical issues regarding security and privacy.Traditional protocols are often too computationally heavy to allow 6G services to achieve their expected Quality-of-Service(QoS).As the transport network is built of ad hoc nodes,there is no guarantee about their trustworthiness or behavior,and transversal functionalities are delegated to the extreme nodes.However,while security can be guaranteed in extreme-to-extreme solutions,privacy cannot,as all intermediate nodes still have to handle the data packets they are transporting.Besides,traditional schemes for private anonymous ad hoc communications are vulnerable against modern intelligent attacks based on learning models.The proposed scheme fulfills this gap.Findings show the probability of a successful intelligent attack reduces by up to 65%compared to ad hoc networks with no privacy protection strategy when used the proposed technology.While congestion probability can remain below 0.001%,as required in 6G services.展开更多
The excavation of deep tunnels crossing faults is highly prone to triggering rockburst disasters,which has become a significant engineering issue.In this study,taking the fault-slip rockbursts from a deep tunnel in so...The excavation of deep tunnels crossing faults is highly prone to triggering rockburst disasters,which has become a significant engineering issue.In this study,taking the fault-slip rockbursts from a deep tunnel in southwestern China as the engineering prototype,large-scale three-dimensional(3D)physical model tests were conducted on a 3D-printed complex geological model containing two faults.Based on the selfdeveloped 3D loading system and excavation device,the macroscopic failure of fault-slip rockbursts was simulated indoors.The stress,strain,and fracturing characteristics of the surrounding rock near the two faults were systematically evaluated during excavation and multistage loading.The test results effectively revealed the evolution and triggering mechanism of fault-slip rockbursts.After the excavation of a highstress tunnel,stress readjustment occurred.Owing to the presence of these two faults,stress continued to accumulate in the rock mass between them,leading to the accumulation of fractures.When the shear stress on a fault surface exceeded its shear strength,sudden fault slip and dislocation occurred,thus triggering rockbursts.Rockbursts occurred twice in the vault between the two faults,showing obvious intermittent characteristics.The rockburst pit was controlled by two faults.When the faults remained stable,tensile failure predominated in the surrounding rock.However,when the fault slip was triggered,shear failure in the surrounding rock increased.These findings provide valuable insights for enhancing the comprehension of fault-slip rockbursts.展开更多
The collection and annotation of lar ge-scale bird datasets are resource-intensive and time-consuming processes that significantly limit the scalability and accuracy of biodiversity monitoring systems.While self-super...The collection and annotation of lar ge-scale bird datasets are resource-intensive and time-consuming processes that significantly limit the scalability and accuracy of biodiversity monitoring systems.While self-supervised learning(SSL)has emerged as a promising approach for leveraging unannotated data,current SSL methods face two critical challenges in bird species recognition:(1)long-tailed data distributions that result in poor performance on underrepresented species;and(2)domain shift issues caused by data augmentation strategies designed to mitigate class imbalance.Here we present SDNet,a novel SSL-based bird recognition framework that integrates diffusion models with large language models(LLMs)to overcome these limitations.SDNet employs LLMs to generate semantically rich textual descriptions for tail-class species by prompting the models with species taxonomy,morphological attributes,and habitat information,producing detailed natural language priors that capture fine-grained visual characteristics(e.g.,plumage patterns,body proportions,and distinctive markings).These textual descriptions are subsequently used by a conditional diffusion model to synthesize new bird image samples through cross-attention mechanisms that fuse textual embeddings with intermediate visual feature representations during the denoising process,ensuring generated images preserve species-specific morphological details while maintaining photorealistic quality.Additionally,we incorporate a Swin Transformer as the feature extraction backbone whose hierarchical window-based attention mechanism and shifted windowing scheme enable multi-scale local feature extraction that proves particularly effective at capturing finegrained discriminative patterns(such as beak shape and feather texture)while mitigating domain shift between synthetic and original images through consistent feature representations across both data sources.SDNet is validated on both a self-constructed dataset(Bird_BXS)an d a publicly available benchmark(Birds_25),demonstrating substantial improvements over conventional SSL approaches.Our results indicate that the synergistic integration of LLMs,diffusion models,and the Swin Transformer architecture contributes significantly to recognition accuracy,particularly for rare and morphologically similar species.These findings highlight the potential of SDNet for addressing fundamental limitations of existing SSL methods in avian recognition tasks and establishing a new paradigm for efficient self-supervised learning in large-scale ornithological vision applications.展开更多
BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suita...BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suitable for rapid clinical application.METHODS:In this multi-center retrospective cohort study,AAS patient data from three hospitals were analyzed.The modeling cohort included data from the First Affiliated Hospital of Zhengzhou University and the People’s Hospital of Xinjiang Uygur Autonomous Region,with Peking University Third Hospital data serving as the external test set.Four machine learning algorithms—logistic regression(LR),multilayer perceptron(MLP),Gaussian naive Bayes(GNB),and random forest(RF)—were used to develop predictive models based on 34 early-accessible clinical variables.A simplifi ed model was then derived based on fi ve key variables(Stanford type,pericardial eff usion,asymmetric peripheral arterial pulsation,decreased bowel sounds,and dyspnea)via Least Absolute Shrinkage and Selection Operator(LASSO)regression to improve ED applicability.RESULTS:A total of 929 patients were included in the modeling cohort,and 210 were included in the external test set.Four machine learning models based on 34 clinical variables were developed,achieving internal and external validation AUCs of 0.85-0.90 and 0.73-0.85,respectively.The simplifi ed model incorporating fi ve key variables demonstrated internal and external validation AUCs of 0.71-0.86 and 0.75-0.78,respectively.Both models showed robust calibration and predictive stability across datasets.CONCLUSION:Both kinds of models were built based on machine learning tools,and proved to have certain prediction performance and extrapolation.展开更多
Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting...Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.展开更多
Early and accurate detection of bone cancer and marrow cell abnormalities is critical for timely intervention and improved patient outcomes.This paper proposes a novel hybrid deep learning framework that integrates a ...Early and accurate detection of bone cancer and marrow cell abnormalities is critical for timely intervention and improved patient outcomes.This paper proposes a novel hybrid deep learning framework that integrates a Convolutional Neural Network(CNN)with a Bidirectional Long Short-Term Memory(BiLSTM)architecture,optimized using the Firefly Optimization algorithm(FO).The proposed CNN-BiLSTM-FO model is tailored for structured biomedical data,capturing both local patterns and sequential dependencies in diagnostic features,while the Firefly Algorithm fine-tunes key hyperparameters to maximize predictive performance.The approach is evaluated on two benchmark biomedical datasets:one comprising diagnostic data for bone cancer detection and another for identifying marrow cell abnormalities.Experimental results demonstrate that the proposed method outperforms standard deep learning models,including CNN,LSTM,BiLSTM,and CNN-LSTM hybrids,significantly.The CNNBiLSTM-FO model achieves an accuracy of 98.55%for bone cancer detection and 96.04%for marrow abnormality classification.The paper also presents a detailed complexity analysis of the proposed algorithm and compares its performance across multiple evaluation metrics such as precision,recall,F1-score,and AUC.The results confirm the effectiveness of the firefly-based optimization strategy in improving classification accuracy and model robustness.This work introduces a scalable and accurate diagnostic solution that holds strong potential for integration into intelligent clinical decision-support systems.展开更多
The research on optimization methods for constellation launch deployment strategies focused on the consideration of mission interval time constraints at the launch site.Firstly,a dynamic modeling of the constellation ...The research on optimization methods for constellation launch deployment strategies focused on the consideration of mission interval time constraints at the launch site.Firstly,a dynamic modeling of the constellation deployment process was established,and the relationship between the deployment window and the phase difference of the orbit insertion point,as well as the cost of phase adjustment after orbit insertion,was derived.Then,the combination of the constellation deployment position sequence was treated as a parameter,together with the sequence of satellite deployment intervals,as optimization variables,simplifying a highdimensional search problem within a wide range of dates to a finite-dimensional integer programming problem.An improved genetic algorithm with local search on deployment dates was introduced to optimize the launch deployment strategy.With the new description of the optimization variables,the total number of elements in the solution space was reduced by N orders of magnitude.Numerical simulation confirms that the proposed optimization method accelerates the convergence speed from hours to minutes.展开更多
Parameter extraction of photovoltaic(PV)models is crucial for the planning,optimization,and control of PV systems.Although some methods using meta-heuristic algorithms have been proposed to determine these parameters,...Parameter extraction of photovoltaic(PV)models is crucial for the planning,optimization,and control of PV systems.Although some methods using meta-heuristic algorithms have been proposed to determine these parameters,the robustness of solutions obtained by these methods faces great challenges when the complexity of the PV model increases.The unstable results will affect the reliable operation and maintenance strategies of PV systems.In response to this challenge,an improved rime optimization algorithm with enhanced exploration and exploitation,termed TERIME,is proposed for robust and accurate parameter identification for various PV models.Specifically,the differential evolution mutation operator is integrated in the exploration phase to enhance the population diversity.Meanwhile,a new exploitation strategy incorporating randomization and neighborhood strategies simultaneously is developed to maintain the balance of exploitation width and depth.The TERIME algorithm is applied to estimate the optimal parameters of the single diode model,double diode model,and triple diode model combined with the Lambert-W function for three PV cell and module types including RTC France,Photo Watt-PWP 201 and S75.According to the statistical analysis in 100 runs,the proposed algorithm achieves more accurate and robust parameter estimations than other techniques to various PV models in varying environmental conditions.All of our source codes are publicly available at https://github.com/dirge1/TERIME.展开更多
The estimation of the probability of informed trading(PIN)model and its extensions poses significant challenges owing to various computational problems.To address these issues,we propose a novel estimation method call...The estimation of the probability of informed trading(PIN)model and its extensions poses significant challenges owing to various computational problems.To address these issues,we propose a novel estimation method called the expectation-conditional-maximization(ECM)algorithm,which can serve as an alternative to the existing methods for estimating PIN models.Our method provides optimal estimates for the original PIN model as well as two of its extensions:the multilayer PIN model and the adjusted PIN model,along with its restricted versions.Our results indicate that estimations using the ECM algorithm are generally faster,more accurate,and more memory-efficient than the standard methods used in the literature,making it a robust alternative.More importantly,the ECM algorithm is not limited to the models discussed and can be easily adapted to estimate future extensions of the PIN model.展开更多
When dealing with expensive multiobjective optimization problems,majority of existing surrogate-assisted evolutionary algorithms(SAEAs)generate solutions in decision space and screen candidate solutions mostly by usin...When dealing with expensive multiobjective optimization problems,majority of existing surrogate-assisted evolutionary algorithms(SAEAs)generate solutions in decision space and screen candidate solutions mostly by using designed surrogate models.The generated solutions exhibit excessive randomness,which tends to reduce the likelihood of generating good-quality solutions and cause a long evolution to the optima.To improve SAEAs greatly,this work proposes an evolutionary algorithm based on surrogate and inverse surrogate models by 1)Employing a surrogate model in lieu of expensive(true)function evaluations;and 2)Proposing and using an inverse surrogate model to generate new solutions.By using the same training data but with its inputs and outputs being reversed,the latter is simple to train.It is then used to generate new vectors in objective space,which are mapped into decision space to obtain their corresponding solutions.Using a particular example,this work shows its advantages over existing SAEAs.The results of comparing it with state-of-the-art algorithms on expensive optimization problems show that it is highly competitive in both solution performance and efficiency.展开更多
“Algorithm Design and Analysis”is not only one of the important courses in the undergraduate teaching of computer science and technology but also a key part of computer professional skills.In recent years,with the r...“Algorithm Design and Analysis”is not only one of the important courses in the undergraduate teaching of computer science and technology but also a key part of computer professional skills.In recent years,with the rise and widespread application of big language models,many teaching reform plans have been produced to promote the quality and efficiency of teaching.This paper studies how to refer to software development professional skills standards,investigates the knowledge points of“Algorithm Design and Analysis”courses in other educational institutions,uses cutting-edge core technology big language models to drive the improvement of teaching evaluation methods,improves teaching efficiency,and carries out reforms and practices in teaching content for undergraduate students in computer science.展开更多
Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection me...Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection methods-rooted in statistical heuristics,feature engineering,and shallow machine learning-struggle to adapt to the increasing sophistication,linguistic mimicry,and adversarial variability of DGA variants.The emergence of Large Language Models(LLMs)marks a transformative shift in this landscape.Leveraging deep contextual understanding,semantic generalization,and few-shot learning capabilities,LLMs such as BERT,GPT,and T5 have shown promising results in detecting both character-based and dictionary-based DGAs,including previously unseen(zeroday)variants.This paper provides a comprehensive and critical review of LLM-driven DGA detection,introducing a structured taxonomy of LLM architectures,evaluating the linguistic and behavioral properties of benchmark datasets,and comparing recent detection frameworks across accuracy,latency,robustness,and multilingual performance.We also highlight key limitations,including challenges in adversarial resilience,model interpretability,deployment scalability,and privacy risks.To address these gaps,we present a forward-looking research roadmap encompassing adversarial training,model compression,cross-lingual benchmarking,and real-time integration with SIEM/SOAR platforms.This survey aims to serve as a foundational resource for advancing the development of scalable,explainable,and operationally viable LLM-based DGA detection systems.展开更多
In this paper, a mathematical model consisting of forward and backward models is built on parallel genetic algorithms (PGAs) for fault diagnosis in a transmission power system. A new method to reduce the scale of faul...In this paper, a mathematical model consisting of forward and backward models is built on parallel genetic algorithms (PGAs) for fault diagnosis in a transmission power system. A new method to reduce the scale of fault sections is developed in the forward model and the message passing interface (MPI) approach is chosen to parallel the genetic algorithms by global sin-gle-population master-slave method (GPGAs). The proposed approach is applied to a sample system consisting of 28 sections, 84 protective relays and 40 circuit breakers. Simulation results show that the new model based on GPGAs can achieve very fast computation in online applications of large-scale power systems.展开更多
Blasting is well-known as an effective method for fragmenting or moving rock in open-pit mines.To evaluate the quality of blasting,the size of rock distribution is used as a critical criterion in blasting operations.A...Blasting is well-known as an effective method for fragmenting or moving rock in open-pit mines.To evaluate the quality of blasting,the size of rock distribution is used as a critical criterion in blasting operations.A high percentage of oversized rocks generated by blasting operations can lead to economic and environmental damage.Therefore,this study proposed four novel intelligent models to predict the size of rock distribution in mine blasting in order to optimize blasting parameters,as well as the efficiency of blasting operation in open mines.Accordingly,a nature-inspired algorithm(i.e.,firefly algorithm-FFA)and different machine learning algorithms(i.e.,gradient boosting machine(GBM),support vector machine(SVM),Gaussian process(GP),and artificial neural network(ANN))were combined for this aim,abbreviated as FFA-GBM,FFA-SVM,FFA-GP,and FFA-ANN,respectively.Subsequently,predicted results from the abovementioned models were compared with each other using three statistical indicators(e.g.,mean absolute error,root-mean-squared error,and correlation coefficient)and color intensity method.For developing and simulating the size of rock in blasting operations,136 blasting events with their images were collected and analyzed by the Split-Desktop software.In which,111 events were randomly selected for the development and optimization of the models.Subsequently,the remaining 25 blasting events were applied to confirm the accuracy of the proposed models.Herein,blast design parameters were regarded as input variables to predict the size of rock in blasting operations.Finally,the obtained results revealed that the FFA is a robust optimization algorithm for estimating rock fragmentation in bench blasting.Among the models developed in this study,FFA-GBM provided the highest accuracy in predicting the size of fragmented rocks.The other techniques(i.e.,FFA-SVM,FFA-GP,and FFA-ANN)yielded lower computational stability and efficiency.Hence,the FFA-GBM model can be used as a powerful and precise soft computing tool that can be applied to practical engineering cases aiming to improve the quality of blasting and rock fragmentation.展开更多
Many biodynamic models have been derived using trial and error curve-fitting technique, such that the error between the computed and measured biodynamic response functions is minimum. This study developed a biomechani...Many biodynamic models have been derived using trial and error curve-fitting technique, such that the error between the computed and measured biodynamic response functions is minimum. This study developed a biomechanical model of the human body in a sitting posture without backrest for evaluating the vibration transmissibility and dynamic response to vertical vibration direction. In describing the human body motion, a three biomechanical models are discussed (two models are 4-DOF and one model 7-DOF). Optimization software based on stochastic techniques search methods, Genetic Algorithms (GAs), is employed to determine the human model parameters imposing some limit constraints on the model parameters. In addition, an objective function is formulated comprising the sum of errors between the computed and actual values (experimental data). The studied functions are the driving-point mechanical impedance, apparent mass and seat- to-head transmissibility functions. The optimization process increased the average goodness of fit and the results of studied functions became much closer to the target values (Experimental data). From the optimized model, the resonant frequencies of the driver parts computed on the basis of biodynamic response functions are found to be within close bounds to that expected for the human body.展开更多
Direct soil temperature(ST)measurement is time-consuming and costly;thus,the use of simple and cost-effective machine learning(ML)tools is helpful.In this study,ML approaches,including KStar,instance-based K-nearest l...Direct soil temperature(ST)measurement is time-consuming and costly;thus,the use of simple and cost-effective machine learning(ML)tools is helpful.In this study,ML approaches,including KStar,instance-based K-nearest learning(IBK),and locally weighted learning(LWL),coupled with resampling algorithms of bagging(BA)and dagging(DA)(BA-IBK,BA-KStar,BA-LWL,DA-IBK,DA-KStar,and DA-LWL)were developed and tested for multi-step ahead(3,6,and 9 d ahead)ST forecasting.In addition,a linear regression(LR)model was used as a benchmark to evaluate the results.A dataset was established,with daily ST time-series at 5 and 50 cm soil depths in a farmland as models’output and meteorological data as models’input,including mean(T_(mean)),minimum(Tmin),and maximum(T_(max))air temperatures,evaporation(Eva),sunshine hours(SSH),and solar radiation(SR),which were collected at Isfahan Synoptic Station(Iran)for 13 years(1992–2005).Six different input combination scenarios were selected based on Pearson’s correlation coefficients between inputs and outputs and fed into the models.We used 70%of the data to train the models,with the remaining 30%used for model evaluation via multiple visual and quantitative metrics.Our?ndings showed that T_(mean)was the most effective input variable for ST forecasting in most of the developed models,while in some cases the combinations of variables,including T_(mean)and T_(max)and T_(mean),T_(max),Tmin,Eva,and SSH proved to be the best input combinations.Among the evaluated models,BA-KStar showed greater compatibility,while in most cases,BA-IBK and-LWL provided more accurate results,depending on soil depth.For the 5 cm soil depth,BA-KStar had superior performance(i.e.,Nash-Sutcliffe efficiency(NSE)=0.90,0.87,and 0.85 for 3,6,and 9 d ahead forecasting,respectively);for the 50 cm soil depth,DA-KStar outperformed the other models(i.e.,NSE=0.88,0.89,and 0.89 for 3,6,and 9 d ahead forecasting,respectively).The results con?rmed that all hybrid models had higher prediction capabilities than the LR model.展开更多
Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to tr...Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to traverse vast expanse with limited computational resources.Furthermore,in the context of sparse,most variables in Pareto optimal solutions are zero,making it difficult for algorithms to identify non-zero variables efficiently.This paper is dedicated to addressing the challenges posed by SLMOPs.To start,we introduce innovative objective functions customized to mine maximum and minimum candidate sets.This substantial enhancement dramatically improves the efficacy of frequent pattern mining.In this way,selecting candidate sets is no longer based on the quantity of nonzero variables they contain but on a higher proportion of nonzero variables within specific dimensions.Additionally,we unveil a novel approach to association rule mining,which delves into the intricate relationships between non-zero variables.This novel methodology aids in identifying sparse distributions that can potentially expedite reductions in the objective function value.We extensively tested our algorithm across eight benchmark problems and four real-world SLMOPs.The results demonstrate that our approach achieves competitive solutions across various challenges.展开更多
This study delineates the development of the optimization framework for the preliminary design phase of Floating Offshore Wind Turbines(FOWTs),and the central challenge addressed is the optimization of the FOWT platfo...This study delineates the development of the optimization framework for the preliminary design phase of Floating Offshore Wind Turbines(FOWTs),and the central challenge addressed is the optimization of the FOWT platform dimensional parameters in relation to motion responses.Although the three-dimensional potential flow(TDPF)panel method is recognized for its precision in calculating FOWT motion responses,its computational intensity necessitates an alternative approach for efficiency.Herein,a novel application of varying fidelity frequency-domain computational strategies is introduced,which synthesizes the strip theory with the TDPF panel method to strike a balance between computational speed and accuracy.The Co-Kriging algorithm is employed to forge a surrogate model that amalgamates these computational strategies.Optimization objectives are centered on the platform’s motion response in heave and pitch directions under general sea conditions.The steel usage,the range of design variables,and geometric considerations are optimization constraints.The angle of the pontoons,the number of columns,the radius of the central column and the parameters of the mooring lines are optimization constants.This informed the structuring of a multi-objective optimization model utilizing the Non-dominated Sorting Genetic Algorithm Ⅱ(NSGA-Ⅱ)algorithm.For the case of the IEA UMaine VolturnUS-S Reference Platform,Pareto fronts are discerned based on the above framework and delineate the relationship between competing motion response objectives.The efficacy of final designs is substantiated through the time-domain calculation model,which ensures that the motion responses in extreme sea conditions are superior to those of the initial design.展开更多
This paper describes a novel algorithm for fragile watermarking of 3D models. Fragile watermarking requires detection of even minute intentional changes to the 3D model along with the location of the change. This pose...This paper describes a novel algorithm for fragile watermarking of 3D models. Fragile watermarking requires detection of even minute intentional changes to the 3D model along with the location of the change. This poses a challenge since inserting random amount of watermark in all the vertices of the model would generally introduce perceptible distortion. The proposed algorithm overcomes this challenge by using genetic algorithm to modify every vertex location in the model so that there is no perceptible distortion. Various experimental results are used to justify the choice of the genetic algorithm design parameters. Experimental results also indicate that the proposed algorithm can accurately detect location of any mesh modification.展开更多
Analyzing rock mass seepage using the discrete fracture network(DFN)flow model poses challenges when dealing with complex fracture networks.This paper presents a novel DFN flow model that incorporates the actual conne...Analyzing rock mass seepage using the discrete fracture network(DFN)flow model poses challenges when dealing with complex fracture networks.This paper presents a novel DFN flow model that incorporates the actual connections of large-scale fractures.Notably,this model efficiently manages over 20,000 fractures without necessitating adjustments to the DFN geometry.All geometric analyses,such as identifying connected fractures,dividing the two-dimensional domain into closed loops,triangulating arbitrary loops,and refining triangular elements,are fully automated.The analysis processes are comprehensively introduced,and core algorithms,along with their pseudo-codes,are outlined and explained to assist readers in their programming endeavors.The accuracy of geometric analyses is validated through topological graphs representing the connection relationships between fractures.In practical application,the proposed model is employed to assess the water-sealing effectiveness of an underground storage cavern project.The analysis results indicate that the existing design scheme can effectively prevent the stored oil from leaking in the presence of both dense and sparse fractures.Furthermore,following extensive modification and optimization,the scale and precision of model computation suggest that the proposed model and developed codes can meet the requirements of engineering applications.展开更多
基金funding from the European Commission by the Ruralities project(grant agreement no.101060876).
文摘In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic heterogeneous infrastructures,unstable links and non-uniform hardware capabilities create critical issues regarding security and privacy.Traditional protocols are often too computationally heavy to allow 6G services to achieve their expected Quality-of-Service(QoS).As the transport network is built of ad hoc nodes,there is no guarantee about their trustworthiness or behavior,and transversal functionalities are delegated to the extreme nodes.However,while security can be guaranteed in extreme-to-extreme solutions,privacy cannot,as all intermediate nodes still have to handle the data packets they are transporting.Besides,traditional schemes for private anonymous ad hoc communications are vulnerable against modern intelligent attacks based on learning models.The proposed scheme fulfills this gap.Findings show the probability of a successful intelligent attack reduces by up to 65%compared to ad hoc networks with no privacy protection strategy when used the proposed technology.While congestion probability can remain below 0.001%,as required in 6G services.
基金funding support from the National Natural Science Foundation of China(Grant Nos.42177136 and 52309126).
文摘The excavation of deep tunnels crossing faults is highly prone to triggering rockburst disasters,which has become a significant engineering issue.In this study,taking the fault-slip rockbursts from a deep tunnel in southwestern China as the engineering prototype,large-scale three-dimensional(3D)physical model tests were conducted on a 3D-printed complex geological model containing two faults.Based on the selfdeveloped 3D loading system and excavation device,the macroscopic failure of fault-slip rockbursts was simulated indoors.The stress,strain,and fracturing characteristics of the surrounding rock near the two faults were systematically evaluated during excavation and multistage loading.The test results effectively revealed the evolution and triggering mechanism of fault-slip rockbursts.After the excavation of a highstress tunnel,stress readjustment occurred.Owing to the presence of these two faults,stress continued to accumulate in the rock mass between them,leading to the accumulation of fractures.When the shear stress on a fault surface exceeded its shear strength,sudden fault slip and dislocation occurred,thus triggering rockbursts.Rockbursts occurred twice in the vault between the two faults,showing obvious intermittent characteristics.The rockburst pit was controlled by two faults.When the faults remained stable,tensile failure predominated in the surrounding rock.However,when the fault slip was triggered,shear failure in the surrounding rock increased.These findings provide valuable insights for enhancing the comprehension of fault-slip rockbursts.
基金supported by the National Natural Science Foundation of China(32471964)。
文摘The collection and annotation of lar ge-scale bird datasets are resource-intensive and time-consuming processes that significantly limit the scalability and accuracy of biodiversity monitoring systems.While self-supervised learning(SSL)has emerged as a promising approach for leveraging unannotated data,current SSL methods face two critical challenges in bird species recognition:(1)long-tailed data distributions that result in poor performance on underrepresented species;and(2)domain shift issues caused by data augmentation strategies designed to mitigate class imbalance.Here we present SDNet,a novel SSL-based bird recognition framework that integrates diffusion models with large language models(LLMs)to overcome these limitations.SDNet employs LLMs to generate semantically rich textual descriptions for tail-class species by prompting the models with species taxonomy,morphological attributes,and habitat information,producing detailed natural language priors that capture fine-grained visual characteristics(e.g.,plumage patterns,body proportions,and distinctive markings).These textual descriptions are subsequently used by a conditional diffusion model to synthesize new bird image samples through cross-attention mechanisms that fuse textual embeddings with intermediate visual feature representations during the denoising process,ensuring generated images preserve species-specific morphological details while maintaining photorealistic quality.Additionally,we incorporate a Swin Transformer as the feature extraction backbone whose hierarchical window-based attention mechanism and shifted windowing scheme enable multi-scale local feature extraction that proves particularly effective at capturing finegrained discriminative patterns(such as beak shape and feather texture)while mitigating domain shift between synthetic and original images through consistent feature representations across both data sources.SDNet is validated on both a self-constructed dataset(Bird_BXS)an d a publicly available benchmark(Birds_25),demonstrating substantial improvements over conventional SSL approaches.Our results indicate that the synergistic integration of LLMs,diffusion models,and the Swin Transformer architecture contributes significantly to recognition accuracy,particularly for rare and morphologically similar species.These findings highlight the potential of SDNet for addressing fundamental limitations of existing SSL methods in avian recognition tasks and establishing a new paradigm for efficient self-supervised learning in large-scale ornithological vision applications.
基金supported by the special fund of the National Clinical Key Specialty Construction Program[(2022)301-2305].
文摘BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suitable for rapid clinical application.METHODS:In this multi-center retrospective cohort study,AAS patient data from three hospitals were analyzed.The modeling cohort included data from the First Affiliated Hospital of Zhengzhou University and the People’s Hospital of Xinjiang Uygur Autonomous Region,with Peking University Third Hospital data serving as the external test set.Four machine learning algorithms—logistic regression(LR),multilayer perceptron(MLP),Gaussian naive Bayes(GNB),and random forest(RF)—were used to develop predictive models based on 34 early-accessible clinical variables.A simplifi ed model was then derived based on fi ve key variables(Stanford type,pericardial eff usion,asymmetric peripheral arterial pulsation,decreased bowel sounds,and dyspnea)via Least Absolute Shrinkage and Selection Operator(LASSO)regression to improve ED applicability.RESULTS:A total of 929 patients were included in the modeling cohort,and 210 were included in the external test set.Four machine learning models based on 34 clinical variables were developed,achieving internal and external validation AUCs of 0.85-0.90 and 0.73-0.85,respectively.The simplifi ed model incorporating fi ve key variables demonstrated internal and external validation AUCs of 0.71-0.86 and 0.75-0.78,respectively.Both models showed robust calibration and predictive stability across datasets.CONCLUSION:Both kinds of models were built based on machine learning tools,and proved to have certain prediction performance and extrapolation.
基金National Key Research and Development Program of China,No.2023YFC3006704National Natural Science Foundation of China,No.42171047CAS-CSIRO Partnership Joint Project of 2024,No.177GJHZ2023097MI。
文摘Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.
文摘Early and accurate detection of bone cancer and marrow cell abnormalities is critical for timely intervention and improved patient outcomes.This paper proposes a novel hybrid deep learning framework that integrates a Convolutional Neural Network(CNN)with a Bidirectional Long Short-Term Memory(BiLSTM)architecture,optimized using the Firefly Optimization algorithm(FO).The proposed CNN-BiLSTM-FO model is tailored for structured biomedical data,capturing both local patterns and sequential dependencies in diagnostic features,while the Firefly Algorithm fine-tunes key hyperparameters to maximize predictive performance.The approach is evaluated on two benchmark biomedical datasets:one comprising diagnostic data for bone cancer detection and another for identifying marrow cell abnormalities.Experimental results demonstrate that the proposed method outperforms standard deep learning models,including CNN,LSTM,BiLSTM,and CNN-LSTM hybrids,significantly.The CNNBiLSTM-FO model achieves an accuracy of 98.55%for bone cancer detection and 96.04%for marrow abnormality classification.The paper also presents a detailed complexity analysis of the proposed algorithm and compares its performance across multiple evaluation metrics such as precision,recall,F1-score,and AUC.The results confirm the effectiveness of the firefly-based optimization strategy in improving classification accuracy and model robustness.This work introduces a scalable and accurate diagnostic solution that holds strong potential for integration into intelligent clinical decision-support systems.
文摘The research on optimization methods for constellation launch deployment strategies focused on the consideration of mission interval time constraints at the launch site.Firstly,a dynamic modeling of the constellation deployment process was established,and the relationship between the deployment window and the phase difference of the orbit insertion point,as well as the cost of phase adjustment after orbit insertion,was derived.Then,the combination of the constellation deployment position sequence was treated as a parameter,together with the sequence of satellite deployment intervals,as optimization variables,simplifying a highdimensional search problem within a wide range of dates to a finite-dimensional integer programming problem.An improved genetic algorithm with local search on deployment dates was introduced to optimize the launch deployment strategy.With the new description of the optimization variables,the total number of elements in the solution space was reduced by N orders of magnitude.Numerical simulation confirms that the proposed optimization method accelerates the convergence speed from hours to minutes.
基金supported by the National Natural Science Foundation of China[grant number 51775020]the Science Challenge Project[grant number.TZ2018007]+2 种基金the National Natural Science Foundation of China[grant number 62073009]the Postdoctoral Fellowship Program of CPSF[grant number GZC20233365]the Fundamental Research Funds for Central Universities[grant number JKF-20240559].
文摘Parameter extraction of photovoltaic(PV)models is crucial for the planning,optimization,and control of PV systems.Although some methods using meta-heuristic algorithms have been proposed to determine these parameters,the robustness of solutions obtained by these methods faces great challenges when the complexity of the PV model increases.The unstable results will affect the reliable operation and maintenance strategies of PV systems.In response to this challenge,an improved rime optimization algorithm with enhanced exploration and exploitation,termed TERIME,is proposed for robust and accurate parameter identification for various PV models.Specifically,the differential evolution mutation operator is integrated in the exploration phase to enhance the population diversity.Meanwhile,a new exploitation strategy incorporating randomization and neighborhood strategies simultaneously is developed to maintain the balance of exploitation width and depth.The TERIME algorithm is applied to estimate the optimal parameters of the single diode model,double diode model,and triple diode model combined with the Lambert-W function for three PV cell and module types including RTC France,Photo Watt-PWP 201 and S75.According to the statistical analysis in 100 runs,the proposed algorithm achieves more accurate and robust parameter estimations than other techniques to various PV models in varying environmental conditions.All of our source codes are publicly available at https://github.com/dirge1/TERIME.
基金supported by the Scientific and Technological Research Council of Turkey(TUBITAK)[grant no 122K637].
文摘The estimation of the probability of informed trading(PIN)model and its extensions poses significant challenges owing to various computational problems.To address these issues,we propose a novel estimation method called the expectation-conditional-maximization(ECM)algorithm,which can serve as an alternative to the existing methods for estimating PIN models.Our method provides optimal estimates for the original PIN model as well as two of its extensions:the multilayer PIN model and the adjusted PIN model,along with its restricted versions.Our results indicate that estimations using the ECM algorithm are generally faster,more accurate,and more memory-efficient than the standard methods used in the literature,making it a robust alternative.More importantly,the ECM algorithm is not limited to the models discussed and can be easily adapted to estimate future extensions of the PIN model.
基金supported in part by the National Natural Science Foundation of China(51775385)the Natural Science Foundation of Shanghai(23ZR1466000)+2 种基金the Shanghai Industrial Collaborative Science and Technology Innovation Project(2021-cyxt2-kj10)the Innovation Program of Shanghai Municipal Education Commission(202101070007E00098)Fundo para o Desenvolvimento das Ciencias e da Tecnologia(FDCT)(0147/2024/AFJ).
文摘When dealing with expensive multiobjective optimization problems,majority of existing surrogate-assisted evolutionary algorithms(SAEAs)generate solutions in decision space and screen candidate solutions mostly by using designed surrogate models.The generated solutions exhibit excessive randomness,which tends to reduce the likelihood of generating good-quality solutions and cause a long evolution to the optima.To improve SAEAs greatly,this work proposes an evolutionary algorithm based on surrogate and inverse surrogate models by 1)Employing a surrogate model in lieu of expensive(true)function evaluations;and 2)Proposing and using an inverse surrogate model to generate new solutions.By using the same training data but with its inputs and outputs being reversed,the latter is simple to train.It is then used to generate new vectors in objective space,which are mapped into decision space to obtain their corresponding solutions.Using a particular example,this work shows its advantages over existing SAEAs.The results of comparing it with state-of-the-art algorithms on expensive optimization problems show that it is highly competitive in both solution performance and efficiency.
基金supported by Harbin Engineering University’s 2021 Education Reform Project“How to Make Computer Theory Teaching Serve Employment”(Grant No.JG2021B0609).
文摘“Algorithm Design and Analysis”is not only one of the important courses in the undergraduate teaching of computer science and technology but also a key part of computer professional skills.In recent years,with the rise and widespread application of big language models,many teaching reform plans have been produced to promote the quality and efficiency of teaching.This paper studies how to refer to software development professional skills standards,investigates the knowledge points of“Algorithm Design and Analysis”courses in other educational institutions,uses cutting-edge core technology big language models to drive the improvement of teaching evaluation methods,improves teaching efficiency,and carries out reforms and practices in teaching content for undergraduate students in computer science.
基金the Deanship of Scientific Research at King Khalid University for funding this work through large group under grant number(GRP.2/663/46).
文摘Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection methods-rooted in statistical heuristics,feature engineering,and shallow machine learning-struggle to adapt to the increasing sophistication,linguistic mimicry,and adversarial variability of DGA variants.The emergence of Large Language Models(LLMs)marks a transformative shift in this landscape.Leveraging deep contextual understanding,semantic generalization,and few-shot learning capabilities,LLMs such as BERT,GPT,and T5 have shown promising results in detecting both character-based and dictionary-based DGAs,including previously unseen(zeroday)variants.This paper provides a comprehensive and critical review of LLM-driven DGA detection,introducing a structured taxonomy of LLM architectures,evaluating the linguistic and behavioral properties of benchmark datasets,and comparing recent detection frameworks across accuracy,latency,robustness,and multilingual performance.We also highlight key limitations,including challenges in adversarial resilience,model interpretability,deployment scalability,and privacy risks.To address these gaps,we present a forward-looking research roadmap encompassing adversarial training,model compression,cross-lingual benchmarking,and real-time integration with SIEM/SOAR platforms.This survey aims to serve as a foundational resource for advancing the development of scalable,explainable,and operationally viable LLM-based DGA detection systems.
基金the National Natural Science Foundation of China (No. 50677062)the New Century Excellent Talents in Uni-versity of China (No. NCET-07-0745)the Natural Science Foundation of Zhejiang Province, China (No. R107062)
文摘In this paper, a mathematical model consisting of forward and backward models is built on parallel genetic algorithms (PGAs) for fault diagnosis in a transmission power system. A new method to reduce the scale of fault sections is developed in the forward model and the message passing interface (MPI) approach is chosen to parallel the genetic algorithms by global sin-gle-population master-slave method (GPGAs). The proposed approach is applied to a sample system consisting of 28 sections, 84 protective relays and 40 circuit breakers. Simulation results show that the new model based on GPGAs can achieve very fast computation in online applications of large-scale power systems.
基金supported by the Center for Mining,Electro-Mechanical research of Hanoi University of Mining and Geology(HUMG),Hanoi,Vietnamfinancially supported by the Hunan Provincial Department of Education General Project(19C1744)+1 种基金Hunan Province Science Foundation for Youth Scholars of China fund(2018JJ3510)the Innovation-Driven Project of Central South University(2020CX040)。
文摘Blasting is well-known as an effective method for fragmenting or moving rock in open-pit mines.To evaluate the quality of blasting,the size of rock distribution is used as a critical criterion in blasting operations.A high percentage of oversized rocks generated by blasting operations can lead to economic and environmental damage.Therefore,this study proposed four novel intelligent models to predict the size of rock distribution in mine blasting in order to optimize blasting parameters,as well as the efficiency of blasting operation in open mines.Accordingly,a nature-inspired algorithm(i.e.,firefly algorithm-FFA)and different machine learning algorithms(i.e.,gradient boosting machine(GBM),support vector machine(SVM),Gaussian process(GP),and artificial neural network(ANN))were combined for this aim,abbreviated as FFA-GBM,FFA-SVM,FFA-GP,and FFA-ANN,respectively.Subsequently,predicted results from the abovementioned models were compared with each other using three statistical indicators(e.g.,mean absolute error,root-mean-squared error,and correlation coefficient)and color intensity method.For developing and simulating the size of rock in blasting operations,136 blasting events with their images were collected and analyzed by the Split-Desktop software.In which,111 events were randomly selected for the development and optimization of the models.Subsequently,the remaining 25 blasting events were applied to confirm the accuracy of the proposed models.Herein,blast design parameters were regarded as input variables to predict the size of rock in blasting operations.Finally,the obtained results revealed that the FFA is a robust optimization algorithm for estimating rock fragmentation in bench blasting.Among the models developed in this study,FFA-GBM provided the highest accuracy in predicting the size of fragmented rocks.The other techniques(i.e.,FFA-SVM,FFA-GP,and FFA-ANN)yielded lower computational stability and efficiency.Hence,the FFA-GBM model can be used as a powerful and precise soft computing tool that can be applied to practical engineering cases aiming to improve the quality of blasting and rock fragmentation.
文摘Many biodynamic models have been derived using trial and error curve-fitting technique, such that the error between the computed and measured biodynamic response functions is minimum. This study developed a biomechanical model of the human body in a sitting posture without backrest for evaluating the vibration transmissibility and dynamic response to vertical vibration direction. In describing the human body motion, a three biomechanical models are discussed (two models are 4-DOF and one model 7-DOF). Optimization software based on stochastic techniques search methods, Genetic Algorithms (GAs), is employed to determine the human model parameters imposing some limit constraints on the model parameters. In addition, an objective function is formulated comprising the sum of errors between the computed and actual values (experimental data). The studied functions are the driving-point mechanical impedance, apparent mass and seat- to-head transmissibility functions. The optimization process increased the average goodness of fit and the results of studied functions became much closer to the target values (Experimental data). From the optimized model, the resonant frequencies of the driver parts computed on the basis of biodynamic response functions are found to be within close bounds to that expected for the human body.
文摘Direct soil temperature(ST)measurement is time-consuming and costly;thus,the use of simple and cost-effective machine learning(ML)tools is helpful.In this study,ML approaches,including KStar,instance-based K-nearest learning(IBK),and locally weighted learning(LWL),coupled with resampling algorithms of bagging(BA)and dagging(DA)(BA-IBK,BA-KStar,BA-LWL,DA-IBK,DA-KStar,and DA-LWL)were developed and tested for multi-step ahead(3,6,and 9 d ahead)ST forecasting.In addition,a linear regression(LR)model was used as a benchmark to evaluate the results.A dataset was established,with daily ST time-series at 5 and 50 cm soil depths in a farmland as models’output and meteorological data as models’input,including mean(T_(mean)),minimum(Tmin),and maximum(T_(max))air temperatures,evaporation(Eva),sunshine hours(SSH),and solar radiation(SR),which were collected at Isfahan Synoptic Station(Iran)for 13 years(1992–2005).Six different input combination scenarios were selected based on Pearson’s correlation coefficients between inputs and outputs and fed into the models.We used 70%of the data to train the models,with the remaining 30%used for model evaluation via multiple visual and quantitative metrics.Our?ndings showed that T_(mean)was the most effective input variable for ST forecasting in most of the developed models,while in some cases the combinations of variables,including T_(mean)and T_(max)and T_(mean),T_(max),Tmin,Eva,and SSH proved to be the best input combinations.Among the evaluated models,BA-KStar showed greater compatibility,while in most cases,BA-IBK and-LWL provided more accurate results,depending on soil depth.For the 5 cm soil depth,BA-KStar had superior performance(i.e.,Nash-Sutcliffe efficiency(NSE)=0.90,0.87,and 0.85 for 3,6,and 9 d ahead forecasting,respectively);for the 50 cm soil depth,DA-KStar outperformed the other models(i.e.,NSE=0.88,0.89,and 0.89 for 3,6,and 9 d ahead forecasting,respectively).The results con?rmed that all hybrid models had higher prediction capabilities than the LR model.
基金support by the Open Project of Xiangjiang Laboratory(22XJ02003)the University Fundamental Research Fund(23-ZZCX-JDZ-28,ZK21-07)+5 种基金the National Science Fund for Outstanding Young Scholars(62122093)the National Natural Science Foundation of China(72071205)the Hunan Graduate Research Innovation Project(CX20230074)the Hunan Natural Science Foundation Regional Joint Project(2023JJ50490)the Science and Technology Project for Young and Middle-aged Talents of Hunan(2023TJZ03)the Science and Technology Innovation Program of Humnan Province(2023RC1002).
文摘Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to traverse vast expanse with limited computational resources.Furthermore,in the context of sparse,most variables in Pareto optimal solutions are zero,making it difficult for algorithms to identify non-zero variables efficiently.This paper is dedicated to addressing the challenges posed by SLMOPs.To start,we introduce innovative objective functions customized to mine maximum and minimum candidate sets.This substantial enhancement dramatically improves the efficacy of frequent pattern mining.In this way,selecting candidate sets is no longer based on the quantity of nonzero variables they contain but on a higher proportion of nonzero variables within specific dimensions.Additionally,we unveil a novel approach to association rule mining,which delves into the intricate relationships between non-zero variables.This novel methodology aids in identifying sparse distributions that can potentially expedite reductions in the objective function value.We extensively tested our algorithm across eight benchmark problems and four real-world SLMOPs.The results demonstrate that our approach achieves competitive solutions across various challenges.
基金financially supported by the National Natural Science Foundation of China(Grant No.52371261)the Science and Technology Projects of Liaoning Province(Grant No.2023011352-JH1/110).
文摘This study delineates the development of the optimization framework for the preliminary design phase of Floating Offshore Wind Turbines(FOWTs),and the central challenge addressed is the optimization of the FOWT platform dimensional parameters in relation to motion responses.Although the three-dimensional potential flow(TDPF)panel method is recognized for its precision in calculating FOWT motion responses,its computational intensity necessitates an alternative approach for efficiency.Herein,a novel application of varying fidelity frequency-domain computational strategies is introduced,which synthesizes the strip theory with the TDPF panel method to strike a balance between computational speed and accuracy.The Co-Kriging algorithm is employed to forge a surrogate model that amalgamates these computational strategies.Optimization objectives are centered on the platform’s motion response in heave and pitch directions under general sea conditions.The steel usage,the range of design variables,and geometric considerations are optimization constraints.The angle of the pontoons,the number of columns,the radius of the central column and the parameters of the mooring lines are optimization constants.This informed the structuring of a multi-objective optimization model utilizing the Non-dominated Sorting Genetic Algorithm Ⅱ(NSGA-Ⅱ)algorithm.For the case of the IEA UMaine VolturnUS-S Reference Platform,Pareto fronts are discerned based on the above framework and delineate the relationship between competing motion response objectives.The efficacy of final designs is substantiated through the time-domain calculation model,which ensures that the motion responses in extreme sea conditions are superior to those of the initial design.
文摘This paper describes a novel algorithm for fragile watermarking of 3D models. Fragile watermarking requires detection of even minute intentional changes to the 3D model along with the location of the change. This poses a challenge since inserting random amount of watermark in all the vertices of the model would generally introduce perceptible distortion. The proposed algorithm overcomes this challenge by using genetic algorithm to modify every vertex location in the model so that there is no perceptible distortion. Various experimental results are used to justify the choice of the genetic algorithm design parameters. Experimental results also indicate that the proposed algorithm can accurately detect location of any mesh modification.
基金sponsored by the General Program of the National Natural Science Foundation of China(Grant Nos.52079129 and 52209148)the Hubei Provincial General Fund,China(Grant No.2023AFB567)。
文摘Analyzing rock mass seepage using the discrete fracture network(DFN)flow model poses challenges when dealing with complex fracture networks.This paper presents a novel DFN flow model that incorporates the actual connections of large-scale fractures.Notably,this model efficiently manages over 20,000 fractures without necessitating adjustments to the DFN geometry.All geometric analyses,such as identifying connected fractures,dividing the two-dimensional domain into closed loops,triangulating arbitrary loops,and refining triangular elements,are fully automated.The analysis processes are comprehensively introduced,and core algorithms,along with their pseudo-codes,are outlined and explained to assist readers in their programming endeavors.The accuracy of geometric analyses is validated through topological graphs representing the connection relationships between fractures.In practical application,the proposed model is employed to assess the water-sealing effectiveness of an underground storage cavern project.The analysis results indicate that the existing design scheme can effectively prevent the stored oil from leaking in the presence of both dense and sparse fractures.Furthermore,following extensive modification and optimization,the scale and precision of model computation suggest that the proposed model and developed codes can meet the requirements of engineering applications.