Topology optimization(TO),a numerical technique to find the optimalmaterial layoutwith a given design domain,has attracted interest from researchers in the field of structural optimization in recent years.For beginner...Topology optimization(TO),a numerical technique to find the optimalmaterial layoutwith a given design domain,has attracted interest from researchers in the field of structural optimization in recent years.For beginners,opensource codes are undoubtedly the best alternative to learning TO,which can elaborate the implementation of a method in detail and easily engage more people to employ and extend the method.In this paper,we present a summary of various open-source codes and related literature on TO methods,including solid isotropic material with penalization(SIMP),evolutionary method,level set method(LSM),moving morphable components/voids(MMC/MMV)methods,multiscale topology optimization method,etc.Simultaneously,we classify the codes into five levels,fromeasy to difficult,depending on their difficulty,so that beginners can get started and understand the form of code implementation more quickly.展开更多
An effective energy management strategy(EMS)is essential to optimize the energy efficiency of electric vehicles(EVs).With the advent of advanced machine learning techniques,the focus on developing sophisticated EMS fo...An effective energy management strategy(EMS)is essential to optimize the energy efficiency of electric vehicles(EVs).With the advent of advanced machine learning techniques,the focus on developing sophisticated EMS for EVs is increasing.Here,we introduce LearningEMS:a unified framework and open-source benchmark designed to facilitate rapid development and assessment of EMS.LearningEMS is distinguished by its ability to support a variety of EV configurations,including hybrid EVs,fuel cell EVs,and plug-in EVs,offering a general platform for the development of EMS.The framework enables detailed comparisons of several EMS algorithms,encompassing imitation learning,deep reinforcement learning(RL),offline RL,model predictive control,and dynamic programming.We rigorously evaluated these algorithms across multiple perspectives:energy efficiency,consistency,adaptability,and practicability.Furthermore,we discuss state,reward,and action settings for RL in EV energy management,introduce a policy extraction and reconstruction method for learning-based EMS deployment,and conduct hardware-in-the-loop experiments.In summary,we offer a unified and comprehensive framework that comes with three distinct EV platforms,over 10000 km of EMS policy data set,ten state-of-the-art algorithms,and over 160 benchmark tasks,along with three learning libraries.Its flexible design allows easy expansion for additional tasks and applications.The open-source algorithms,models,data sets,and deployment processes foster additional research and innovation in EV and broader engineering domains.展开更多
介绍了高温蠕变工况下运行的压力容器可能出现的失效模式,结合工程设计现状,指出了我国当前压力容器标准体系在确定高温蠕变工况许用压应力时存在的技术瓶颈,在此基础之上引出ASME Code Case 3029,对其适用范围、发展历程、产生背景及...介绍了高温蠕变工况下运行的压力容器可能出现的失效模式,结合工程设计现状,指出了我国当前压力容器标准体系在确定高温蠕变工况许用压应力时存在的技术瓶颈,在此基础之上引出ASME Code Case 3029,对其适用范围、发展历程、产生背景及工程意义进行了简单的介绍,以某工程设计项目中的实际结构为例,介绍了该方法的使用过程及注意事项,并结合压力容器工程设计领域的实际需求,对我国标准体系下一步的制定或修订方向提出了展望。展开更多
Mobile communications are reaching out to every aspect of our daily life,necessitating highefficiency data transmission and support for diverse data types and communication scenarios.Polar codes have emerged as a prom...Mobile communications are reaching out to every aspect of our daily life,necessitating highefficiency data transmission and support for diverse data types and communication scenarios.Polar codes have emerged as a promising solution due to their outstanding error-correction performance and low complexity.Unequal error protection(UEP)involves nonuniform error safeguarding for distinct data segments,achieving a fine balance between error resilience and resource allocation,which ultimately enhancing system performance and efficiency.In this paper,we propose a novel class of UEP rateless polar codes.The codes are designed based on matrix extension of polar codes,and elegant mapping and duplication operations are designed to achieve UEP property while preserving the overall performance of conventional polar codes.Superior UEP performance is attained without significant modifications to conventional polar codes,making it straightforward for compatibility with existing polar codes.A theoretical analysis is conducted on the block error rate and throughput efficiency performance.To the best of our knowledge,this work provides the first theoretical performance analysis of UEP rateless polar codes.Simulation results show that the proposed codes significantly outperform existing polar coding schemes in both block error rate and throughput efficiency.展开更多
Open-wheeled race car aerodynamics is unquestionably challenging insofar as it involves many physical phenomena,such as slender and blunt body aerodynamics,ground effect,vortex management and interaction between diffe...Open-wheeled race car aerodynamics is unquestionably challenging insofar as it involves many physical phenomena,such as slender and blunt body aerodynamics,ground effect,vortex management and interaction between different sophisticated aero devices.In the current work,a 2017 F1 car aerodynamics has been investigated from a numerical point of view by using an open-source code.The vehicle project was developed by PERRINN(Copyright.2011—Present PERRINN),an engineering community founded by Nicolas Perrin in 2011.The racing car performance is quantitatively evaluated in terms of drag,downforce,efficiency and front balance.The goals of the present CFD(computational fluid dynamics)-based research are the following:analyzing the capabilities of the open-source software OpenFOAM in dealing with complex meshes and external aerodynamics calculation,and developing a reliable workflow from CAD(computer aided design)model to the post-processing of the results,in order to meet production demands.展开更多
Differential pulse-position modulation(DP PM)can achieve a good compromise between power and bandwidth requirements.However,the output sequence has undetectable insertions and deletions.This paper proposes a successiv...Differential pulse-position modulation(DP PM)can achieve a good compromise between power and bandwidth requirements.However,the output sequence has undetectable insertions and deletions.This paper proposes a successive cancellation(SC)decoding scheme based on the weighted levenshtein distance(WLD)of polar codes for correcting insertions/deletions in DPPM systems.In this method,the WLD is used to calculate the transfer probabilities recursively to obtain likelihood ratios,and the low-complexity SC decoding method is built according to the error characteristics to match the DPPM system.Additionally,the proposed SC decoding scheme is extended to list decoding,which can further improve error correction performance.Simulation results show that the proposed scheme can effectively correct insertions/deletions in the DPPM system,which enhances its reliability and performance.展开更多
As artificial Intelligence(AI)continues to expand exponentially,particularly with the emergence of generative pre-trained transformers(GPT)based on a transformer’s architecture,which has revolutionized data processin...As artificial Intelligence(AI)continues to expand exponentially,particularly with the emergence of generative pre-trained transformers(GPT)based on a transformer’s architecture,which has revolutionized data processing and enabled significant improvements in various applications.This document seeks to investigate the security vulnerabilities detection in the source code using a range of large language models(LLM).Our primary objective is to evaluate the effectiveness of Static Application Security Testing(SAST)by applying various techniques such as prompt persona,structure outputs and zero-shot.To the selection of the LLMs(CodeLlama 7B,DeepSeek coder 7B,Gemini 1.5 Flash,Gemini 2.0 Flash,Mistral 7b Instruct,Phi 38b Mini 128K instruct,Qwen 2.5 coder,StartCoder 27B)with comparison and combination with Find Security Bugs.The evaluation method will involve using a selected dataset containing vulnerabilities,and the results to provide insights for different scenarios according to the software criticality(Business critical,non-critical,minimum effort,best effort)In detail,the main objectives of this study are to investigate if large language models outperform or exceed the capabilities of traditional static analysis tools,if the combining LLMs with Static Application Security Testing(SAST)tools lead to an improvement and the possibility that local machine learning models on a normal computer produce reliable results.Summarizing the most important conclusions of the research,it can be said that while it is true that the results have improved depending on the size of the LLM for business-critical software,the best results have been obtained by SAST analysis.This differs in“NonCritical,”“Best Effort,”and“Minimum Effort”scenarios,where the combination of LLM(Gemini)+SAST has obtained better results.展开更多
Transformer-based models have significantly advanced binary code similarity detection(BCSD)by leveraging their semantic encoding capabilities for efficient function matching across diverse compilation settings.Althoug...Transformer-based models have significantly advanced binary code similarity detection(BCSD)by leveraging their semantic encoding capabilities for efficient function matching across diverse compilation settings.Although adversarial examples can strategically undermine the accuracy of BCSD models and protect critical code,existing techniques predominantly depend on inserting artificial instructions,which incur high computational costs and offer limited diversity of perturbations.To address these limitations,we propose AIMA,a novel gradient-guided assembly instruction relocation method.Our method decouples the detection model into tokenization,embedding,and encoding layers to enable efficient gradient computation.Since token IDs of instructions are discrete and nondifferentiable,we compute gradients in the continuous embedding space to evaluate the influence of each token.The most critical tokens are identified by calculating the L2 norm of their embedding gradients.We then establish a mapping between instructions and their corresponding tokens to aggregate token-level importance into instructionlevel significance.To maximize adversarial impact,a sliding window algorithm selects the most influential contiguous segments for relocation,ensuring optimal perturbation with minimal length.This approach efficiently locates critical code regions without expensive search operations.The selected segments are relocated outside their original function boundaries via a jump mechanism,which preserves runtime control flow and functionality while introducing“deletion”effects in the static instruction sequence.Extensive experiments show that AIMA reduces similarity scores by up to 35.8%in state-of-the-art BCSD models.When incorporated into training data,it also enhances model robustness,achieving a 5.9%improvement in AUROC.展开更多
The ultracold neutron(UCN)transport code,MCUCN,designed initially for simulating UCN transportation from a solid deuterium(SD_2)source and neutron electric dipole moment experiments,could not simulate UCN storage and ...The ultracold neutron(UCN)transport code,MCUCN,designed initially for simulating UCN transportation from a solid deuterium(SD_2)source and neutron electric dipole moment experiments,could not simulate UCN storage and transportation in a superfluid^(4)He(SFHe,He-Ⅱ)source accurately.This limitation arose from the absence of an^(4)He upscattering mechanism and the absorption of^(3)He.And the provided source energy distribution in MCUCN is different from that in SFHe source.This study introduced enhancements to MCUCN to address these constraints,explicitly incorporating the^(4)He upscattering effect,the absorption of^(3)He,the loss caused by impurities on converter wall,UCN source energy distribution in SFHe,and the transmission through negative optical potential.Additionally,a Python-based visualization code for intermediate states and results was developed.To validate these enhancements,we systematically compared the simulation results of the Lujan Center Mark3 UCN system by MCUCN and the improved MCUCN code(iMCUCN)with UCNtransport simulations.Additionally,we compared the results of the SUN1 system simulated by MCUCN and iMCUCN with measurement results.The study demonstrates that iMCUCN effectively simulates the storage and transportation of ultracold neutrons in He-Ⅱ.展开更多
Blind recognition of low-density paritycheck(LDPC)codes has gradually attracted more attention with the development of military and civil communications.However,in the case of the paritycheck matrices with relatively ...Blind recognition of low-density paritycheck(LDPC)codes has gradually attracted more attention with the development of military and civil communications.However,in the case of the paritycheck matrices with relatively high row weights,the existing blind recognition algorithms based on a candidate set generally perform worse.In this paper,we propose a blind recognition method for LDPC codes,called as tangent function assisted least square(TLS)method,which improves recognition performances by constructing a new cost function.To characterize the constraint degree among received vectors and paritycheck vectors,a feature function based on tangent function is constructed in the proposed algorithm.A cost function based on least square method is also established according to the feature function values satisfying the parity-check relationship.Moreover,the minimum average value in TLS is obtained on the candidate set.Numerical analysis and simulation results show that recognition performances of TLS algorithm are consistent with theoretical results.Compared with existing algorithms,the proposed method possesses better recognition performances.展开更多
LargeLanguageModels(LLMs)are increasingly appliedinthe fieldof code translation.However,existing evaluation methodologies suffer from two major limitations:(1)the high overlap between test data and pretraining corpora...LargeLanguageModels(LLMs)are increasingly appliedinthe fieldof code translation.However,existing evaluation methodologies suffer from two major limitations:(1)the high overlap between test data and pretraining corpora,which introduces significant bias in performance evaluation;and(2)mainstream metrics focus primarily on surface-level accuracy,failing to uncover the underlying factors that constrain model capabilities.To address these issues,this paper presents TCode(Translation-Oriented Code Evaluation benchmark)—a complexity-controllable,contamination-free benchmark dataset for code translation—alongside a dedicated static feature sensitivity evaluation framework.The dataset is carefully designed to control complexity along multiple dimensions—including syntactic nesting and expression intricacy—enabling both broad coverage and fine-grained differentiation of sample difficulty.This design supports precise evaluation of model capabilities across a wide spectrum of translation challenges.The proposed evaluation framework introduces a correlation-driven analysis mechanism based on static program features,enabling predictive modeling of translation success from two perspectives:Code Form Complexity(e.g.,code length and character density)and Semantic Modeling Complexity(e.g.,syntactic depth,control-flow nesting,and type system complexity).Empirical evaluations across representative LLMs—including Qwen2.5-72B and Llama3.3-70B—demonstrate that even state-of-the-art models achieve over 80% compilation success on simple samples,but their accuracy drops sharply below 40% on complex cases.Further correlation analysis indicates that Semantic Modeling Complexity alone is correlated with up to 60% of the variance in translation success,with static program features exhibiting nonlinear threshold effects that highlight clear capability boundaries.This study departs fromthe traditional accuracy-centric evaluation paradigm and,for the first time,systematically characterizes the capabilities of large languagemodels in translation tasks through the lens of programstatic features.The findings provide actionable insights for model refinement and training strategy development.展开更多
Automated Program Repair(APR)techniques have shown significant potential in mitigating the cost and complexity associated with debugging by automatically generating corrective patches for software defects.Despite cons...Automated Program Repair(APR)techniques have shown significant potential in mitigating the cost and complexity associated with debugging by automatically generating corrective patches for software defects.Despite considerable progress in APR methodologies,existing approaches frequently lack contextual awareness of runtime behaviors and structural intricacies inherent in buggy source code.In this paper,we propose a novel APR approach that integrates attention mechanisms within an autoencoder-based framework,explicitly utilizing structural code affinity and execution context correlation derived from stack trace analysis.Our approach begins with an innovative preprocessing pipeline,where code segments and stack traces are transformed into tokenized representations.Subsequently,the BM25 ranking algorithm is employed to quantitatively measure structural code affinity and execution context correlation,identifying syntactically and semantically analogous buggy code snippets and relevant runtime error contexts from extensive repositories.These extracted features are then encoded via an attention-enhanced autoencoder model,specifically designed to capture significant patterns and correlations essential for effective patch generation.To assess the efficacy and generalizability of our proposed method,we conducted rigorous experimental comparisons against DeepFix,a state-of-the-art APR system,using a substantial dataset comprising 53,478 studentdeveloped C programs.Experimental outcomes indicate that our model achieves a notable bug repair success rate of approximately 62.36%,representing a statistically significant performance improvement of over 6%compared to the baseline.Furthermore,a thorough K-fold cross-validation reinforced the consistency,robustness,and reliability of our method across diverse subsets of the dataset.Our findings present the critical advantage of integrating attentionbased learning with code structural and execution context features in APR tasks,leading to improved accuracy and practical applicability.Future work aims to extend the model’s applicability across different programming languages,systematically optimize hyperparameters,and explore alternative feature representation methods to further enhance debugging efficiency and effectiveness.展开更多
In erasure-coded storage systems,updating data requires parity maintenance,which often leads to significant I/O amplification due to“write-after-read”operations.Furthermore,scattered parity placement increases disk ...In erasure-coded storage systems,updating data requires parity maintenance,which often leads to significant I/O amplification due to“write-after-read”operations.Furthermore,scattered parity placement increases disk seek overhead during repair,resulting in degraded system performance.To address these challenges,this paper proposes a Cognitive Update and Repair Method(CURM)that leverages machine learning to classify files into writeonly,read-only,and read-write categories,enabling tailored update and repair strategies.For write-only and read-write files,CURM employs a data-differencemechanism combined with fine-grained I/O scheduling to minimize redundant read operations and mitigate I/O amplification.For read-write files,CURM further reserves adjacent disk space near parity blocks,supporting parallel reads and reducing disk seek overhead during repair.We implement CURM in a prototype system,Cognitive Update and Repair File System(CURFS),and conduct extensive experiments using realworld Network File System(NFS)and Microsoft Research(MSR)workloads on a 25-node cluster.Experimental results demonstrate that CURMimproves data update throughput by up to 82.52%,reduces recovery time by up to 47.47%,and decreases long-term storage overhead by more than 15% compared to state-of-the-art methods including Full Logging(FL),ParityLogging(PL),ParityLoggingwithReservedspace(PLR),andPARIX.These results validate the effectiveness of CURM in enhancing both update and repair performance,providing a scalable and efficient solution for large-scale erasure-coded storage systems.展开更多
We present a high performance modularly-built open-source software-OpenIFEM.OpenIFEM is a C++implementation of the modified immersed finite element method(mIFEM)to solve fluid-structure interaction(FSI)problems.This s...We present a high performance modularly-built open-source software-OpenIFEM.OpenIFEM is a C++implementation of the modified immersed finite element method(mIFEM)to solve fluid-structure interaction(FSI)problems.This software is modularly built to perform multiple tasks including fluid dynamics(incompressible and slightly compressible fluid models),linear and nonlinear solid mechanics,and fully coupled fluid-structure interactions.Most of open-source software packages are restricted to certain discretization methods;some are under-tested,under-documented,and lack modularity as well as extensibility.OpenIFEM is designed and built to include a set of generic classes for users to adapt so that any fluid and solid solvers can be coupled through the FSI algorithm.In addition,the package utilizes well-developed and tested libraries.It also comes with standard test cases that serve as software and algorithm validation.The software can be built on cross-platform,i.e.,Linux,Windows,and Mac OS,using CMake.Efficient parallelization is also implemented for high-performance computing for large-sized problems.OpenIFEM is documented using Doxygen and publicly available to download on GitHub.It is expected to benefit the future development of FSI algorithms and be applied to a variety of FSI applications.展开更多
With the rapid development of Open-Source(OS),more and more software projects are maintained and developed in the form of OS.These Open-Source projects depend on and influence each other,gradually forming a huge OS pr...With the rapid development of Open-Source(OS),more and more software projects are maintained and developed in the form of OS.These Open-Source projects depend on and influence each other,gradually forming a huge OS project network,namely an Open-Source Software ECOsystem(OSSECO).Unfortunately,not all OS projects in the open-source ecosystem can be healthy and stable in the long term,and more projects will go from active to inactive and gradually die.In a tightly connected ecosystem,the death of one project can potentially cause the collapse of the entire ecosystem network.How can we effectively prevent such situations from happening?In this paper,we first identify the basic project characteristics that affect the survival of OS projects at both project and ecosystem levels through the proportional hazards model.Then,we utilize graph convolutional networks based on the ecosystem network to extract the ecosystem environment characteristics of OS projects.Finally,we fuse basic project characteristics and environmental project characteristics and construct a Hybrid Structured Prediction Model(HSPM)to predict the OS project survival state.The experimental results show that HSPM significantly improved compared to the traditional prediction model.Our work can substantially assist OS project managers in maintaining their projects’health.It can also provide an essential reference for developers when choosing the right open-source project for their production activities.展开更多
Scientific research requires the collection of data in order to study, monitor, analyze, describe, or understand a particular process or event. Data collection efforts are often a compromise: manual measurements can b...Scientific research requires the collection of data in order to study, monitor, analyze, describe, or understand a particular process or event. Data collection efforts are often a compromise: manual measurements can be time-consuming and labor-intensive, resulting in data being collected at a low frequency, while automating the data-collection process can reduce labor requirements and increase the frequency of measurements, but at the cost of added expense of electronic data-collecting instrumentation. Rapid advances in electronic technologies have resulted in a variety of new and inexpensive sensing, monitoring, and control capabilities which offer opportunities for implementation in agricultural and natural-resource research applications. An Open Source Hardware project called Arduino consists of a programmable microcontroller development platform, expansion capability through add-on boards, and a programming development environment for creating custom microcontroller software. All circuit-board and electronic component specifications, as well as the programming software, are open-source and freely available for anyone to use or modify. Inexpensive sensors and the Arduino development platform were used to develop several inexpensive, automated sensing and datalogging systems for use in agricultural and natural-resources related research projects. Systems were developed and implemented to monitor soil-moisture status of field crops for irrigation scheduling and crop-water use studies, to measure daily evaporation-pan water levels for quantifying evaporative demand, and to monitor environmental parameters under forested conditions. These studies demonstrate the usefulness of automated measurements, and offer guidance for other researchers in developing inexpensive sensing and monitoring systems to further their research.展开更多
Resource-scarce regions with serious COVID-19 outbreaks do not have enough ventilators to support critically ill patients,and these shortages are especially devastating in developing countries.To help alleviate this s...Resource-scarce regions with serious COVID-19 outbreaks do not have enough ventilators to support critically ill patients,and these shortages are especially devastating in developing countries.To help alleviate this strain,we have designed and tested the accessible low-barrier in vivo-validated economical ventilator(ALIVE Vent),a COVID-19-inspired,cost-effective,open-source,in vivo-validated solution made from commercially available components.The ALIVE Vent operates using compressed oxygen and air to drive inspiration,while two solenoid valves ensure one-way flow and precise cycle timing.The device was functionally tested and profiled using a variable resistance and compliance artificial lung and validated in anesthetized large animals.Our functional test results revealed its effective operation under a wide variety of ventilation conditions defined by the American Association of Respiratory Care guidelines for ventilator stockpiling.The large animal test showed that our ventilator performed similarly if not better than a standard ventilator in maintaining optimal ventilation status.The FiO2,respiratory rate,inspiratory to expiratory time ratio,positive-end expiratory pressure,and peak inspiratory pressure were successfully maintained within normal,clinically validated ranges,and the animals were recovered without any complications.In regions with limited access to ventilators,the ALIVE Vent can help alleviate shortages,and we have ensured that all used materials are publicly available.While this pandemic has elucidated enormous global inequalities in healthcare,innovative,cost-effective solutions aimed at reducing socio-economic barriers,such as the ALIVE Vent,can help enable access to prompt healthcare and life saving technology on a global scale and beyond COVID-19.展开更多
In today’s society with advanced Internet,the amount of information increases dramatically with each passing day,which leads to increasingly complex processes of open-source intelligence.Therefore,it is more importan...In today’s society with advanced Internet,the amount of information increases dramatically with each passing day,which leads to increasingly complex processes of open-source intelligence.Therefore,it is more important to rationalize the operation mode and improve the operation efficiency of open-source intelligence under the premise of satisfying users’needs.This paper focuses on the simulation study of the process system of opensource intelligence from the user’s perspective.First,the basic concept and development status of open-source intelligence are introduced in details.Second,six existing intelligence operation process models are summarized and their advantages and disadvantages are compared in focus.Based on users’preference,the open-source intelligence system simulation theory model is constructed from four aspects:intelligence collection,intelligence processing,intelligence analysis,and intelligence delivery.Meanwhile,the dynamics model of the open-source intelligence process system is constructed based on the open-source intelligence system simulation theoretical model,which specifically includes five parts:determination of system boundary,construction of causal loop diagram,construction of stock flow diagram,writing ofmathematical equations,and system sensitivity test.Finally,the system simulation results were analyzed.It was found that improving the system of intelligence agencies,opening up government affairs,improving the professional level of intelligence personnel,strengthening the communication and cooperation among personnel of various intelligence departments,and expressing intelligence products through diverse forms can effectively improve the operational efficiency of the open-source intelligence process system.展开更多
基金supported by the National Key R&D Program of China[Grant Number 2020YFB1708300]the National Natural Science Foundation of China[Grant Number 52075184].
文摘Topology optimization(TO),a numerical technique to find the optimalmaterial layoutwith a given design domain,has attracted interest from researchers in the field of structural optimization in recent years.For beginners,opensource codes are undoubtedly the best alternative to learning TO,which can elaborate the implementation of a method in detail and easily engage more people to employ and extend the method.In this paper,we present a summary of various open-source codes and related literature on TO methods,including solid isotropic material with penalization(SIMP),evolutionary method,level set method(LSM),moving morphable components/voids(MMC/MMV)methods,multiscale topology optimization method,etc.Simultaneously,we classify the codes into five levels,fromeasy to difficult,depending on their difficulty,so that beginners can get started and understand the form of code implementation more quickly.
基金supported in part by the National Natural Science Foundation of China(52172377).
文摘An effective energy management strategy(EMS)is essential to optimize the energy efficiency of electric vehicles(EVs).With the advent of advanced machine learning techniques,the focus on developing sophisticated EMS for EVs is increasing.Here,we introduce LearningEMS:a unified framework and open-source benchmark designed to facilitate rapid development and assessment of EMS.LearningEMS is distinguished by its ability to support a variety of EV configurations,including hybrid EVs,fuel cell EVs,and plug-in EVs,offering a general platform for the development of EMS.The framework enables detailed comparisons of several EMS algorithms,encompassing imitation learning,deep reinforcement learning(RL),offline RL,model predictive control,and dynamic programming.We rigorously evaluated these algorithms across multiple perspectives:energy efficiency,consistency,adaptability,and practicability.Furthermore,we discuss state,reward,and action settings for RL in EV energy management,introduce a policy extraction and reconstruction method for learning-based EMS deployment,and conduct hardware-in-the-loop experiments.In summary,we offer a unified and comprehensive framework that comes with three distinct EV platforms,over 10000 km of EMS policy data set,ten state-of-the-art algorithms,and over 160 benchmark tasks,along with three learning libraries.Its flexible design allows easy expansion for additional tasks and applications.The open-source algorithms,models,data sets,and deployment processes foster additional research and innovation in EV and broader engineering domains.
文摘介绍了高温蠕变工况下运行的压力容器可能出现的失效模式,结合工程设计现状,指出了我国当前压力容器标准体系在确定高温蠕变工况许用压应力时存在的技术瓶颈,在此基础之上引出ASME Code Case 3029,对其适用范围、发展历程、产生背景及工程意义进行了简单的介绍,以某工程设计项目中的实际结构为例,介绍了该方法的使用过程及注意事项,并结合压力容器工程设计领域的实际需求,对我国标准体系下一步的制定或修订方向提出了展望。
基金supported by National Natural Science Foundation of China(No.62301008)China Postdoctoral Science Foundation(No.2022M720272)New Cornerstone Science Foundation through the XPLORER PRIZE。
文摘Mobile communications are reaching out to every aspect of our daily life,necessitating highefficiency data transmission and support for diverse data types and communication scenarios.Polar codes have emerged as a promising solution due to their outstanding error-correction performance and low complexity.Unequal error protection(UEP)involves nonuniform error safeguarding for distinct data segments,achieving a fine balance between error resilience and resource allocation,which ultimately enhancing system performance and efficiency.In this paper,we propose a novel class of UEP rateless polar codes.The codes are designed based on matrix extension of polar codes,and elegant mapping and duplication operations are designed to achieve UEP property while preserving the overall performance of conventional polar codes.Superior UEP performance is attained without significant modifications to conventional polar codes,making it straightforward for compatibility with existing polar codes.A theoretical analysis is conducted on the block error rate and throughput efficiency performance.To the best of our knowledge,this work provides the first theoretical performance analysis of UEP rateless polar codes.Simulation results show that the proposed codes significantly outperform existing polar coding schemes in both block error rate and throughput efficiency.
文摘Open-wheeled race car aerodynamics is unquestionably challenging insofar as it involves many physical phenomena,such as slender and blunt body aerodynamics,ground effect,vortex management and interaction between different sophisticated aero devices.In the current work,a 2017 F1 car aerodynamics has been investigated from a numerical point of view by using an open-source code.The vehicle project was developed by PERRINN(Copyright.2011—Present PERRINN),an engineering community founded by Nicolas Perrin in 2011.The racing car performance is quantitatively evaluated in terms of drag,downforce,efficiency and front balance.The goals of the present CFD(computational fluid dynamics)-based research are the following:analyzing the capabilities of the open-source software OpenFOAM in dealing with complex meshes and external aerodynamics calculation,and developing a reliable workflow from CAD(computer aided design)model to the post-processing of the results,in order to meet production demands.
基金supported by National Natural Science Foundation of China(No.61801327).
文摘Differential pulse-position modulation(DP PM)can achieve a good compromise between power and bandwidth requirements.However,the output sequence has undetectable insertions and deletions.This paper proposes a successive cancellation(SC)decoding scheme based on the weighted levenshtein distance(WLD)of polar codes for correcting insertions/deletions in DPPM systems.In this method,the WLD is used to calculate the transfer probabilities recursively to obtain likelihood ratios,and the low-complexity SC decoding method is built according to the error characteristics to match the DPPM system.Additionally,the proposed SC decoding scheme is extended to list decoding,which can further improve error correction performance.Simulation results show that the proposed scheme can effectively correct insertions/deletions in the DPPM system,which enhances its reliability and performance.
文摘As artificial Intelligence(AI)continues to expand exponentially,particularly with the emergence of generative pre-trained transformers(GPT)based on a transformer’s architecture,which has revolutionized data processing and enabled significant improvements in various applications.This document seeks to investigate the security vulnerabilities detection in the source code using a range of large language models(LLM).Our primary objective is to evaluate the effectiveness of Static Application Security Testing(SAST)by applying various techniques such as prompt persona,structure outputs and zero-shot.To the selection of the LLMs(CodeLlama 7B,DeepSeek coder 7B,Gemini 1.5 Flash,Gemini 2.0 Flash,Mistral 7b Instruct,Phi 38b Mini 128K instruct,Qwen 2.5 coder,StartCoder 27B)with comparison and combination with Find Security Bugs.The evaluation method will involve using a selected dataset containing vulnerabilities,and the results to provide insights for different scenarios according to the software criticality(Business critical,non-critical,minimum effort,best effort)In detail,the main objectives of this study are to investigate if large language models outperform or exceed the capabilities of traditional static analysis tools,if the combining LLMs with Static Application Security Testing(SAST)tools lead to an improvement and the possibility that local machine learning models on a normal computer produce reliable results.Summarizing the most important conclusions of the research,it can be said that while it is true that the results have improved depending on the size of the LLM for business-critical software,the best results have been obtained by SAST analysis.This differs in“NonCritical,”“Best Effort,”and“Minimum Effort”scenarios,where the combination of LLM(Gemini)+SAST has obtained better results.
基金supported by Key Laboratory of Cyberspace Security,Ministry of Education,China。
文摘Transformer-based models have significantly advanced binary code similarity detection(BCSD)by leveraging their semantic encoding capabilities for efficient function matching across diverse compilation settings.Although adversarial examples can strategically undermine the accuracy of BCSD models and protect critical code,existing techniques predominantly depend on inserting artificial instructions,which incur high computational costs and offer limited diversity of perturbations.To address these limitations,we propose AIMA,a novel gradient-guided assembly instruction relocation method.Our method decouples the detection model into tokenization,embedding,and encoding layers to enable efficient gradient computation.Since token IDs of instructions are discrete and nondifferentiable,we compute gradients in the continuous embedding space to evaluate the influence of each token.The most critical tokens are identified by calculating the L2 norm of their embedding gradients.We then establish a mapping between instructions and their corresponding tokens to aggregate token-level importance into instructionlevel significance.To maximize adversarial impact,a sliding window algorithm selects the most influential contiguous segments for relocation,ensuring optimal perturbation with minimal length.This approach efficiently locates critical code regions without expensive search operations.The selected segments are relocated outside their original function boundaries via a jump mechanism,which preserves runtime control flow and functionality while introducing“deletion”effects in the static instruction sequence.Extensive experiments show that AIMA reduces similarity scores by up to 35.8%in state-of-the-art BCSD models.When incorporated into training data,it also enhances model robustness,achieving a 5.9%improvement in AUROC.
基金the National Key R&D Program of China(No.2024YFE0110001)the National Natural Science Foundation of China(U1932219)the Mobility Programme endorsed by the Joint Committee of the Sino-German Center(M0728)。
文摘The ultracold neutron(UCN)transport code,MCUCN,designed initially for simulating UCN transportation from a solid deuterium(SD_2)source and neutron electric dipole moment experiments,could not simulate UCN storage and transportation in a superfluid^(4)He(SFHe,He-Ⅱ)source accurately.This limitation arose from the absence of an^(4)He upscattering mechanism and the absorption of^(3)He.And the provided source energy distribution in MCUCN is different from that in SFHe source.This study introduced enhancements to MCUCN to address these constraints,explicitly incorporating the^(4)He upscattering effect,the absorption of^(3)He,the loss caused by impurities on converter wall,UCN source energy distribution in SFHe,and the transmission through negative optical potential.Additionally,a Python-based visualization code for intermediate states and results was developed.To validate these enhancements,we systematically compared the simulation results of the Lujan Center Mark3 UCN system by MCUCN and the improved MCUCN code(iMCUCN)with UCNtransport simulations.Additionally,we compared the results of the SUN1 system simulated by MCUCN and iMCUCN with measurement results.The study demonstrates that iMCUCN effectively simulates the storage and transportation of ultracold neutrons in He-Ⅱ.
基金Fundamental Research Funds for the Central Universities under Grant 3072025YC0802the National Natural Science Foundation of China under Grant 62001138Heilongjiang Provincial Natural Science Foundation of China under Grant LH2021F009。
文摘Blind recognition of low-density paritycheck(LDPC)codes has gradually attracted more attention with the development of military and civil communications.However,in the case of the paritycheck matrices with relatively high row weights,the existing blind recognition algorithms based on a candidate set generally perform worse.In this paper,we propose a blind recognition method for LDPC codes,called as tangent function assisted least square(TLS)method,which improves recognition performances by constructing a new cost function.To characterize the constraint degree among received vectors and paritycheck vectors,a feature function based on tangent function is constructed in the proposed algorithm.A cost function based on least square method is also established according to the feature function values satisfying the parity-check relationship.Moreover,the minimum average value in TLS is obtained on the candidate set.Numerical analysis and simulation results show that recognition performances of TLS algorithm are consistent with theoretical results.Compared with existing algorithms,the proposed method possesses better recognition performances.
文摘LargeLanguageModels(LLMs)are increasingly appliedinthe fieldof code translation.However,existing evaluation methodologies suffer from two major limitations:(1)the high overlap between test data and pretraining corpora,which introduces significant bias in performance evaluation;and(2)mainstream metrics focus primarily on surface-level accuracy,failing to uncover the underlying factors that constrain model capabilities.To address these issues,this paper presents TCode(Translation-Oriented Code Evaluation benchmark)—a complexity-controllable,contamination-free benchmark dataset for code translation—alongside a dedicated static feature sensitivity evaluation framework.The dataset is carefully designed to control complexity along multiple dimensions—including syntactic nesting and expression intricacy—enabling both broad coverage and fine-grained differentiation of sample difficulty.This design supports precise evaluation of model capabilities across a wide spectrum of translation challenges.The proposed evaluation framework introduces a correlation-driven analysis mechanism based on static program features,enabling predictive modeling of translation success from two perspectives:Code Form Complexity(e.g.,code length and character density)and Semantic Modeling Complexity(e.g.,syntactic depth,control-flow nesting,and type system complexity).Empirical evaluations across representative LLMs—including Qwen2.5-72B and Llama3.3-70B—demonstrate that even state-of-the-art models achieve over 80% compilation success on simple samples,but their accuracy drops sharply below 40% on complex cases.Further correlation analysis indicates that Semantic Modeling Complexity alone is correlated with up to 60% of the variance in translation success,with static program features exhibiting nonlinear threshold effects that highlight clear capability boundaries.This study departs fromthe traditional accuracy-centric evaluation paradigm and,for the first time,systematically characterizes the capabilities of large languagemodels in translation tasks through the lens of programstatic features.The findings provide actionable insights for model refinement and training strategy development.
文摘Automated Program Repair(APR)techniques have shown significant potential in mitigating the cost and complexity associated with debugging by automatically generating corrective patches for software defects.Despite considerable progress in APR methodologies,existing approaches frequently lack contextual awareness of runtime behaviors and structural intricacies inherent in buggy source code.In this paper,we propose a novel APR approach that integrates attention mechanisms within an autoencoder-based framework,explicitly utilizing structural code affinity and execution context correlation derived from stack trace analysis.Our approach begins with an innovative preprocessing pipeline,where code segments and stack traces are transformed into tokenized representations.Subsequently,the BM25 ranking algorithm is employed to quantitatively measure structural code affinity and execution context correlation,identifying syntactically and semantically analogous buggy code snippets and relevant runtime error contexts from extensive repositories.These extracted features are then encoded via an attention-enhanced autoencoder model,specifically designed to capture significant patterns and correlations essential for effective patch generation.To assess the efficacy and generalizability of our proposed method,we conducted rigorous experimental comparisons against DeepFix,a state-of-the-art APR system,using a substantial dataset comprising 53,478 studentdeveloped C programs.Experimental outcomes indicate that our model achieves a notable bug repair success rate of approximately 62.36%,representing a statistically significant performance improvement of over 6%compared to the baseline.Furthermore,a thorough K-fold cross-validation reinforced the consistency,robustness,and reliability of our method across diverse subsets of the dataset.Our findings present the critical advantage of integrating attentionbased learning with code structural and execution context features in APR tasks,leading to improved accuracy and practical applicability.Future work aims to extend the model’s applicability across different programming languages,systematically optimize hyperparameters,and explore alternative feature representation methods to further enhance debugging efficiency and effectiveness.
基金supported by the National Natural Science Foundation of China(Grant No.62362019)the Natural Science Foundation of Hainan Province(Grant No.624RC482)the Hainan Provincial Higher Education Teaching Reform Research Project(Grant Hnjg2024-27).
文摘In erasure-coded storage systems,updating data requires parity maintenance,which often leads to significant I/O amplification due to“write-after-read”operations.Furthermore,scattered parity placement increases disk seek overhead during repair,resulting in degraded system performance.To address these challenges,this paper proposes a Cognitive Update and Repair Method(CURM)that leverages machine learning to classify files into writeonly,read-only,and read-write categories,enabling tailored update and repair strategies.For write-only and read-write files,CURM employs a data-differencemechanism combined with fine-grained I/O scheduling to minimize redundant read operations and mitigate I/O amplification.For read-write files,CURM further reserves adjacent disk space near parity blocks,supporting parallel reads and reducing disk seek overhead during repair.We implement CURM in a prototype system,Cognitive Update and Repair File System(CURFS),and conduct extensive experiments using realworld Network File System(NFS)and Microsoft Research(MSR)workloads on a 25-node cluster.Experimental results demonstrate that CURMimproves data update throughput by up to 82.52%,reduces recovery time by up to 47.47%,and decreases long-term storage overhead by more than 15% compared to state-of-the-art methods including Full Logging(FL),ParityLogging(PL),ParityLoggingwithReservedspace(PLR),andPARIX.These results validate the effectiveness of CURM in enhancing both update and repair performance,providing a scalable and efficient solution for large-scale erasure-coded storage systems.
文摘We present a high performance modularly-built open-source software-OpenIFEM.OpenIFEM is a C++implementation of the modified immersed finite element method(mIFEM)to solve fluid-structure interaction(FSI)problems.This software is modularly built to perform multiple tasks including fluid dynamics(incompressible and slightly compressible fluid models),linear and nonlinear solid mechanics,and fully coupled fluid-structure interactions.Most of open-source software packages are restricted to certain discretization methods;some are under-tested,under-documented,and lack modularity as well as extensibility.OpenIFEM is designed and built to include a set of generic classes for users to adapt so that any fluid and solid solvers can be coupled through the FSI algorithm.In addition,the package utilizes well-developed and tested libraries.It also comes with standard test cases that serve as software and algorithm validation.The software can be built on cross-platform,i.e.,Linux,Windows,and Mac OS,using CMake.Efficient parallelization is also implemented for high-performance computing for large-sized problems.OpenIFEM is documented using Doxygen and publicly available to download on GitHub.It is expected to benefit the future development of FSI algorithms and be applied to a variety of FSI applications.
基金This work was supported by the National Social Science Foundation(NSSF)Research on intelligent recommendation of multi-modal resources for children’s graded reading in smart library(22BTQ033)the Science and Technology Research and Development Program Project of China railway group limited(Project No.2021-Special-08).
文摘With the rapid development of Open-Source(OS),more and more software projects are maintained and developed in the form of OS.These Open-Source projects depend on and influence each other,gradually forming a huge OS project network,namely an Open-Source Software ECOsystem(OSSECO).Unfortunately,not all OS projects in the open-source ecosystem can be healthy and stable in the long term,and more projects will go from active to inactive and gradually die.In a tightly connected ecosystem,the death of one project can potentially cause the collapse of the entire ecosystem network.How can we effectively prevent such situations from happening?In this paper,we first identify the basic project characteristics that affect the survival of OS projects at both project and ecosystem levels through the proportional hazards model.Then,we utilize graph convolutional networks based on the ecosystem network to extract the ecosystem environment characteristics of OS projects.Finally,we fuse basic project characteristics and environmental project characteristics and construct a Hybrid Structured Prediction Model(HSPM)to predict the OS project survival state.The experimental results show that HSPM significantly improved compared to the traditional prediction model.Our work can substantially assist OS project managers in maintaining their projects’health.It can also provide an essential reference for developers when choosing the right open-source project for their production activities.
文摘Scientific research requires the collection of data in order to study, monitor, analyze, describe, or understand a particular process or event. Data collection efforts are often a compromise: manual measurements can be time-consuming and labor-intensive, resulting in data being collected at a low frequency, while automating the data-collection process can reduce labor requirements and increase the frequency of measurements, but at the cost of added expense of electronic data-collecting instrumentation. Rapid advances in electronic technologies have resulted in a variety of new and inexpensive sensing, monitoring, and control capabilities which offer opportunities for implementation in agricultural and natural-resource research applications. An Open Source Hardware project called Arduino consists of a programmable microcontroller development platform, expansion capability through add-on boards, and a programming development environment for creating custom microcontroller software. All circuit-board and electronic component specifications, as well as the programming software, are open-source and freely available for anyone to use or modify. Inexpensive sensors and the Arduino development platform were used to develop several inexpensive, automated sensing and datalogging systems for use in agricultural and natural-resources related research projects. Systems were developed and implemented to monitor soil-moisture status of field crops for irrigation scheduling and crop-water use studies, to measure daily evaporation-pan water levels for quantifying evaporative demand, and to monitor environmental parameters under forested conditions. These studies demonstrate the usefulness of automated measurements, and offer guidance for other researchers in developing inexpensive sensing and monitoring systems to further their research.
基金the National Institutes of Health(NIH R01 HL089315-01 and NIH R01 HL152155,YJW)the Thoracic Surgery Foundation Resident Research Fellowship(YZ)the National Science Foundation Graduate Research Fellowship Program(AMI).
文摘Resource-scarce regions with serious COVID-19 outbreaks do not have enough ventilators to support critically ill patients,and these shortages are especially devastating in developing countries.To help alleviate this strain,we have designed and tested the accessible low-barrier in vivo-validated economical ventilator(ALIVE Vent),a COVID-19-inspired,cost-effective,open-source,in vivo-validated solution made from commercially available components.The ALIVE Vent operates using compressed oxygen and air to drive inspiration,while two solenoid valves ensure one-way flow and precise cycle timing.The device was functionally tested and profiled using a variable resistance and compliance artificial lung and validated in anesthetized large animals.Our functional test results revealed its effective operation under a wide variety of ventilation conditions defined by the American Association of Respiratory Care guidelines for ventilator stockpiling.The large animal test showed that our ventilator performed similarly if not better than a standard ventilator in maintaining optimal ventilation status.The FiO2,respiratory rate,inspiratory to expiratory time ratio,positive-end expiratory pressure,and peak inspiratory pressure were successfully maintained within normal,clinically validated ranges,and the animals were recovered without any complications.In regions with limited access to ventilators,the ALIVE Vent can help alleviate shortages,and we have ensured that all used materials are publicly available.While this pandemic has elucidated enormous global inequalities in healthcare,innovative,cost-effective solutions aimed at reducing socio-economic barriers,such as the ALIVE Vent,can help enable access to prompt healthcare and life saving technology on a global scale and beyond COVID-19.
基金supported by the National Social Science Foundation of China under the project“Research on the mechanism of developing and utilizing domestic and foreign open-source intelligence under product-oriented thinking(20BTQ049)”.
文摘In today’s society with advanced Internet,the amount of information increases dramatically with each passing day,which leads to increasingly complex processes of open-source intelligence.Therefore,it is more important to rationalize the operation mode and improve the operation efficiency of open-source intelligence under the premise of satisfying users’needs.This paper focuses on the simulation study of the process system of opensource intelligence from the user’s perspective.First,the basic concept and development status of open-source intelligence are introduced in details.Second,six existing intelligence operation process models are summarized and their advantages and disadvantages are compared in focus.Based on users’preference,the open-source intelligence system simulation theory model is constructed from four aspects:intelligence collection,intelligence processing,intelligence analysis,and intelligence delivery.Meanwhile,the dynamics model of the open-source intelligence process system is constructed based on the open-source intelligence system simulation theoretical model,which specifically includes five parts:determination of system boundary,construction of causal loop diagram,construction of stock flow diagram,writing ofmathematical equations,and system sensitivity test.Finally,the system simulation results were analyzed.It was found that improving the system of intelligence agencies,opening up government affairs,improving the professional level of intelligence personnel,strengthening the communication and cooperation among personnel of various intelligence departments,and expressing intelligence products through diverse forms can effectively improve the operational efficiency of the open-source intelligence process system.