Understanding the complex interactions between human activities and ecosystem functions is a prerequisite for achieving sustainable development.Since the implementation of the“Grain for Green”Project in 1999,ecosyst...Understanding the complex interactions between human activities and ecosystem functions is a prerequisite for achieving sustainable development.Since the implementation of the“Grain for Green”Project in 1999,ecosystem functions in China’s Loess Plateau have significantly improved.However,intensified human activities have also exacerbated the pressures on the region’s fragile ecological environment.This study investigates the spatiotemporal variations in the human activity intensity index(HAI)and net ecosystem benefits(NEB)from 2000 to 2020,using expert-based assessments and an enhanced cost-benefit evaluation framework.Results indicate that HAI increased by 16.7% and 16.6% at the grid and county levels,respectively.NEB exhibited pronounced spatial heterogeneity,with a total increase of USD 36.2 trillion at the grid scale.At the county level,the average NEB rose by 75%.The degree of trade-off was higher at the grid scale than at the county scale,while the synergistic areas initially expanded and then declined at both scales.Key areas for improvement and regions of lagging development were identified as priority zones for ecological management and spatial planning at both spatial resolutions.This study offers scientific insights and practical guidance for harmonizing ecological conservation with high-quality development in ecologically vulnerable regions.展开更多
Differential Code Bias(DCB)is the time delays between two different GNSS signals,which is crucial for GNSS positioning.Previous studies have shown that it can be significantly affected by the flex power operations in ...Differential Code Bias(DCB)is the time delays between two different GNSS signals,which is crucial for GNSS positioning.Previous studies have shown that it can be significantly affected by the flex power operations in satellites.This study proposes a 15-min short-term DCB estimation method to analyze flex power's impact on DCB variations.The method jointly estimates satellite DCB,receiver DCB,and ionospheric parameters using over 300 MGEX stations.We examined three representative flex power events in 2024,achieving average internal RMS values of 0.042 ns and 0.0068 ns for inter-frequency and intra-frequency scenarios respectively.Results show that intra-frequency DCB exhibits clear shift biases synchronized with flex power state transitions while maintaining stability within 0.20 ns during nontransition periods.No definitive impact on inter-frequency DCB was observed at current estimation precision levels.展开更多
Differential pulse-position modulation(DP PM)can achieve a good compromise between power and bandwidth requirements.However,the output sequence has undetectable insertions and deletions.This paper proposes a successiv...Differential pulse-position modulation(DP PM)can achieve a good compromise between power and bandwidth requirements.However,the output sequence has undetectable insertions and deletions.This paper proposes a successive cancellation(SC)decoding scheme based on the weighted levenshtein distance(WLD)of polar codes for correcting insertions/deletions in DPPM systems.In this method,the WLD is used to calculate the transfer probabilities recursively to obtain likelihood ratios,and the low-complexity SC decoding method is built according to the error characteristics to match the DPPM system.Additionally,the proposed SC decoding scheme is extended to list decoding,which can further improve error correction performance.Simulation results show that the proposed scheme can effectively correct insertions/deletions in the DPPM system,which enhances its reliability and performance.展开更多
As artificial Intelligence(AI)continues to expand exponentially,particularly with the emergence of generative pre-trained transformers(GPT)based on a transformer’s architecture,which has revolutionized data processin...As artificial Intelligence(AI)continues to expand exponentially,particularly with the emergence of generative pre-trained transformers(GPT)based on a transformer’s architecture,which has revolutionized data processing and enabled significant improvements in various applications.This document seeks to investigate the security vulnerabilities detection in the source code using a range of large language models(LLM).Our primary objective is to evaluate the effectiveness of Static Application Security Testing(SAST)by applying various techniques such as prompt persona,structure outputs and zero-shot.To the selection of the LLMs(CodeLlama 7B,DeepSeek coder 7B,Gemini 1.5 Flash,Gemini 2.0 Flash,Mistral 7b Instruct,Phi 38b Mini 128K instruct,Qwen 2.5 coder,StartCoder 27B)with comparison and combination with Find Security Bugs.The evaluation method will involve using a selected dataset containing vulnerabilities,and the results to provide insights for different scenarios according to the software criticality(Business critical,non-critical,minimum effort,best effort)In detail,the main objectives of this study are to investigate if large language models outperform or exceed the capabilities of traditional static analysis tools,if the combining LLMs with Static Application Security Testing(SAST)tools lead to an improvement and the possibility that local machine learning models on a normal computer produce reliable results.Summarizing the most important conclusions of the research,it can be said that while it is true that the results have improved depending on the size of the LLM for business-critical software,the best results have been obtained by SAST analysis.This differs in“NonCritical,”“Best Effort,”and“Minimum Effort”scenarios,where the combination of LLM(Gemini)+SAST has obtained better results.展开更多
Transformer-based models have significantly advanced binary code similarity detection(BCSD)by leveraging their semantic encoding capabilities for efficient function matching across diverse compilation settings.Althoug...Transformer-based models have significantly advanced binary code similarity detection(BCSD)by leveraging their semantic encoding capabilities for efficient function matching across diverse compilation settings.Although adversarial examples can strategically undermine the accuracy of BCSD models and protect critical code,existing techniques predominantly depend on inserting artificial instructions,which incur high computational costs and offer limited diversity of perturbations.To address these limitations,we propose AIMA,a novel gradient-guided assembly instruction relocation method.Our method decouples the detection model into tokenization,embedding,and encoding layers to enable efficient gradient computation.Since token IDs of instructions are discrete and nondifferentiable,we compute gradients in the continuous embedding space to evaluate the influence of each token.The most critical tokens are identified by calculating the L2 norm of their embedding gradients.We then establish a mapping between instructions and their corresponding tokens to aggregate token-level importance into instructionlevel significance.To maximize adversarial impact,a sliding window algorithm selects the most influential contiguous segments for relocation,ensuring optimal perturbation with minimal length.This approach efficiently locates critical code regions without expensive search operations.The selected segments are relocated outside their original function boundaries via a jump mechanism,which preserves runtime control flow and functionality while introducing“deletion”effects in the static instruction sequence.Extensive experiments show that AIMA reduces similarity scores by up to 35.8%in state-of-the-art BCSD models.When incorporated into training data,it also enhances model robustness,achieving a 5.9%improvement in AUROC.展开更多
The ultracold neutron(UCN)transport code,MCUCN,designed initially for simulating UCN transportation from a solid deuterium(SD_2)source and neutron electric dipole moment experiments,could not simulate UCN storage and ...The ultracold neutron(UCN)transport code,MCUCN,designed initially for simulating UCN transportation from a solid deuterium(SD_2)source and neutron electric dipole moment experiments,could not simulate UCN storage and transportation in a superfluid^(4)He(SFHe,He-Ⅱ)source accurately.This limitation arose from the absence of an^(4)He upscattering mechanism and the absorption of^(3)He.And the provided source energy distribution in MCUCN is different from that in SFHe source.This study introduced enhancements to MCUCN to address these constraints,explicitly incorporating the^(4)He upscattering effect,the absorption of^(3)He,the loss caused by impurities on converter wall,UCN source energy distribution in SFHe,and the transmission through negative optical potential.Additionally,a Python-based visualization code for intermediate states and results was developed.To validate these enhancements,we systematically compared the simulation results of the Lujan Center Mark3 UCN system by MCUCN and the improved MCUCN code(iMCUCN)with UCNtransport simulations.Additionally,we compared the results of the SUN1 system simulated by MCUCN and iMCUCN with measurement results.The study demonstrates that iMCUCN effectively simulates the storage and transportation of ultracold neutrons in He-Ⅱ.展开更多
Blind recognition of low-density paritycheck(LDPC)codes has gradually attracted more attention with the development of military and civil communications.However,in the case of the paritycheck matrices with relatively ...Blind recognition of low-density paritycheck(LDPC)codes has gradually attracted more attention with the development of military and civil communications.However,in the case of the paritycheck matrices with relatively high row weights,the existing blind recognition algorithms based on a candidate set generally perform worse.In this paper,we propose a blind recognition method for LDPC codes,called as tangent function assisted least square(TLS)method,which improves recognition performances by constructing a new cost function.To characterize the constraint degree among received vectors and paritycheck vectors,a feature function based on tangent function is constructed in the proposed algorithm.A cost function based on least square method is also established according to the feature function values satisfying the parity-check relationship.Moreover,the minimum average value in TLS is obtained on the candidate set.Numerical analysis and simulation results show that recognition performances of TLS algorithm are consistent with theoretical results.Compared with existing algorithms,the proposed method possesses better recognition performances.展开更多
LargeLanguageModels(LLMs)are increasingly appliedinthe fieldof code translation.However,existing evaluation methodologies suffer from two major limitations:(1)the high overlap between test data and pretraining corpora...LargeLanguageModels(LLMs)are increasingly appliedinthe fieldof code translation.However,existing evaluation methodologies suffer from two major limitations:(1)the high overlap between test data and pretraining corpora,which introduces significant bias in performance evaluation;and(2)mainstream metrics focus primarily on surface-level accuracy,failing to uncover the underlying factors that constrain model capabilities.To address these issues,this paper presents TCode(Translation-Oriented Code Evaluation benchmark)—a complexity-controllable,contamination-free benchmark dataset for code translation—alongside a dedicated static feature sensitivity evaluation framework.The dataset is carefully designed to control complexity along multiple dimensions—including syntactic nesting and expression intricacy—enabling both broad coverage and fine-grained differentiation of sample difficulty.This design supports precise evaluation of model capabilities across a wide spectrum of translation challenges.The proposed evaluation framework introduces a correlation-driven analysis mechanism based on static program features,enabling predictive modeling of translation success from two perspectives:Code Form Complexity(e.g.,code length and character density)and Semantic Modeling Complexity(e.g.,syntactic depth,control-flow nesting,and type system complexity).Empirical evaluations across representative LLMs—including Qwen2.5-72B and Llama3.3-70B—demonstrate that even state-of-the-art models achieve over 80% compilation success on simple samples,but their accuracy drops sharply below 40% on complex cases.Further correlation analysis indicates that Semantic Modeling Complexity alone is correlated with up to 60% of the variance in translation success,with static program features exhibiting nonlinear threshold effects that highlight clear capability boundaries.This study departs fromthe traditional accuracy-centric evaluation paradigm and,for the first time,systematically characterizes the capabilities of large languagemodels in translation tasks through the lens of programstatic features.The findings provide actionable insights for model refinement and training strategy development.展开更多
The rapid growth of distributed data-centric applications and AI workloads increases demand for low-latency,high-throughput communication,necessitating frequent and flexible updates to network routing configurations.H...The rapid growth of distributed data-centric applications and AI workloads increases demand for low-latency,high-throughput communication,necessitating frequent and flexible updates to network routing configurations.However,maintaining consistent forwarding states during these updates is challenging,particularly when rerouting multiple flows simultaneously.Existing approaches pay little attention to multi-flow update,where improper update sequences across data plane nodes may construct deadlock dependencies.Moreover,these methods typically involve excessive control-data plane interactions,incurring significant resource overhead and performance degradation.This paper presents P4LoF,an efficient loop-free update approach that enables the controller to reroute multiple flows through minimal interactions.P4LoF first utilizes a greedy-based algorithm to generate the shortest update dependency chain for the single-flow update.These chains are then dynamically merged into a dependency graph and resolved as a Shortest Common Super-sequence(SCS)problem to produce the update sequence of multi-flow update.To address deadlock dependencies in multi-flow updates,P4LoF builds a deadlock-fix forwarding model that leverages the flexible packet processing capabilities of the programmable data plane.Experimental results show that P4LoF reduces control-data plane interactions by at least 32.6%with modest overhead,while effectively guaranteeing loop-free consistency.展开更多
Automated Program Repair(APR)techniques have shown significant potential in mitigating the cost and complexity associated with debugging by automatically generating corrective patches for software defects.Despite cons...Automated Program Repair(APR)techniques have shown significant potential in mitigating the cost and complexity associated with debugging by automatically generating corrective patches for software defects.Despite considerable progress in APR methodologies,existing approaches frequently lack contextual awareness of runtime behaviors and structural intricacies inherent in buggy source code.In this paper,we propose a novel APR approach that integrates attention mechanisms within an autoencoder-based framework,explicitly utilizing structural code affinity and execution context correlation derived from stack trace analysis.Our approach begins with an innovative preprocessing pipeline,where code segments and stack traces are transformed into tokenized representations.Subsequently,the BM25 ranking algorithm is employed to quantitatively measure structural code affinity and execution context correlation,identifying syntactically and semantically analogous buggy code snippets and relevant runtime error contexts from extensive repositories.These extracted features are then encoded via an attention-enhanced autoencoder model,specifically designed to capture significant patterns and correlations essential for effective patch generation.To assess the efficacy and generalizability of our proposed method,we conducted rigorous experimental comparisons against DeepFix,a state-of-the-art APR system,using a substantial dataset comprising 53,478 studentdeveloped C programs.Experimental outcomes indicate that our model achieves a notable bug repair success rate of approximately 62.36%,representing a statistically significant performance improvement of over 6%compared to the baseline.Furthermore,a thorough K-fold cross-validation reinforced the consistency,robustness,and reliability of our method across diverse subsets of the dataset.Our findings present the critical advantage of integrating attentionbased learning with code structural and execution context features in APR tasks,leading to improved accuracy and practical applicability.Future work aims to extend the model’s applicability across different programming languages,systematically optimize hyperparameters,and explore alternative feature representation methods to further enhance debugging efficiency and effectiveness.展开更多
In erasure-coded storage systems,updating data requires parity maintenance,which often leads to significant I/O amplification due to“write-after-read”operations.Furthermore,scattered parity placement increases disk ...In erasure-coded storage systems,updating data requires parity maintenance,which often leads to significant I/O amplification due to“write-after-read”operations.Furthermore,scattered parity placement increases disk seek overhead during repair,resulting in degraded system performance.To address these challenges,this paper proposes a Cognitive Update and Repair Method(CURM)that leverages machine learning to classify files into writeonly,read-only,and read-write categories,enabling tailored update and repair strategies.For write-only and read-write files,CURM employs a data-differencemechanism combined with fine-grained I/O scheduling to minimize redundant read operations and mitigate I/O amplification.For read-write files,CURM further reserves adjacent disk space near parity blocks,supporting parallel reads and reducing disk seek overhead during repair.We implement CURM in a prototype system,Cognitive Update and Repair File System(CURFS),and conduct extensive experiments using realworld Network File System(NFS)and Microsoft Research(MSR)workloads on a 25-node cluster.Experimental results demonstrate that CURMimproves data update throughput by up to 82.52%,reduces recovery time by up to 47.47%,and decreases long-term storage overhead by more than 15% compared to state-of-the-art methods including Full Logging(FL),ParityLogging(PL),ParityLoggingwithReservedspace(PLR),andPARIX.These results validate the effectiveness of CURM in enhancing both update and repair performance,providing a scalable and efficient solution for large-scale erasure-coded storage systems.展开更多
Recently,the Internet of Things(IoT)technology has been utilized in a wide range of services and applications which significantly transforms digital ecosystems through seamless interconnectivity between various smart ...Recently,the Internet of Things(IoT)technology has been utilized in a wide range of services and applications which significantly transforms digital ecosystems through seamless interconnectivity between various smart devices.Furthermore,the IoT plays a key role in multiple domains,including industrial automation,smart homes,and intelligent transportation systems.However,an increasing number of connected devices presents significant challenges related to efficient resource allocation and system responsiveness.To address these issue,this research proposes a Modified Walrus Optimization Algorithm(MWaOA)for effective resource management in smart IoT systems.In the proposed MWaOA,a crowding process is incorporated to maintain diversity and avoid premature convergence thereby enhancing the global search capability.During resource allocation,the MWaOA prevents early convergence,which aids in achieving a better balance between the exploration and exploitation phases during optimization.Empirical evaluations show that the MWaOA reduces energy consumption by approximately 4% to 34%and minimizes the response time by 6% to 33% across different service arrival rates.Compared to traditional optimization algorithms,MWaOA reduces energy consumption by 5% to 30%and minimizes the response time by 4% to 28% across different simulation epochs.The proposed MWaOA provides adaptive and robust resource allocation,thereby minimizing transmission cost while considering network constraints and real-time performance parameters.展开更多
To shield electronics from complicated electromagnetic environments caused by wireless electromagnetic waves,achieving elaborately structural manufacturing while not sacrificing electromagnetic interference shielding ...To shield electronics from complicated electromagnetic environments caused by wireless electromagnetic waves,achieving elaborately structural manufacturing while not sacrificing electromagnetic interference shielding performances remains crucial challenges.Herein,we propose a hierarchical manufacturing method that combines the use of 3D printing shear flow field and layer-by-layer assembly for fabricating the structurally customizable and multifunctional polylactic acid@graphene nanoparticle(PLA@GNs)materials.The dynamic behavior of polymer fluids is firstly explored via computational fluid dynamic simulation,and a Weissenberg number is employed to quantitatively analyze the disordered-to-ordered structural evolution of molecular chains and nanoparticles,allowing to tailor the micro-scale ordered structures.Subsequently,the macro-scale 3D architectures of PLA@GNs modules are fabricated by layer-by-layer assembly.Owing to the aligned GNs,the shielding performance reaches 41.2 d B,simultaneously accompanied by a directional thermal conductivity of 3.2 W m^(-1)K^(-1).Moreover,the potential application of 3D-printed shielding modules in specific civilian frequency bands such as 4G(1800–2100 MHz),Bluetooth(2402–2480 MHz),and 5G(3300–3800 MHz)is fully demonstrated.Overall,this work not only establishes a universal methodology about 3D printing shear flow field-driven orientation of two-dimensional nanoparticles within polymer fluids,but also gives a scientific method for advanced manufacturing of the next-generation electromagnetic functional modules for smart electronics.展开更多
BACKGROUND Humeral shaft fractures are common and vary by age,with high-energy trauma observed in younger adults and low-impact injuries in older adults.Radial nerve palsy is a frequent complication.Treatment ranges f...BACKGROUND Humeral shaft fractures are common and vary by age,with high-energy trauma observed in younger adults and low-impact injuries in older adults.Radial nerve palsy is a frequent complication.Treatment ranges from nonoperative methods to surgical interventions such as intramedullary K-wires,which promote faster rehabilitation and improved elbow mobility.AIM To evaluate the outcomes of managing humeral shaft fractures using closed reduction and internal fixation with flexible intramedullary K-wires.METHODS This was a retrospective cohort study analyzing the medical records of patients with humeral shaft fractures managed with flexible intramedullary K-wires at King Abdulaziz Medical City,using non-random sampling and descriptive analysis for outcome evaluation.RESULTS This study assessed the clinical outcomes of 20 patients treated for humeral shaft fractures with intramedullary K-wires.Patients were predominantly male(n=16,80%),had an average age of 39.2 years,and a mean body mass index of 29.5 kg/m^(2).The fractures most frequently occurred in the middle third of the humerus(n=14,70%),with oblique fractures being the most common type(n=7,35%).All surgeries used general anesthesia and a posterior approach,with no intraoperative complications reported.Postoperatively,all patients achieved clinical and radiological union(n=20,100%),and the majority(n=13,65%)reached an elbow range of motion from 0 to 150 degrees.CONCLUSION These results suggest that intramedullary K-wire fixation may be an effective option for treating humeral shaft fractures,with favorable outcomes in range of motion recovery,fracture union,and a low rate of intraoperative complications.展开更多
The developing Sixth-Generation(6G)network aims to establish seamless global connectivity for billions of humans,machines,and devices.However,the rich digital service and the explosive heterogeneous connection between...The developing Sixth-Generation(6G)network aims to establish seamless global connectivity for billions of humans,machines,and devices.However,the rich digital service and the explosive heterogeneous connection between various entities in 6G networks can not only induce increasing complications of digital identity management,but also raise material concerns about the security and privacy of the user identity.In this paper,we design a user-centric identity management that returns the sole control to the user self and achieves identity sovereignty toward 6G networks.Specifically,we propose a blockchain-based Identity Management(IDM)architecture for 6G networks,which provides a practical method to secure digital identity management.Subsequently,we develop a fully privacy-preserving identity attribute management scheme by using zero-knowledge proof to protect the privacy-sensitive identity attribute.In particular,the scheme achieves an identity attribute hiding and verification protocol to support users in obtaining and applying their identity attributes without revealing concrete data.Finally,we analyze the security of the proposed architecture and implement a prototype system to evaluate its performance.The results demonstrate that our architecture ensures effective user digital identity management in 6G networks.展开更多
Sustainable energy systems will entail a change in the carbon intensity projections,which should be carried out in a proper manner to facilitate the smooth running of the grid and reduce greenhouse emissions.The prese...Sustainable energy systems will entail a change in the carbon intensity projections,which should be carried out in a proper manner to facilitate the smooth running of the grid and reduce greenhouse emissions.The present article outlines the TransCarbonNet,a novel hybrid deep learning framework with self-attention characteristics added to the bidirectional Long Short-Term Memory(Bi-LSTM)network to forecast the carbon intensity of the grid several days.The proposed temporal fusion model not only learns the local temporal interactions but also the long-term patterns of the carbon emission data;hence,it is able to give suitable forecasts over a period of seven days.TransCarbonNet takes advantage of a multi-head self-attention element to identify significant temporal connections,which means the Bi-LSTM element calculates sequential dependencies in both directions.Massive tests on two actual data sets indicate much improved results in comparison with the existing results,with mean relative errors of 15.3 percent and 12.7 percent,respectively.The framework has given explicable weights of attention that reveal critical periods that influence carbon intensity alterations,and informed decisions on the management of carbon sustainability.The effectiveness of the proposed solution has been validated in numerous cases of operations,and TransCarbonNet is established to be an effective tool when it comes to carbon-friendly optimization of the grid.展开更多
Modern power systems increasingly depend on interconnected microgrids to enhance reliability and renewable energy utilization.However,the high penetration of intermittent renewable sources often causes frequency devia...Modern power systems increasingly depend on interconnected microgrids to enhance reliability and renewable energy utilization.However,the high penetration of intermittent renewable sources often causes frequency deviations,voltage fluctuations,and poor reactive power coordination,posing serious challenges to grid stability.Conventional Interconnection FlowControllers(IFCs)primarily regulate active power flowand fail to effectively handle dynamic frequency variations or reactive power sharing in multi-microgrid networks.To overcome these limitations,this study proposes an enhanced Interconnection Flow Controller(e-IFC)that integrates frequency response balancing and an Interconnection Reactive Power Flow Controller(IRFC)within a unified adaptive control structure.The proposed e-IFC is implemented and analyzed in DIgSILENT PowerFactory to evaluate its performance under various grid disturbances,including frequency drops,load changes,and reactive power fluctuations.Simulation results reveal that the e-IFC achieves 27.4% higher active power sharing accuracy,19.6% lower reactive power deviation,and 18.2% improved frequency stability compared to the conventional IFC.The adaptive controller ensures seamless transitions between grid-connected and islanded modes and maintains stable operation even under communication delays and data noise.Overall,the proposed e-IFCsignificantly enhances active-reactive power coordination and dynamic stability in renewable-integrated multi-microgrid systems.Future research will focus on coupling the e-IFC with tertiary-level optimization frameworks and conducting hardware-in-the-loop validation to enable its application in large-scale smart microgrid environments.展开更多
Digital content such as games,extended reality(XR),and movies has been widely and easily distributed over wireless networks.As a result,unauthorized access,copyright infringement by third parties or eavesdroppers,and ...Digital content such as games,extended reality(XR),and movies has been widely and easily distributed over wireless networks.As a result,unauthorized access,copyright infringement by third parties or eavesdroppers,and cyberattacks over these networks have become pressing concerns.Therefore,protecting copyrighted content and preventing illegal distribution in wireless communications has garnered significant attention.The Intelligent Reflecting Surface(IRS)is regarded as a promising technology for future wireless and mobile networks due to its ability to reconfigure the radio propagation environment.This study investigates the security performance of an uplink Non-Orthogonal Multiple Access(NOMA)system integrated with an IRS and employing Fountain Codes(FCs).Specifically,two users send signals to the base station at separate distances.A relay receives the signal from the nearby user first and then relays it to the base station.The IRS receives the signal from the distant user and reflects it to the relay,which then sends the reflected signal to the base station.Furthermore,a malevolent eavesdropper intercepts both user and relay communications.We construct mathematical equations for Outage Probability(OP),throughput,diversity evaluation,and Interception Probability(IP),offering quantitative insights to assess system security and performance.Additionally,OP and IP are analyzed using a Deep Neural Network(DNN)model.A deeper comprehension of the security performance of the IRS-assisted NOMA systemin signal transmission is provided by Monte Carlo simulations,which are also carried out to confirm the theoretical conclusions.展开更多
Quantum error correction is a technique that enhances a system’s ability to combat noise by encoding logical information into additional quantum bits,which plays a key role in building practical quantum computers.The...Quantum error correction is a technique that enhances a system’s ability to combat noise by encoding logical information into additional quantum bits,which plays a key role in building practical quantum computers.The XZZX surface code,with only one stabilizer generator on each face,demonstrates significant application potential under biased noise.However,the existing minimum weight perfect matching(MWPM)algorithm has high computational complexity and lacks flexibility in large-scale systems.Therefore,this paper proposes a decoding method that combines graph neural networks(GNN)with multi-classifiers,the syndrome is transformed into an undirected graph,and the features are aggregated by convolutional layers,providing a more efficient and accurate decoding strategy.In the experiments,we evaluated the performance of the XZZX code under different biased noise conditions(bias=1,20,200)and different code distances(d=3,5,7,9,11).The experimental results show that under low bias noise(bias=1),the GNN decoder achieves a threshold of 0.18386,an improvement of approximately 19.12%compared to the MWPM decoder.Under high bias noise(bias=200),the GNN decoder reaches a threshold of 0.40542,improving by approximately 20.76%,overcoming the limitations of the conventional decoder.They demonstrate that the GNN decoding method exhibits superior performance and has broad application potential in the error correction of XZZX code.展开更多
This paper proposes a genetic optimization method for the construction of non-binary quasi-cyclic low-density parity-check(NB-QC-LDPC)codes with short block lengths.In our scheme,the initial template base matrices and...This paper proposes a genetic optimization method for the construction of non-binary quasi-cyclic low-density parity-check(NB-QC-LDPC)codes with short block lengths.In our scheme,the initial template base matrices and the corresponding non-binary replacement matrices are constructed by the progressive edge growth algorithm and randomly generated,respectively.The genetic algorithm is then utilized to optimize the base matrices and the replacement ones.The simulation results show that the NB-QC-LDPC codes constructed by the proposed method achieve better decoding performance and lower implementation complexity compared to the existing NB-LDPC codes such as consultative committee for space data system and Bei Dou satellite navigation system.展开更多
基金National Natural Science Foundation of China(Grant No.U2243225)the Strategic Priority Research Program of the Chinese Academy of Sciences(Grant No.XDB40000000)+2 种基金the Natural Science Basic Research Program of Shaanxi(Grant No.Z2024-ZYFS-0065)the Funding of Top Young talents of Ten Thousand talents Plan in China(2021)the Fundamental Research Funds for the Central Universities(Grants No.2452023071 and 2023HHZX002).
文摘Understanding the complex interactions between human activities and ecosystem functions is a prerequisite for achieving sustainable development.Since the implementation of the“Grain for Green”Project in 1999,ecosystem functions in China’s Loess Plateau have significantly improved.However,intensified human activities have also exacerbated the pressures on the region’s fragile ecological environment.This study investigates the spatiotemporal variations in the human activity intensity index(HAI)and net ecosystem benefits(NEB)from 2000 to 2020,using expert-based assessments and an enhanced cost-benefit evaluation framework.Results indicate that HAI increased by 16.7% and 16.6% at the grid and county levels,respectively.NEB exhibited pronounced spatial heterogeneity,with a total increase of USD 36.2 trillion at the grid scale.At the county level,the average NEB rose by 75%.The degree of trade-off was higher at the grid scale than at the county scale,while the synergistic areas initially expanded and then declined at both scales.Key areas for improvement and regions of lagging development were identified as priority zones for ecological management and spatial planning at both spatial resolutions.This study offers scientific insights and practical guidance for harmonizing ecological conservation with high-quality development in ecologically vulnerable regions.
基金the funds from the Key Laboratory of Smart Earth(KF2023YB01-07)Shanghai Collaborative Innovation Fund(XTCX-KJ-2024-17)the National Natural Science Foundation of China(42388102,62303311,and 62231010)。
文摘Differential Code Bias(DCB)is the time delays between two different GNSS signals,which is crucial for GNSS positioning.Previous studies have shown that it can be significantly affected by the flex power operations in satellites.This study proposes a 15-min short-term DCB estimation method to analyze flex power's impact on DCB variations.The method jointly estimates satellite DCB,receiver DCB,and ionospheric parameters using over 300 MGEX stations.We examined three representative flex power events in 2024,achieving average internal RMS values of 0.042 ns and 0.0068 ns for inter-frequency and intra-frequency scenarios respectively.Results show that intra-frequency DCB exhibits clear shift biases synchronized with flex power state transitions while maintaining stability within 0.20 ns during nontransition periods.No definitive impact on inter-frequency DCB was observed at current estimation precision levels.
基金supported by National Natural Science Foundation of China(No.61801327).
文摘Differential pulse-position modulation(DP PM)can achieve a good compromise between power and bandwidth requirements.However,the output sequence has undetectable insertions and deletions.This paper proposes a successive cancellation(SC)decoding scheme based on the weighted levenshtein distance(WLD)of polar codes for correcting insertions/deletions in DPPM systems.In this method,the WLD is used to calculate the transfer probabilities recursively to obtain likelihood ratios,and the low-complexity SC decoding method is built according to the error characteristics to match the DPPM system.Additionally,the proposed SC decoding scheme is extended to list decoding,which can further improve error correction performance.Simulation results show that the proposed scheme can effectively correct insertions/deletions in the DPPM system,which enhances its reliability and performance.
文摘As artificial Intelligence(AI)continues to expand exponentially,particularly with the emergence of generative pre-trained transformers(GPT)based on a transformer’s architecture,which has revolutionized data processing and enabled significant improvements in various applications.This document seeks to investigate the security vulnerabilities detection in the source code using a range of large language models(LLM).Our primary objective is to evaluate the effectiveness of Static Application Security Testing(SAST)by applying various techniques such as prompt persona,structure outputs and zero-shot.To the selection of the LLMs(CodeLlama 7B,DeepSeek coder 7B,Gemini 1.5 Flash,Gemini 2.0 Flash,Mistral 7b Instruct,Phi 38b Mini 128K instruct,Qwen 2.5 coder,StartCoder 27B)with comparison and combination with Find Security Bugs.The evaluation method will involve using a selected dataset containing vulnerabilities,and the results to provide insights for different scenarios according to the software criticality(Business critical,non-critical,minimum effort,best effort)In detail,the main objectives of this study are to investigate if large language models outperform or exceed the capabilities of traditional static analysis tools,if the combining LLMs with Static Application Security Testing(SAST)tools lead to an improvement and the possibility that local machine learning models on a normal computer produce reliable results.Summarizing the most important conclusions of the research,it can be said that while it is true that the results have improved depending on the size of the LLM for business-critical software,the best results have been obtained by SAST analysis.This differs in“NonCritical,”“Best Effort,”and“Minimum Effort”scenarios,where the combination of LLM(Gemini)+SAST has obtained better results.
基金supported by Key Laboratory of Cyberspace Security,Ministry of Education,China。
文摘Transformer-based models have significantly advanced binary code similarity detection(BCSD)by leveraging their semantic encoding capabilities for efficient function matching across diverse compilation settings.Although adversarial examples can strategically undermine the accuracy of BCSD models and protect critical code,existing techniques predominantly depend on inserting artificial instructions,which incur high computational costs and offer limited diversity of perturbations.To address these limitations,we propose AIMA,a novel gradient-guided assembly instruction relocation method.Our method decouples the detection model into tokenization,embedding,and encoding layers to enable efficient gradient computation.Since token IDs of instructions are discrete and nondifferentiable,we compute gradients in the continuous embedding space to evaluate the influence of each token.The most critical tokens are identified by calculating the L2 norm of their embedding gradients.We then establish a mapping between instructions and their corresponding tokens to aggregate token-level importance into instructionlevel significance.To maximize adversarial impact,a sliding window algorithm selects the most influential contiguous segments for relocation,ensuring optimal perturbation with minimal length.This approach efficiently locates critical code regions without expensive search operations.The selected segments are relocated outside their original function boundaries via a jump mechanism,which preserves runtime control flow and functionality while introducing“deletion”effects in the static instruction sequence.Extensive experiments show that AIMA reduces similarity scores by up to 35.8%in state-of-the-art BCSD models.When incorporated into training data,it also enhances model robustness,achieving a 5.9%improvement in AUROC.
基金the National Key R&D Program of China(No.2024YFE0110001)the National Natural Science Foundation of China(U1932219)the Mobility Programme endorsed by the Joint Committee of the Sino-German Center(M0728)。
文摘The ultracold neutron(UCN)transport code,MCUCN,designed initially for simulating UCN transportation from a solid deuterium(SD_2)source and neutron electric dipole moment experiments,could not simulate UCN storage and transportation in a superfluid^(4)He(SFHe,He-Ⅱ)source accurately.This limitation arose from the absence of an^(4)He upscattering mechanism and the absorption of^(3)He.And the provided source energy distribution in MCUCN is different from that in SFHe source.This study introduced enhancements to MCUCN to address these constraints,explicitly incorporating the^(4)He upscattering effect,the absorption of^(3)He,the loss caused by impurities on converter wall,UCN source energy distribution in SFHe,and the transmission through negative optical potential.Additionally,a Python-based visualization code for intermediate states and results was developed.To validate these enhancements,we systematically compared the simulation results of the Lujan Center Mark3 UCN system by MCUCN and the improved MCUCN code(iMCUCN)with UCNtransport simulations.Additionally,we compared the results of the SUN1 system simulated by MCUCN and iMCUCN with measurement results.The study demonstrates that iMCUCN effectively simulates the storage and transportation of ultracold neutrons in He-Ⅱ.
基金Fundamental Research Funds for the Central Universities under Grant 3072025YC0802the National Natural Science Foundation of China under Grant 62001138Heilongjiang Provincial Natural Science Foundation of China under Grant LH2021F009。
文摘Blind recognition of low-density paritycheck(LDPC)codes has gradually attracted more attention with the development of military and civil communications.However,in the case of the paritycheck matrices with relatively high row weights,the existing blind recognition algorithms based on a candidate set generally perform worse.In this paper,we propose a blind recognition method for LDPC codes,called as tangent function assisted least square(TLS)method,which improves recognition performances by constructing a new cost function.To characterize the constraint degree among received vectors and paritycheck vectors,a feature function based on tangent function is constructed in the proposed algorithm.A cost function based on least square method is also established according to the feature function values satisfying the parity-check relationship.Moreover,the minimum average value in TLS is obtained on the candidate set.Numerical analysis and simulation results show that recognition performances of TLS algorithm are consistent with theoretical results.Compared with existing algorithms,the proposed method possesses better recognition performances.
文摘LargeLanguageModels(LLMs)are increasingly appliedinthe fieldof code translation.However,existing evaluation methodologies suffer from two major limitations:(1)the high overlap between test data and pretraining corpora,which introduces significant bias in performance evaluation;and(2)mainstream metrics focus primarily on surface-level accuracy,failing to uncover the underlying factors that constrain model capabilities.To address these issues,this paper presents TCode(Translation-Oriented Code Evaluation benchmark)—a complexity-controllable,contamination-free benchmark dataset for code translation—alongside a dedicated static feature sensitivity evaluation framework.The dataset is carefully designed to control complexity along multiple dimensions—including syntactic nesting and expression intricacy—enabling both broad coverage and fine-grained differentiation of sample difficulty.This design supports precise evaluation of model capabilities across a wide spectrum of translation challenges.The proposed evaluation framework introduces a correlation-driven analysis mechanism based on static program features,enabling predictive modeling of translation success from two perspectives:Code Form Complexity(e.g.,code length and character density)and Semantic Modeling Complexity(e.g.,syntactic depth,control-flow nesting,and type system complexity).Empirical evaluations across representative LLMs—including Qwen2.5-72B and Llama3.3-70B—demonstrate that even state-of-the-art models achieve over 80% compilation success on simple samples,but their accuracy drops sharply below 40% on complex cases.Further correlation analysis indicates that Semantic Modeling Complexity alone is correlated with up to 60% of the variance in translation success,with static program features exhibiting nonlinear threshold effects that highlight clear capability boundaries.This study departs fromthe traditional accuracy-centric evaluation paradigm and,for the first time,systematically characterizes the capabilities of large languagemodels in translation tasks through the lens of programstatic features.The findings provide actionable insights for model refinement and training strategy development.
基金supported by the National Key Research and Development Program of China under Grant 2022YFB2901501in part by the Science and Technology Innovation leading Talents Subsidy Project of Central Plains under Grant 244200510038.
文摘The rapid growth of distributed data-centric applications and AI workloads increases demand for low-latency,high-throughput communication,necessitating frequent and flexible updates to network routing configurations.However,maintaining consistent forwarding states during these updates is challenging,particularly when rerouting multiple flows simultaneously.Existing approaches pay little attention to multi-flow update,where improper update sequences across data plane nodes may construct deadlock dependencies.Moreover,these methods typically involve excessive control-data plane interactions,incurring significant resource overhead and performance degradation.This paper presents P4LoF,an efficient loop-free update approach that enables the controller to reroute multiple flows through minimal interactions.P4LoF first utilizes a greedy-based algorithm to generate the shortest update dependency chain for the single-flow update.These chains are then dynamically merged into a dependency graph and resolved as a Shortest Common Super-sequence(SCS)problem to produce the update sequence of multi-flow update.To address deadlock dependencies in multi-flow updates,P4LoF builds a deadlock-fix forwarding model that leverages the flexible packet processing capabilities of the programmable data plane.Experimental results show that P4LoF reduces control-data plane interactions by at least 32.6%with modest overhead,while effectively guaranteeing loop-free consistency.
文摘Automated Program Repair(APR)techniques have shown significant potential in mitigating the cost and complexity associated with debugging by automatically generating corrective patches for software defects.Despite considerable progress in APR methodologies,existing approaches frequently lack contextual awareness of runtime behaviors and structural intricacies inherent in buggy source code.In this paper,we propose a novel APR approach that integrates attention mechanisms within an autoencoder-based framework,explicitly utilizing structural code affinity and execution context correlation derived from stack trace analysis.Our approach begins with an innovative preprocessing pipeline,where code segments and stack traces are transformed into tokenized representations.Subsequently,the BM25 ranking algorithm is employed to quantitatively measure structural code affinity and execution context correlation,identifying syntactically and semantically analogous buggy code snippets and relevant runtime error contexts from extensive repositories.These extracted features are then encoded via an attention-enhanced autoencoder model,specifically designed to capture significant patterns and correlations essential for effective patch generation.To assess the efficacy and generalizability of our proposed method,we conducted rigorous experimental comparisons against DeepFix,a state-of-the-art APR system,using a substantial dataset comprising 53,478 studentdeveloped C programs.Experimental outcomes indicate that our model achieves a notable bug repair success rate of approximately 62.36%,representing a statistically significant performance improvement of over 6%compared to the baseline.Furthermore,a thorough K-fold cross-validation reinforced the consistency,robustness,and reliability of our method across diverse subsets of the dataset.Our findings present the critical advantage of integrating attentionbased learning with code structural and execution context features in APR tasks,leading to improved accuracy and practical applicability.Future work aims to extend the model’s applicability across different programming languages,systematically optimize hyperparameters,and explore alternative feature representation methods to further enhance debugging efficiency and effectiveness.
基金supported by the National Natural Science Foundation of China(Grant No.62362019)the Natural Science Foundation of Hainan Province(Grant No.624RC482)the Hainan Provincial Higher Education Teaching Reform Research Project(Grant Hnjg2024-27).
文摘In erasure-coded storage systems,updating data requires parity maintenance,which often leads to significant I/O amplification due to“write-after-read”operations.Furthermore,scattered parity placement increases disk seek overhead during repair,resulting in degraded system performance.To address these challenges,this paper proposes a Cognitive Update and Repair Method(CURM)that leverages machine learning to classify files into writeonly,read-only,and read-write categories,enabling tailored update and repair strategies.For write-only and read-write files,CURM employs a data-differencemechanism combined with fine-grained I/O scheduling to minimize redundant read operations and mitigate I/O amplification.For read-write files,CURM further reserves adjacent disk space near parity blocks,supporting parallel reads and reducing disk seek overhead during repair.We implement CURM in a prototype system,Cognitive Update and Repair File System(CURFS),and conduct extensive experiments using realworld Network File System(NFS)and Microsoft Research(MSR)workloads on a 25-node cluster.Experimental results demonstrate that CURMimproves data update throughput by up to 82.52%,reduces recovery time by up to 47.47%,and decreases long-term storage overhead by more than 15% compared to state-of-the-art methods including Full Logging(FL),ParityLogging(PL),ParityLoggingwithReservedspace(PLR),andPARIX.These results validate the effectiveness of CURM in enhancing both update and repair performance,providing a scalable and efficient solution for large-scale erasure-coded storage systems.
文摘Recently,the Internet of Things(IoT)technology has been utilized in a wide range of services and applications which significantly transforms digital ecosystems through seamless interconnectivity between various smart devices.Furthermore,the IoT plays a key role in multiple domains,including industrial automation,smart homes,and intelligent transportation systems.However,an increasing number of connected devices presents significant challenges related to efficient resource allocation and system responsiveness.To address these issue,this research proposes a Modified Walrus Optimization Algorithm(MWaOA)for effective resource management in smart IoT systems.In the proposed MWaOA,a crowding process is incorporated to maintain diversity and avoid premature convergence thereby enhancing the global search capability.During resource allocation,the MWaOA prevents early convergence,which aids in achieving a better balance between the exploration and exploitation phases during optimization.Empirical evaluations show that the MWaOA reduces energy consumption by approximately 4% to 34%and minimizes the response time by 6% to 33% across different service arrival rates.Compared to traditional optimization algorithms,MWaOA reduces energy consumption by 5% to 30%and minimizes the response time by 4% to 28% across different simulation epochs.The proposed MWaOA provides adaptive and robust resource allocation,thereby minimizing transmission cost while considering network constraints and real-time performance parameters.
基金financially supported by the National Natural Science Foundation of China(52303036)the Natural Science Foundation of Guangxi(2024GXNSFBA010123)+2 种基金the International Science&Technology Innovation Cooperation Project of Sichuan Province(2024YFHZ0232)the International Science&Technology Cooperation Project of Chengdu(2021-GH03-00009-HZ)the Opening Project of State Key Laboratory of Polymer Materials Engineering(Sichuan University)(Sklpme2023-3-18)。
文摘To shield electronics from complicated electromagnetic environments caused by wireless electromagnetic waves,achieving elaborately structural manufacturing while not sacrificing electromagnetic interference shielding performances remains crucial challenges.Herein,we propose a hierarchical manufacturing method that combines the use of 3D printing shear flow field and layer-by-layer assembly for fabricating the structurally customizable and multifunctional polylactic acid@graphene nanoparticle(PLA@GNs)materials.The dynamic behavior of polymer fluids is firstly explored via computational fluid dynamic simulation,and a Weissenberg number is employed to quantitatively analyze the disordered-to-ordered structural evolution of molecular chains and nanoparticles,allowing to tailor the micro-scale ordered structures.Subsequently,the macro-scale 3D architectures of PLA@GNs modules are fabricated by layer-by-layer assembly.Owing to the aligned GNs,the shielding performance reaches 41.2 d B,simultaneously accompanied by a directional thermal conductivity of 3.2 W m^(-1)K^(-1).Moreover,the potential application of 3D-printed shielding modules in specific civilian frequency bands such as 4G(1800–2100 MHz),Bluetooth(2402–2480 MHz),and 5G(3300–3800 MHz)is fully demonstrated.Overall,this work not only establishes a universal methodology about 3D printing shear flow field-driven orientation of two-dimensional nanoparticles within polymer fluids,but also gives a scientific method for advanced manufacturing of the next-generation electromagnetic functional modules for smart electronics.
基金approved by King Abdullah International Medical Research Center Ethics Committee(approval No.0000074524).
文摘BACKGROUND Humeral shaft fractures are common and vary by age,with high-energy trauma observed in younger adults and low-impact injuries in older adults.Radial nerve palsy is a frequent complication.Treatment ranges from nonoperative methods to surgical interventions such as intramedullary K-wires,which promote faster rehabilitation and improved elbow mobility.AIM To evaluate the outcomes of managing humeral shaft fractures using closed reduction and internal fixation with flexible intramedullary K-wires.METHODS This was a retrospective cohort study analyzing the medical records of patients with humeral shaft fractures managed with flexible intramedullary K-wires at King Abdulaziz Medical City,using non-random sampling and descriptive analysis for outcome evaluation.RESULTS This study assessed the clinical outcomes of 20 patients treated for humeral shaft fractures with intramedullary K-wires.Patients were predominantly male(n=16,80%),had an average age of 39.2 years,and a mean body mass index of 29.5 kg/m^(2).The fractures most frequently occurred in the middle third of the humerus(n=14,70%),with oblique fractures being the most common type(n=7,35%).All surgeries used general anesthesia and a posterior approach,with no intraoperative complications reported.Postoperatively,all patients achieved clinical and radiological union(n=20,100%),and the majority(n=13,65%)reached an elbow range of motion from 0 to 150 degrees.CONCLUSION These results suggest that intramedullary K-wire fixation may be an effective option for treating humeral shaft fractures,with favorable outcomes in range of motion recovery,fracture union,and a low rate of intraoperative complications.
文摘The developing Sixth-Generation(6G)network aims to establish seamless global connectivity for billions of humans,machines,and devices.However,the rich digital service and the explosive heterogeneous connection between various entities in 6G networks can not only induce increasing complications of digital identity management,but also raise material concerns about the security and privacy of the user identity.In this paper,we design a user-centric identity management that returns the sole control to the user self and achieves identity sovereignty toward 6G networks.Specifically,we propose a blockchain-based Identity Management(IDM)architecture for 6G networks,which provides a practical method to secure digital identity management.Subsequently,we develop a fully privacy-preserving identity attribute management scheme by using zero-knowledge proof to protect the privacy-sensitive identity attribute.In particular,the scheme achieves an identity attribute hiding and verification protocol to support users in obtaining and applying their identity attributes without revealing concrete data.Finally,we analyze the security of the proposed architecture and implement a prototype system to evaluate its performance.The results demonstrate that our architecture ensures effective user digital identity management in 6G networks.
基金funded by the Deanship of Scientific Research and Libraries at Princess Nourah bint Abdulrahman University,through the“Nafea”Program,Grant No.(NP-45-082).
文摘Sustainable energy systems will entail a change in the carbon intensity projections,which should be carried out in a proper manner to facilitate the smooth running of the grid and reduce greenhouse emissions.The present article outlines the TransCarbonNet,a novel hybrid deep learning framework with self-attention characteristics added to the bidirectional Long Short-Term Memory(Bi-LSTM)network to forecast the carbon intensity of the grid several days.The proposed temporal fusion model not only learns the local temporal interactions but also the long-term patterns of the carbon emission data;hence,it is able to give suitable forecasts over a period of seven days.TransCarbonNet takes advantage of a multi-head self-attention element to identify significant temporal connections,which means the Bi-LSTM element calculates sequential dependencies in both directions.Massive tests on two actual data sets indicate much improved results in comparison with the existing results,with mean relative errors of 15.3 percent and 12.7 percent,respectively.The framework has given explicable weights of attention that reveal critical periods that influence carbon intensity alterations,and informed decisions on the management of carbon sustainability.The effectiveness of the proposed solution has been validated in numerous cases of operations,and TransCarbonNet is established to be an effective tool when it comes to carbon-friendly optimization of the grid.
基金the Deanship of Scientific Research at Northern Border University,Arar,Saudi Arabia,for funding this research work through the project number“NBU-FFR-2025-3623-11”.
文摘Modern power systems increasingly depend on interconnected microgrids to enhance reliability and renewable energy utilization.However,the high penetration of intermittent renewable sources often causes frequency deviations,voltage fluctuations,and poor reactive power coordination,posing serious challenges to grid stability.Conventional Interconnection FlowControllers(IFCs)primarily regulate active power flowand fail to effectively handle dynamic frequency variations or reactive power sharing in multi-microgrid networks.To overcome these limitations,this study proposes an enhanced Interconnection Flow Controller(e-IFC)that integrates frequency response balancing and an Interconnection Reactive Power Flow Controller(IRFC)within a unified adaptive control structure.The proposed e-IFC is implemented and analyzed in DIgSILENT PowerFactory to evaluate its performance under various grid disturbances,including frequency drops,load changes,and reactive power fluctuations.Simulation results reveal that the e-IFC achieves 27.4% higher active power sharing accuracy,19.6% lower reactive power deviation,and 18.2% improved frequency stability compared to the conventional IFC.The adaptive controller ensures seamless transitions between grid-connected and islanded modes and maintains stable operation even under communication delays and data noise.Overall,the proposed e-IFCsignificantly enhances active-reactive power coordination and dynamic stability in renewable-integrated multi-microgrid systems.Future research will focus on coupling the e-IFC with tertiary-level optimization frameworks and conducting hardware-in-the-loop validation to enable its application in large-scale smart microgrid environments.
基金supported in part by Vietnam National Foundation for Science and Technology Development(NAFOSTED)under Grant 102.04-2021.57in part by Culture,Sports and Tourism R&D Program through the Korea Creative Content Agency grant funded by the Ministry of Culture,Sports and Tourism in 2024(Project Name:Global Talent Training Program for Copyright Management Technology in Game Contents,Project Number:RS-2024-00396709,Contribution Rate:100%).
文摘Digital content such as games,extended reality(XR),and movies has been widely and easily distributed over wireless networks.As a result,unauthorized access,copyright infringement by third parties or eavesdroppers,and cyberattacks over these networks have become pressing concerns.Therefore,protecting copyrighted content and preventing illegal distribution in wireless communications has garnered significant attention.The Intelligent Reflecting Surface(IRS)is regarded as a promising technology for future wireless and mobile networks due to its ability to reconfigure the radio propagation environment.This study investigates the security performance of an uplink Non-Orthogonal Multiple Access(NOMA)system integrated with an IRS and employing Fountain Codes(FCs).Specifically,two users send signals to the base station at separate distances.A relay receives the signal from the nearby user first and then relays it to the base station.The IRS receives the signal from the distant user and reflects it to the relay,which then sends the reflected signal to the base station.Furthermore,a malevolent eavesdropper intercepts both user and relay communications.We construct mathematical equations for Outage Probability(OP),throughput,diversity evaluation,and Interception Probability(IP),offering quantitative insights to assess system security and performance.Additionally,OP and IP are analyzed using a Deep Neural Network(DNN)model.A deeper comprehension of the security performance of the IRS-assisted NOMA systemin signal transmission is provided by Monte Carlo simulations,which are also carried out to confirm the theoretical conclusions.
基金supported by the Natural Science Foundation of Shandong Province,China(Grant No.ZR2021MF049)the Joint Fund of Natural Science Foundation of Shandong Province,China(Grant Nos.ZR2022LL.Z012 and ZR2021LLZ001)the Key Research and Development Program of Shandong Province,China(Grant No.2023CXGC010901).
文摘Quantum error correction is a technique that enhances a system’s ability to combat noise by encoding logical information into additional quantum bits,which plays a key role in building practical quantum computers.The XZZX surface code,with only one stabilizer generator on each face,demonstrates significant application potential under biased noise.However,the existing minimum weight perfect matching(MWPM)algorithm has high computational complexity and lacks flexibility in large-scale systems.Therefore,this paper proposes a decoding method that combines graph neural networks(GNN)with multi-classifiers,the syndrome is transformed into an undirected graph,and the features are aggregated by convolutional layers,providing a more efficient and accurate decoding strategy.In the experiments,we evaluated the performance of the XZZX code under different biased noise conditions(bias=1,20,200)and different code distances(d=3,5,7,9,11).The experimental results show that under low bias noise(bias=1),the GNN decoder achieves a threshold of 0.18386,an improvement of approximately 19.12%compared to the MWPM decoder.Under high bias noise(bias=200),the GNN decoder reaches a threshold of 0.40542,improving by approximately 20.76%,overcoming the limitations of the conventional decoder.They demonstrate that the GNN decoding method exhibits superior performance and has broad application potential in the error correction of XZZX code.
基金supported by the National Natural Science Foundation of China under Grants 62331002。
文摘This paper proposes a genetic optimization method for the construction of non-binary quasi-cyclic low-density parity-check(NB-QC-LDPC)codes with short block lengths.In our scheme,the initial template base matrices and the corresponding non-binary replacement matrices are constructed by the progressive edge growth algorithm and randomly generated,respectively.The genetic algorithm is then utilized to optimize the base matrices and the replacement ones.The simulation results show that the NB-QC-LDPC codes constructed by the proposed method achieve better decoding performance and lower implementation complexity compared to the existing NB-LDPC codes such as consultative committee for space data system and Bei Dou satellite navigation system.