Accurately counting dense objects in complex and diverse backgrounds is a significant challenge in computer vision,with applications ranging from crowd counting to various other object counting tasks.To address this,w...Accurately counting dense objects in complex and diverse backgrounds is a significant challenge in computer vision,with applications ranging from crowd counting to various other object counting tasks.To address this,we propose HUANNet(High-Resolution Unified Attention Network),a convolutional neural network designed to capture both local features and rich semantic information through a high-resolution representation learning framework,while optimizing computational distribution across parallel branches.HUANNet introduces three core modules:the High-Resolution Attention Module(HRAM),which enhances feature extraction by optimizing multiresolution feature fusion;the Unified Multi-Scale Attention Module(UMAM),which integrates spatial,channel,and convolutional kernel information through an attention mechanism applied across multiple levels of the network;and the Grid-Assisted Point Matching Module(GPMM),which stabilizes and improves point-to-point matching by leveraging grid-based mechanisms.Extensive experiments show that HUANNet achieves competitive results on the ShanghaiTech Part A/B crowd counting datasets and sets new state-of-the-art performance on dense object counting datasets such as CARPK and XRAY-IECCD,demonstrating the effectiveness and versatility of HUANNet.展开更多
Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrain...Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.展开更多
The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous c...The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous computing networks.Finding an optimal computational resource for task offloading and then executing efficiently is a critical issue to achieve a trade-off between energy consumption and transmission delay.In this network,the task processed at fog nodes reduces transmission delay.Still,it increases energy consumption,while routing tasks to the cloud server saves energy at the cost of higher communication delay.Moreover,the order in which offloaded tasks are executed affects the system’s efficiency.For instance,executing lower-priority tasks before higher-priority jobs can disturb the reliability and stability of the system.Therefore,an efficient strategy of optimal computation offloading and task scheduling is required for operational efficacy.In this paper,we introduced a multi-objective and enhanced version of Cheeta Optimizer(CO),namely(MoECO),to jointly optimize the computation offloading and task scheduling in cloud-fog networks to minimize two competing objectives,i.e.,energy consumption and communication delay.MoECO first assigns tasks to the optimal computational nodes and then the allocated tasks are scheduled for processing based on the task priority.The mathematical modelling of CO needs improvement in computation time and convergence speed.Therefore,MoECO is proposed to increase the search capability of agents by controlling the search strategy based on a leader’s location.The adaptive step length operator is adjusted to diversify the solution and thus improves the exploration phase,i.e.,global search strategy.Consequently,this prevents the algorithm from getting trapped in the local optimal solution.Moreover,the interaction factor during the exploitation phase is also adjusted based on the location of the prey instead of the adjacent Cheetah.This increases the exploitation capability of agents,i.e.,local search capability.Furthermore,MoECO employs a multi-objective Pareto-optimal front to simultaneously minimize designated objectives.Comprehensive simulations in MATLAB demonstrate that the proposed algorithm obtains multiple solutions via a Pareto-optimal front and achieves an efficient trade-off between optimization objectives compared to baseline methods.展开更多
In the field of edge computing,achieving low-latency computational task offloading with limited resources is a critical research challenge,particularly in resource-constrained and latency-sensitive vehicular network e...In the field of edge computing,achieving low-latency computational task offloading with limited resources is a critical research challenge,particularly in resource-constrained and latency-sensitive vehicular network environments where rapid response is mandatory for safety-critical applications.In scenarios where edge servers are sparsely deployed,the lack of coordination and information sharing often leads to load imbalance,thereby increasing system latency.Furthermore,in regions without edge server coverage,tasks must be processed locally,which further exacerbates latency issues.To address these challenges,we propose a novel and efficient Deep Reinforcement Learning(DRL)-based approach aimed at minimizing average task latency.The proposed method incorporates three offloading strategies:local computation,direct offloading to the edge server in local region,and device-to-device(D2D)-assisted offloading to edge servers in other regions.We formulate the task offloading process as a complex latency minimization optimization problem.To solve it,we propose an advanced algorithm based on the Dueling Double Deep Q-Network(D3QN)architecture and incorporating the Prioritized Experience Replay(PER)mechanism.Experimental results demonstrate that,compared with existing offloading algorithms,the proposed method significantly reduces average task latency,enhances user experience,and offers an effective strategy for latency optimization in future edge computing systems under dynamic workloads.展开更多
Background:Diabetic foot,a severe complication of diabetes,is characterized by chronic refractory wounds.Sanhuang Oil,a topical herbal formula,demonstrates significant therapeutic effects including antibacterial,anti-...Background:Diabetic foot,a severe complication of diabetes,is characterized by chronic refractory wounds.Sanhuang Oil,a topical herbal formula,demonstrates significant therapeutic effects including antibacterial,anti-inflammatory,and immunomodulatory activities.However,its active constituents and mechanisms of action against diabetic foot remain to be elucidated.Methods:In this study,the chemical constituents of Sanhuang Oil were identified using UPLC-QE-Orbitrap-MS.Subsequently,the mechanism by which Sanhuang Oil promotes diabetic foot ulcer healing was predicted by integrating network pharmacology and molecular docking.Additionally,diabetic mouse model was established in ICR mice using a combination of a high-fat diet(HFD)and streptozotocin(STZ)chemical induction.A full-thickness skin defect was created on the dorsum of the mice.Wound healing and the healing rate were observed following Sanhuang Oil intervention.The mechanism underlying Sanhuang Oil’s promotion of diabetic ulcer healing was further investigated using transcriptomics and histopathological examination(H&E staining).Results:A total of 97 active ingredients were identified from Sanhuang Oil.Network pharmacology analysis predicted 543 common targets,and Kyoto Encyclopedia of Genes and Genomes(KEGG)pathway enrichment analysis identified 203 relevant pathways.Molecular docking further confirmed high binding affinity(binding energy≤−5.0 kcal/mol)between specific active components in Sanhuang Oil(e.g.,coptisine,phellodendrine,baicalein)and key targets associated with diabetic foot ulcers(e.g.,EGFR,AKT1,STAT3).In vivo experimental results demonstrated that the wound healing rate was significantly higher in Sanhuang Oil-treated groups compared to the model group(P<0.001).HE staining revealed that the high-dose Sanhuang Oil group exhibited more pronounced epithelial tissue coverage over the wound,reduced inflammatory cell infiltration,and increased collagen deposition and fibroblast proliferation.transcriptomic analysis identified Pdk4,Ttn,Csrp3,Actn2,Myoz2,Tnnc2,Myod1,Myog,Myot,and Myf6 as key regulatory proteins involved in promoting wound healing.Conclusion:Sanhuang Oil promotes wound healing in diabetic ulcer mice,potentially by mitigating inflammation and regulating key targets such as Pdk4 to enhance fibroblast function.These findings provide novel insights into the multi-target,multi-pathway mechanism of Sanhuang Oil for treating diabetic foot ulcers.展开更多
The advent of quantum computing poses a significant challenge to traditional cryptographic protocols,particularly those used in SecureMultiparty Computation(MPC),a fundamental cryptographic primitive for privacypreser...The advent of quantum computing poses a significant challenge to traditional cryptographic protocols,particularly those used in SecureMultiparty Computation(MPC),a fundamental cryptographic primitive for privacypreserving computation.Classical MPC relies on cryptographic techniques such as homomorphic encryption,secret sharing,and oblivious transfer,which may become vulnerable in the post-quantum era due to the computational power of quantum adversaries.This study presents a review of 140 peer-reviewed articles published between 2000 and 2025 that used different databases like MDPI,IEEE Explore,Springer,and Elsevier,examining the applications,types,and security issues with the solution of Quantum computing in different fields.This review explores the impact of quantum computing on MPC security,assesses emerging quantum-resistant MPC protocols,and examines hybrid classicalquantum approaches aimed at mitigating quantum threats.We analyze the role of Quantum Key Distribution(QKD),post-quantum cryptography(PQC),and quantum homomorphic encryption in securing multiparty computations.Additionally,we discuss the challenges of scalability,computational efficiency,and practical deployment of quantumsecure MPC frameworks in real-world applications such as privacy-preserving AI,secure blockchain transactions,and confidential data analysis.This review provides insights into the future research directions and open challenges in ensuring secure,scalable,and quantum-resistant multiparty computation.展开更多
The Ordos Basin is a large superimposed hydrocarbon-bearing basin in China,and further research on the sedimentary characteristics and sedimentary evolution of the sequence framework of target layers is of great theor...The Ordos Basin is a large superimposed hydrocarbon-bearing basin in China,and further research on the sedimentary characteristics and sedimentary evolution of the sequence framework of target layers is of great theoretical and practical significance for guiding oil and gas exploration.The sedimentary facies and sedimentary evolution of the high-resolution sequence framework of the Carboniferous Taiyuan Formation in the Hangjinqi area have been systematically analyzed for the first time by drilling,logging and seismic data.The results show that four types of sequence interfaces can be identified in the Taiyuan Formation:regional unconformity surfaces,scour surfaces,lithologic-lithofacies transformation surfaces and flooding surfaces.According to the sedimentary response caused by the upward and downward movements of the base level at different levels,the Taiyuan Formation can be divided into 2 long-term cycles(LSC_(1)-LSC_(2)),4 mid-term cycles(MSC_(1)-MSC4)and 7 short-term cycles(SSC_(1)-SSC7).The long-and mid-term cycles correspond to members T_(1)and T_(2)and layers T_(1)-1,T_(1-2),T_(2-1),and T_(2)-2,respectively.Long-term cycles are dominated by C_(1);mid-term cycles are dominated by C_(1)and C_(2),followed by A2;and short-term cycles are dominated by C_(1),C_(2),A1 and A2.Under the high-resolution sequence stratigraphic framework,the Hangjinqi area underwent a transformation of fan delta and tidal flat depositional systems during the Taiyuan Formation sedimentary period.In the MSC_(1)-MSC_(2)stage,owing to a large-scale paleocontinent,the fan delta sedimentary body,which was limited in scale and scope,developed only in the southeastern corner and gradually transitioned basinward to tidal flat facies.In the MSC3-MSC4 stage,as the paleocontinent continuously decreased and the sedimentary range expanded,fan-delta plain sedimentation began in the study area.Several braided distributary channels with poor connectivity developed on the fan-delta plain,and between them were floodplains and peat swamps.展开更多
Beam-tracking simulations have been extensively utilized in the study of collective beam instabilities in circular accelerators.Traditionally,many simulation codes have relied on central processing unit(CPU)-based met...Beam-tracking simulations have been extensively utilized in the study of collective beam instabilities in circular accelerators.Traditionally,many simulation codes have relied on central processing unit(CPU)-based methods,tracking on a single CPU core,or parallelizing the computation across multiple cores via the message passing interface(MPI).Although these approaches work well for single-bunch tracking,scaling them to multiple bunches significantly increases the computational load,which often necessitates the use of a dedicated multi-CPU cluster.To address this challenge,alternative methods leveraging General-Purpose computing on Graphics Processing Units(GPGPU)have been proposed,enabling tracking studies on a standalone desktop personal computer(PC).However,frequent CPU-GPU interactions,including data transfers and synchronization operations during tracking,can introduce communication overheads,potentially reducing the overall effectiveness of GPU-based computations.In this study,we propose a novel approach that eliminates this overhead by performing the entire tracking simulation process exclusively on the GPU,thereby enabling the simultaneous processing of all bunches and their macro-particles.Specifically,we introduce MBTRACK2-CUDA,a Compute Unified Device Architecture(CUDA)ported version of MBTRACK2,which facilitates efficient tracking of single-and multi-bunch collective effects by leveraging the full GPU-resident computation.展开更多
As Internet of Things(IoT)applications expand,Mobile Edge Computing(MEC)has emerged as a promising architecture to overcome the real-time processing limitations of mobile devices.Edge-side computation offloading plays...As Internet of Things(IoT)applications expand,Mobile Edge Computing(MEC)has emerged as a promising architecture to overcome the real-time processing limitations of mobile devices.Edge-side computation offloading plays a pivotal role in MEC performance but remains challenging due to complex task topologies,conflicting objectives,and limited resources.This paper addresses high-dimensional multi-objective offloading for serial heterogeneous tasks in MEC.We jointly consider task heterogeneity,high-dimensional objectives,and flexible resource scheduling,modeling the problem as a Many-objective optimization.To solve it,we propose a flexible framework integrating an improved cooperative co-evolutionary algorithm based on decomposition(MOCC/D)and a flexible scheduling strategy.Experimental results on benchmark functions and simulation scenarios show that the proposed method outperforms existing approaches in both convergence and solution quality.展开更多
Physics-informed neural networks(PINNs)have emerged as a promising class of scientific machine learning techniques that integrate governing physical laws into neural network training.Their ability to enforce different...Physics-informed neural networks(PINNs)have emerged as a promising class of scientific machine learning techniques that integrate governing physical laws into neural network training.Their ability to enforce differential equations,constitutive relations,and boundary conditions within the loss function provides a physically grounded alternative to traditional data-driven models,particularly for solid and structural mechanics,where data are often limited or noisy.This review offers a comprehensive assessment of recent developments in PINNs,combining bibliometric analysis,theoretical foundations,application-oriented insights,and methodological innovations.A biblio-metric survey indicates a rapid increase in publications on PINNs since 2018,with prominent research clusters focused on numerical methods,structural analysis,and forecasting.Building upon this trend,the review consolidates advance-ments across five principal application domains,including forward structural analysis,inverse modeling and parameter identification,structural and topology optimization,assessment of structural integrity,and manufacturing processes.These applications are propelled by substantial methodological advancements,encompassing rigorous enforcement of boundary conditions,modified loss functions,adaptive training,domain decomposition strategies,multi-fidelity and transfer learning approaches,as well as hybrid finite element–PINN integration.These advances address recurring challenges in solid mechanics,such as high-order governing equations,material heterogeneity,complex geometries,localized phenomena,and limited experimental data.Despite remaining challenges in computational cost,scalability,and experimental validation,PINNs are increasingly evolving into specialized,physics-aware tools for practical solid and structural mechanics applications.展开更多
Classical computation of electronic properties in large-scale materials remains challenging.Quantum computation has the potential to offer advantages in memory footprint and computational scaling.However,general and v...Classical computation of electronic properties in large-scale materials remains challenging.Quantum computation has the potential to offer advantages in memory footprint and computational scaling.However,general and viable quantum algorithms for simulating large-scale materials are still limited.We propose and implement random-state quantum algorithms to calculate electronic-structure properties of real materials.Using a random state circuit on a small number of qubits,we employ real-time evolution with first-order Trotter decomposition and Hadamard test to obtain electronic density of states,and we develop a modified quantum phase estimation algorithm to calculate real-space local density of states via direct quantum measurements.Furthermore,we validate these algorithms by numerically computing the density of states and spatial distributions of electronic states in graphene,twisted bilayer graphene quasicrystals,and fractal lattices,covering system sizes from hundreds to thousands of atoms.Our results manifest that the random-state quantum algorithms provide a general and qubit-efficient route to scalable simulations of electronic properties in large-scale periodic and aperiodic materials.展开更多
This paper describes the model speed and model In/Out (I/O) efficiency of the high-resolution atmospheric general circulation model FAMIL (Finite- volume Atmospheric Model of IAP/LASG) at the National Supercompute...This paper describes the model speed and model In/Out (I/O) efficiency of the high-resolution atmospheric general circulation model FAMIL (Finite- volume Atmospheric Model of IAP/LASG) at the National Supercomputer Center in Tianjin, China, on its Tianhe-lA supercomputer platform. A series of three- model-day simulations were carried out with standard Aqua Planet Experiment (APE) designed within FAMIL to obtain the time stamp for the calculation of model speed, simulation cost, and model 1/O efficiency. The results of the simulation demonstrate that FAMIL has remarkable scalability below 3456 and 6144 cores, and the lowest simulation costs are 1536 and 3456 cores for 12.5 km and 6.25 krn resolutions, respectively. Furthermore, FAMIL has excellent I/O scalability and an efficiency of more than 80% on 6 I/Os and more than 99% on 1536 I/Os.展开更多
Mineral dissemination and pore space distribution in ore particles are important features that influence heap leaching performance.To quantify the mineral dissemination and pore space distribution of an ore particle,a...Mineral dissemination and pore space distribution in ore particles are important features that influence heap leaching performance.To quantify the mineral dissemination and pore space distribution of an ore particle,a cylindrical copper oxide ore sample(I center dot 4.6 mm x 5.6 mm)was scanned using high-resolution X-ray computed tomography(HRXCT),a nondestructive imaging technology,at a spatial resolution of 4.85 mu m.Combined with three-dimensional(3D)image analysis techniques,the main mineral phases and pore space were segmented and the volume fraction of each phase was calculated.In addition,the mass fraction of each mineral phase was estimated and the result was validated with that obtained using traditional techniques.Furthermore,the pore phase features,including the pore size distribution,pore surface area,pore fractal dimension,pore centerline,and the pore connectivity,were investigated quantitatively.The pore space analysis results indicate that the pore size distribution closely fits a log-normal distribution and that the pore space morphology is complicated,with a large surface area and low connectivity.This study demonstrates that the combination of HRXCT and 3D image analysis is an effective tool for acquiring 3D mineralogical and pore structural data.展开更多
The purpose of this review is to explore the intersection of computational engineering and biomedical science,highlighting the transformative potential this convergence holds for innovation in healthcare and medical r...The purpose of this review is to explore the intersection of computational engineering and biomedical science,highlighting the transformative potential this convergence holds for innovation in healthcare and medical research.The review covers key topics such as computational modelling,bioinformatics,machine learning in medical diagnostics,and the integration of wearable technology for real-time health monitoring.Major findings indicate that computational models have significantly enhanced the understanding of complex biological systems,while machine learning algorithms have improved the accuracy of disease prediction and diagnosis.The synergy between bioinformatics and computational techniques has led to breakthroughs in personalized medicine,enabling more precise treatment strategies.Additionally,the integration of wearable devices with advanced computational methods has opened new avenues for continuous health monitoring and early disease detection.The review emphasizes the need for interdisciplinary collaboration to further advance this field.Future research should focus on developing more robust and scalable computational models,enhancing data integration techniques,and addressing ethical considerations related to data privacy and security.By fostering innovation at the intersection of these disciplines,the potential to revolutionize healthcare delivery and outcomes becomes increasingly attainable.展开更多
High-resolution transmission electron microscopy(HRTEM)promises rapid atomic-scale dynamic structure imaging.Yet,the precision limitations of aberration parameters and the challenge of eliminating aberrations in Cs-co...High-resolution transmission electron microscopy(HRTEM)promises rapid atomic-scale dynamic structure imaging.Yet,the precision limitations of aberration parameters and the challenge of eliminating aberrations in Cs-corrected transmission electron microscopy constrain resolution.A machine learning algorithm is developed to determine the aberration parameters with higher precision from small,lattice-periodic crystal images.The proposed algorithm is then validated with simulated HRTEM images of graphene and applied to the experimental images of a molybdenum disulfide(MoS_(2))monolayer with 25 variables(14 aberrations)resolved in wide ranges.Using these measured parameters,the phases of the exit-wave functions are reconstructed for each image in a focal series of MoS_(2)monolayers.The images were acquired due to the unexpected movement of the specimen holder.Four-dimensional data extraction reveals time-varying atomic structures and ripple.In particular,the atomic evolution of the sulfur-vacancy point and line defects,as well as the edge structure near the amorphous,is visualized as the resolution has been improved from about 1.75?to 0.9 A.This method can help salvage important transmission electron microscope images and is beneficial for the images obtained from electron microscopes with average stability.展开更多
Low earth orbit(LEO)satellites with wide coverage can carry the mobile edge computing(MEC)servers with powerful computing capabilities to form the LEO satellite edge computing system,providing computing services for t...Low earth orbit(LEO)satellites with wide coverage can carry the mobile edge computing(MEC)servers with powerful computing capabilities to form the LEO satellite edge computing system,providing computing services for the global ground users.In this paper,the computation offloading problem and resource allocation problem are formulated as a mixed integer nonlinear program(MINLP)problem.This paper proposes a computation offloading algorithm based on deep deterministic policy gradient(DDPG)to obtain the user offloading decisions and user uplink transmission power.This paper uses the convex optimization algorithm based on Lagrange multiplier method to obtain the optimal MEC server resource allocation scheme.In addition,the expression of suboptimal user local CPU cycles is derived by relaxation method.Simulation results show that the proposed algorithm can achieve excellent convergence effect,and the proposed algorithm significantly reduces the system utility values at considerable time cost compared with other algorithms.展开更多
Background and objective Atypical 'cardiac' chest pain (ACCP) is not usually caused by myocardial ischaemia. Current noninvasive investigations for these symptoms are not yet as accurate as invasive coronary a...Background and objective Atypical 'cardiac' chest pain (ACCP) is not usually caused by myocardial ischaemia. Current noninvasive investigations for these symptoms are not yet as accurate as invasive coronary angiography. The latest 64-row multi-detector computed tomography (MDCT) technology is non-invasive, has high specificity and negative predictive values for the detection of significant coronary disease. Our aim was to investigate if this modality can provide more information in the assessment of outpatients with ACCP in addition to established cardiovascular risk scores. Methods Seventy consecutive patients presenting to the outpatient clinic with ACCP underwent 64-row MDCT scan of the coronary arteries. They were categorized into low, medium or high risk groups based upon the Framingham and PROCAM scores. We defined a clinically abnormal MDCT scan as coronary stenosis =50% or calcium score >400 Agatston. Results Fifty-three (75.7%) patients did not have clinically abnormal scans. Framingham score classified 43 patients as low-risk while PROCAM classified 59 patients as low-risk. MDCT scans were abnormal for 18.6% and 22.0% of the respective low-risk group of patients. For patients with medium-to-high risk, 33.3% and 36.4% of Framingham and PROCAM patient groups respectively had abnormal MDCT scans. Conclusion MDCT adds valuable information in the assessment of patients with ACCP by identifying a significant proportion of patients categorized as low-risk to have underlying significant coronary stenosis and coronary calcification by established cardiovascular risk scores.展开更多
AIM To compare high-resolution computed tomography(HRCT) findings between humoral primary immunodeficiencies(hPIDs) subtypes; to correlate these findings to pulmonary function tests(PFTs).METHODS We retrospectively id...AIM To compare high-resolution computed tomography(HRCT) findings between humoral primary immunodeficiencies(hPIDs) subtypes; to correlate these findings to pulmonary function tests(PFTs).METHODS We retrospectively identified 52 consecutive adult patients with hPIDs who underwent 64-row HRCT and PFTs at the time of diagnosis. On a per-patient basis, an experienced radiologist recorded airway abnormalities(bronchiectasis,airway wall thickening, mucus plugging, tree-in-bud, and air-trapping) and parenchymal-interstitial abnormalities(consolidations, ground-glass opacities,linear and/or irregular opacities, nodules, and bullae/cysts) found on HRCT.The chi-square test was performed to compare the prevalence of each abnormality among patients with different subtypes of hPIDs. Overall logistic regression analysis was performed to assess whether HRCT findings predicted obstructive and/or restrictive PFTs results(absent-to-mild vs moderate-tosevere).RESULTS Thirty-eight of the 52 patients with hPIDs showed common variable immunodeficiency disorders(CVID), while the remaining 14 had CVID-like conditions(i.e., 11 had isolated IgG subclass deficiencies and 3 had selective IgA deficiencies). The prevalence of most HRCT abnormalities was not significantly different between CVID and CVID-like patients(P > 0.05), except for linear and/or irregular opacities(prevalence of 31.6% in the CVID group and 0 in the CVID-like group; P = 0.0427). Airway wall thickening was the most frequent HRCT abnormality found in both CVID and CVID-like patients(71% of cases in both groups). The presence of tree-in-bud abnormalities was an independent predictor of moderate-to-severe obstructive defects at PFTs(Odds Ratio, OR, of 18.75, P < 0.05), while the presence of linear and/or irregular opacities was an independent predictor of restrictive defects at PFTs(OR = 13.00; P < 0.05).CONCLUSION CVID and CVID-like patients showed similar HRCT findings. Tree-in-bud and linear and/or irregular opacities predicted higher risks of, respectively,obstructive and restrictive defects at PFTs.展开更多
In 6th Generation Mobile Networks(6G),the Space-Integrated-Ground(SIG)Radio Access Network(RAN)promises seamless coverage and exceptionally high Quality of Service(QoS)for diverse services.However,achieving this neces...In 6th Generation Mobile Networks(6G),the Space-Integrated-Ground(SIG)Radio Access Network(RAN)promises seamless coverage and exceptionally high Quality of Service(QoS)for diverse services.However,achieving this necessitates effective management of computation and wireless resources tailored to the requirements of various services.The heterogeneity of computation resources and interference among shared wireless resources pose significant coordination and management challenges.To solve these problems,this work provides an overview of multi-dimensional resource management in 6G SIG RAN,including computation and wireless resource.Firstly it provides with a review of current investigations on computation and wireless resource management and an analysis of existing deficiencies and challenges.Then focusing on the provided challenges,the work proposes an MEC-based computation resource management scheme and a mixed numerology-based wireless resource management scheme.Furthermore,it outlines promising future technologies,including joint model-driven and data-driven resource management technology,and blockchain-based resource management technology within the 6G SIG network.The work also highlights remaining challenges,such as reducing communication costs associated with unstable ground-to-satellite links and overcoming barriers posed by spectrum isolation.Overall,this comprehensive approach aims to pave the way for efficient and effective resource management in future 6G networks.展开更多
The Literary Lab at Stanford University is one of the birthplaces of digital humanities and has maintained significant influence in this field over the years.Professor Hui Haifeng has been engaged in research on digit...The Literary Lab at Stanford University is one of the birthplaces of digital humanities and has maintained significant influence in this field over the years.Professor Hui Haifeng has been engaged in research on digital humanities and computational criticism in recent years.During his visiting scholarship at Stanford University,he participated in the activities of the Literary Lab.Taking this opportunity,he interviewed Professor Mark Algee-Hewitt,the director of the Literary Lab,discussing important topics such as the current state and reception of DH(digital humanities)in the English Department,the operations of the Literary Lab,and the landscape of computational criticism.Mark Algee-Hewitt's research focuses on the eighteenth and early nineteenth centuries in England and Germany and seeks to combine literary criticism with digital and quantitative analyses of literary texts.In particular,he is interested in the history of aesthetic theory and the development and transmission of aesthetic and philosophical concepts during the Enlightenment and Romantic periods.He is also interested in the relationship between aesthetic theory and the poetry of the long eighteenth century.Although his primary background is English literature,he also has a degree in computer science.He believes that the influence of digital humanities within the humanities disciplines is growing increasingly significant.This impact is evident in both the attraction and assistance it offers to students,as well as in the new interpretations it brings to traditional literary studies.He argues that the key to effectively integrating digital humanities into the English Department is to focus on literary research questions,exploring how digital tools can raise new questions or provide new insights into traditional research.展开更多
基金funded by the National Natural Science Foundation of China(62273213,62472262,62572287)Natural Science Foundation of Shandong Province(ZR2024MF144)+1 种基金Natural Science Foundation of Shandong Province for Innovation and Development Joint Funds(ZR2022LZH001)Taishan Scholarship Construction Engineering.
文摘Accurately counting dense objects in complex and diverse backgrounds is a significant challenge in computer vision,with applications ranging from crowd counting to various other object counting tasks.To address this,we propose HUANNet(High-Resolution Unified Attention Network),a convolutional neural network designed to capture both local features and rich semantic information through a high-resolution representation learning framework,while optimizing computational distribution across parallel branches.HUANNet introduces three core modules:the High-Resolution Attention Module(HRAM),which enhances feature extraction by optimizing multiresolution feature fusion;the Unified Multi-Scale Attention Module(UMAM),which integrates spatial,channel,and convolutional kernel information through an attention mechanism applied across multiple levels of the network;and the Grid-Assisted Point Matching Module(GPMM),which stabilizes and improves point-to-point matching by leveraging grid-based mechanisms.Extensive experiments show that HUANNet achieves competitive results on the ShanghaiTech Part A/B crowd counting datasets and sets new state-of-the-art performance on dense object counting datasets such as CARPK and XRAY-IECCD,demonstrating the effectiveness and versatility of HUANNet.
基金supported by Key Science and Technology Program of Henan Province,China(Grant Nos.242102210147,242102210027)Fujian Province Young and Middle aged Teacher Education Research Project(Science and Technology Category)(No.JZ240101)(Corresponding author:Dong Yuan).
文摘Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.
基金appreciation to the Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R384)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous computing networks.Finding an optimal computational resource for task offloading and then executing efficiently is a critical issue to achieve a trade-off between energy consumption and transmission delay.In this network,the task processed at fog nodes reduces transmission delay.Still,it increases energy consumption,while routing tasks to the cloud server saves energy at the cost of higher communication delay.Moreover,the order in which offloaded tasks are executed affects the system’s efficiency.For instance,executing lower-priority tasks before higher-priority jobs can disturb the reliability and stability of the system.Therefore,an efficient strategy of optimal computation offloading and task scheduling is required for operational efficacy.In this paper,we introduced a multi-objective and enhanced version of Cheeta Optimizer(CO),namely(MoECO),to jointly optimize the computation offloading and task scheduling in cloud-fog networks to minimize two competing objectives,i.e.,energy consumption and communication delay.MoECO first assigns tasks to the optimal computational nodes and then the allocated tasks are scheduled for processing based on the task priority.The mathematical modelling of CO needs improvement in computation time and convergence speed.Therefore,MoECO is proposed to increase the search capability of agents by controlling the search strategy based on a leader’s location.The adaptive step length operator is adjusted to diversify the solution and thus improves the exploration phase,i.e.,global search strategy.Consequently,this prevents the algorithm from getting trapped in the local optimal solution.Moreover,the interaction factor during the exploitation phase is also adjusted based on the location of the prey instead of the adjacent Cheetah.This increases the exploitation capability of agents,i.e.,local search capability.Furthermore,MoECO employs a multi-objective Pareto-optimal front to simultaneously minimize designated objectives.Comprehensive simulations in MATLAB demonstrate that the proposed algorithm obtains multiple solutions via a Pareto-optimal front and achieves an efficient trade-off between optimization objectives compared to baseline methods.
基金supported by the National Natural Science Foundation of China(62202215)Liaoning Province Applied Basic Research Program(Youth Special Project,2023JH2/101600038)+4 种基金Shenyang Youth Science and Technology Innovation Talent Support Program(RC220458)Guangxuan Program of Shenyang Ligong University(SYLUGXRC202216)the Basic Research Special Funds for Undergraduate Universities in Liaoning Province(LJ212410144067)the Natural Science Foundation of Liaoning Province(2024-MS-113)the science and technology funds from Liaoning Education Department(LJKZ0242).
文摘In the field of edge computing,achieving low-latency computational task offloading with limited resources is a critical research challenge,particularly in resource-constrained and latency-sensitive vehicular network environments where rapid response is mandatory for safety-critical applications.In scenarios where edge servers are sparsely deployed,the lack of coordination and information sharing often leads to load imbalance,thereby increasing system latency.Furthermore,in regions without edge server coverage,tasks must be processed locally,which further exacerbates latency issues.To address these challenges,we propose a novel and efficient Deep Reinforcement Learning(DRL)-based approach aimed at minimizing average task latency.The proposed method incorporates three offloading strategies:local computation,direct offloading to the edge server in local region,and device-to-device(D2D)-assisted offloading to edge servers in other regions.We formulate the task offloading process as a complex latency minimization optimization problem.To solve it,we propose an advanced algorithm based on the Dueling Double Deep Q-Network(D3QN)architecture and incorporating the Prioritized Experience Replay(PER)mechanism.Experimental results demonstrate that,compared with existing offloading algorithms,the proposed method significantly reduces average task latency,enhances user experience,and offers an effective strategy for latency optimization in future edge computing systems under dynamic workloads.
基金supported by the Natural Science Foundation of Hubei Provincial Department of Education(D20232101)Shandong Second Medical University 2024 Affiliated Hospital(Teaching Hospital)Scientific Research Development Fund Project(2024FYQ026)+3 种基金the innovative Research Programme of Xiangyang No.1 People’s Hospital(XYY2023ZY01)Faculty Development Grants of Xiangyang No.1 People’s Hospital Affiliated to Hubei University of Medicine(XYY2023D05)Joint supported by Hubei Provincial Natural Science Foundation and Xiangyang of China(2025AFD091)Traditional Chinese Medicine Scientific Research Project of Hubei Provincial Administration of Traditional Chinese Medicine(ZY2025D019).
文摘Background:Diabetic foot,a severe complication of diabetes,is characterized by chronic refractory wounds.Sanhuang Oil,a topical herbal formula,demonstrates significant therapeutic effects including antibacterial,anti-inflammatory,and immunomodulatory activities.However,its active constituents and mechanisms of action against diabetic foot remain to be elucidated.Methods:In this study,the chemical constituents of Sanhuang Oil were identified using UPLC-QE-Orbitrap-MS.Subsequently,the mechanism by which Sanhuang Oil promotes diabetic foot ulcer healing was predicted by integrating network pharmacology and molecular docking.Additionally,diabetic mouse model was established in ICR mice using a combination of a high-fat diet(HFD)and streptozotocin(STZ)chemical induction.A full-thickness skin defect was created on the dorsum of the mice.Wound healing and the healing rate were observed following Sanhuang Oil intervention.The mechanism underlying Sanhuang Oil’s promotion of diabetic ulcer healing was further investigated using transcriptomics and histopathological examination(H&E staining).Results:A total of 97 active ingredients were identified from Sanhuang Oil.Network pharmacology analysis predicted 543 common targets,and Kyoto Encyclopedia of Genes and Genomes(KEGG)pathway enrichment analysis identified 203 relevant pathways.Molecular docking further confirmed high binding affinity(binding energy≤−5.0 kcal/mol)between specific active components in Sanhuang Oil(e.g.,coptisine,phellodendrine,baicalein)and key targets associated with diabetic foot ulcers(e.g.,EGFR,AKT1,STAT3).In vivo experimental results demonstrated that the wound healing rate was significantly higher in Sanhuang Oil-treated groups compared to the model group(P<0.001).HE staining revealed that the high-dose Sanhuang Oil group exhibited more pronounced epithelial tissue coverage over the wound,reduced inflammatory cell infiltration,and increased collagen deposition and fibroblast proliferation.transcriptomic analysis identified Pdk4,Ttn,Csrp3,Actn2,Myoz2,Tnnc2,Myod1,Myog,Myot,and Myf6 as key regulatory proteins involved in promoting wound healing.Conclusion:Sanhuang Oil promotes wound healing in diabetic ulcer mice,potentially by mitigating inflammation and regulating key targets such as Pdk4 to enhance fibroblast function.These findings provide novel insights into the multi-target,multi-pathway mechanism of Sanhuang Oil for treating diabetic foot ulcers.
文摘The advent of quantum computing poses a significant challenge to traditional cryptographic protocols,particularly those used in SecureMultiparty Computation(MPC),a fundamental cryptographic primitive for privacypreserving computation.Classical MPC relies on cryptographic techniques such as homomorphic encryption,secret sharing,and oblivious transfer,which may become vulnerable in the post-quantum era due to the computational power of quantum adversaries.This study presents a review of 140 peer-reviewed articles published between 2000 and 2025 that used different databases like MDPI,IEEE Explore,Springer,and Elsevier,examining the applications,types,and security issues with the solution of Quantum computing in different fields.This review explores the impact of quantum computing on MPC security,assesses emerging quantum-resistant MPC protocols,and examines hybrid classicalquantum approaches aimed at mitigating quantum threats.We analyze the role of Quantum Key Distribution(QKD),post-quantum cryptography(PQC),and quantum homomorphic encryption in securing multiparty computations.Additionally,we discuss the challenges of scalability,computational efficiency,and practical deployment of quantumsecure MPC frameworks in real-world applications such as privacy-preserving AI,secure blockchain transactions,and confidential data analysis.This review provides insights into the future research directions and open challenges in ensuring secure,scalable,and quantum-resistant multiparty computation.
基金supported by the Fundamental Research Funds for the Liaoning Universities(Grant No.LJ202410166012).
文摘The Ordos Basin is a large superimposed hydrocarbon-bearing basin in China,and further research on the sedimentary characteristics and sedimentary evolution of the sequence framework of target layers is of great theoretical and practical significance for guiding oil and gas exploration.The sedimentary facies and sedimentary evolution of the high-resolution sequence framework of the Carboniferous Taiyuan Formation in the Hangjinqi area have been systematically analyzed for the first time by drilling,logging and seismic data.The results show that four types of sequence interfaces can be identified in the Taiyuan Formation:regional unconformity surfaces,scour surfaces,lithologic-lithofacies transformation surfaces and flooding surfaces.According to the sedimentary response caused by the upward and downward movements of the base level at different levels,the Taiyuan Formation can be divided into 2 long-term cycles(LSC_(1)-LSC_(2)),4 mid-term cycles(MSC_(1)-MSC4)and 7 short-term cycles(SSC_(1)-SSC7).The long-and mid-term cycles correspond to members T_(1)and T_(2)and layers T_(1)-1,T_(1-2),T_(2-1),and T_(2)-2,respectively.Long-term cycles are dominated by C_(1);mid-term cycles are dominated by C_(1)and C_(2),followed by A2;and short-term cycles are dominated by C_(1),C_(2),A1 and A2.Under the high-resolution sequence stratigraphic framework,the Hangjinqi area underwent a transformation of fan delta and tidal flat depositional systems during the Taiyuan Formation sedimentary period.In the MSC_(1)-MSC_(2)stage,owing to a large-scale paleocontinent,the fan delta sedimentary body,which was limited in scale and scope,developed only in the southeastern corner and gradually transitioned basinward to tidal flat facies.In the MSC3-MSC4 stage,as the paleocontinent continuously decreased and the sedimentary range expanded,fan-delta plain sedimentation began in the study area.Several braided distributary channels with poor connectivity developed on the fan-delta plain,and between them were floodplains and peat swamps.
基金supported by the National Research Foundation of Korea(NRF)funded by the Ministry of Science and ICT(MSIT)(No.RS-2022-00143178)the Ministry of Education(MOE)(Nos.2022R1A6A3A13053896 and 2022R1F1A1074616),Republic of Korea.
文摘Beam-tracking simulations have been extensively utilized in the study of collective beam instabilities in circular accelerators.Traditionally,many simulation codes have relied on central processing unit(CPU)-based methods,tracking on a single CPU core,or parallelizing the computation across multiple cores via the message passing interface(MPI).Although these approaches work well for single-bunch tracking,scaling them to multiple bunches significantly increases the computational load,which often necessitates the use of a dedicated multi-CPU cluster.To address this challenge,alternative methods leveraging General-Purpose computing on Graphics Processing Units(GPGPU)have been proposed,enabling tracking studies on a standalone desktop personal computer(PC).However,frequent CPU-GPU interactions,including data transfers and synchronization operations during tracking,can introduce communication overheads,potentially reducing the overall effectiveness of GPU-based computations.In this study,we propose a novel approach that eliminates this overhead by performing the entire tracking simulation process exclusively on the GPU,thereby enabling the simultaneous processing of all bunches and their macro-particles.Specifically,we introduce MBTRACK2-CUDA,a Compute Unified Device Architecture(CUDA)ported version of MBTRACK2,which facilitates efficient tracking of single-and multi-bunch collective effects by leveraging the full GPU-resident computation.
基金supported by Youth Talent Project of Scientific Research Program of Hubei Provincial Department of Education under Grant Q20241809Doctoral Scientific Research Foundation of Hubei University of Automotive Technology under Grant 202404.
文摘As Internet of Things(IoT)applications expand,Mobile Edge Computing(MEC)has emerged as a promising architecture to overcome the real-time processing limitations of mobile devices.Edge-side computation offloading plays a pivotal role in MEC performance but remains challenging due to complex task topologies,conflicting objectives,and limited resources.This paper addresses high-dimensional multi-objective offloading for serial heterogeneous tasks in MEC.We jointly consider task heterogeneity,high-dimensional objectives,and flexible resource scheduling,modeling the problem as a Many-objective optimization.To solve it,we propose a flexible framework integrating an improved cooperative co-evolutionary algorithm based on decomposition(MOCC/D)and a flexible scheduling strategy.Experimental results on benchmark functions and simulation scenarios show that the proposed method outperforms existing approaches in both convergence and solution quality.
基金funded by National Research Council of Thailand(contract No.N42A671047).
文摘Physics-informed neural networks(PINNs)have emerged as a promising class of scientific machine learning techniques that integrate governing physical laws into neural network training.Their ability to enforce differential equations,constitutive relations,and boundary conditions within the loss function provides a physically grounded alternative to traditional data-driven models,particularly for solid and structural mechanics,where data are often limited or noisy.This review offers a comprehensive assessment of recent developments in PINNs,combining bibliometric analysis,theoretical foundations,application-oriented insights,and methodological innovations.A biblio-metric survey indicates a rapid increase in publications on PINNs since 2018,with prominent research clusters focused on numerical methods,structural analysis,and forecasting.Building upon this trend,the review consolidates advance-ments across five principal application domains,including forward structural analysis,inverse modeling and parameter identification,structural and topology optimization,assessment of structural integrity,and manufacturing processes.These applications are propelled by substantial methodological advancements,encompassing rigorous enforcement of boundary conditions,modified loss functions,adaptive training,domain decomposition strategies,multi-fidelity and transfer learning approaches,as well as hybrid finite element–PINN integration.These advances address recurring challenges in solid mechanics,such as high-order governing equations,material heterogeneity,complex geometries,localized phenomena,and limited experimental data.Despite remaining challenges in computational cost,scalability,and experimental validation,PINNs are increasingly evolving into specialized,physics-aware tools for practical solid and structural mechanics applications.
基金supported by the Major Project for the Integration of ScienceEducation and Industry (Grant No.2025ZDZX02)。
文摘Classical computation of electronic properties in large-scale materials remains challenging.Quantum computation has the potential to offer advantages in memory footprint and computational scaling.However,general and viable quantum algorithms for simulating large-scale materials are still limited.We propose and implement random-state quantum algorithms to calculate electronic-structure properties of real materials.Using a random state circuit on a small number of qubits,we employ real-time evolution with first-order Trotter decomposition and Hadamard test to obtain electronic density of states,and we develop a modified quantum phase estimation algorithm to calculate real-space local density of states via direct quantum measurements.Furthermore,we validate these algorithms by numerically computing the density of states and spatial distributions of electronic states in graphene,twisted bilayer graphene quasicrystals,and fractal lattices,covering system sizes from hundreds to thousands of atoms.Our results manifest that the random-state quantum algorithms provide a general and qubit-efficient route to scalable simulations of electronic properties in large-scale periodic and aperiodic materials.
基金supported by the Strategic Priority Research Program of the Chinese Academy of Sciences (Grant No. XDA05110303)the National Basic Research Program of China (973Program, Grant Nos. 2012CB417203 and 2010CB950404)+1 种基金the National High Technology Research and Development Program of China (863 Program, Grant No. 2010AA012305)the National Natural Science Foundation of China (Grant No. 41023002)
文摘This paper describes the model speed and model In/Out (I/O) efficiency of the high-resolution atmospheric general circulation model FAMIL (Finite- volume Atmospheric Model of IAP/LASG) at the National Supercomputer Center in Tianjin, China, on its Tianhe-lA supercomputer platform. A series of three- model-day simulations were carried out with standard Aqua Planet Experiment (APE) designed within FAMIL to obtain the time stamp for the calculation of model speed, simulation cost, and model 1/O efficiency. The results of the simulation demonstrate that FAMIL has remarkable scalability below 3456 and 6144 cores, and the lowest simulation costs are 1536 and 3456 cores for 12.5 km and 6.25 krn resolutions, respectively. Furthermore, FAMIL has excellent I/O scalability and an efficiency of more than 80% on 6 I/Os and more than 99% on 1536 I/Os.
基金financially supported by the National Natural Science Foundation of China(No.51304076)the Natural Science Foundation of Hunan Province,China(No.14JJ4064)
文摘Mineral dissemination and pore space distribution in ore particles are important features that influence heap leaching performance.To quantify the mineral dissemination and pore space distribution of an ore particle,a cylindrical copper oxide ore sample(I center dot 4.6 mm x 5.6 mm)was scanned using high-resolution X-ray computed tomography(HRXCT),a nondestructive imaging technology,at a spatial resolution of 4.85 mu m.Combined with three-dimensional(3D)image analysis techniques,the main mineral phases and pore space were segmented and the volume fraction of each phase was calculated.In addition,the mass fraction of each mineral phase was estimated and the result was validated with that obtained using traditional techniques.Furthermore,the pore phase features,including the pore size distribution,pore surface area,pore fractal dimension,pore centerline,and the pore connectivity,were investigated quantitatively.The pore space analysis results indicate that the pore size distribution closely fits a log-normal distribution and that the pore space morphology is complicated,with a large surface area and low connectivity.This study demonstrates that the combination of HRXCT and 3D image analysis is an effective tool for acquiring 3D mineralogical and pore structural data.
文摘The purpose of this review is to explore the intersection of computational engineering and biomedical science,highlighting the transformative potential this convergence holds for innovation in healthcare and medical research.The review covers key topics such as computational modelling,bioinformatics,machine learning in medical diagnostics,and the integration of wearable technology for real-time health monitoring.Major findings indicate that computational models have significantly enhanced the understanding of complex biological systems,while machine learning algorithms have improved the accuracy of disease prediction and diagnosis.The synergy between bioinformatics and computational techniques has led to breakthroughs in personalized medicine,enabling more precise treatment strategies.Additionally,the integration of wearable devices with advanced computational methods has opened new avenues for continuous health monitoring and early disease detection.The review emphasizes the need for interdisciplinary collaboration to further advance this field.Future research should focus on developing more robust and scalable computational models,enhancing data integration techniques,and addressing ethical considerations related to data privacy and security.By fostering innovation at the intersection of these disciplines,the potential to revolutionize healthcare delivery and outcomes becomes increasingly attainable.
基金financial support from the National Natural Science Foundation of China(Grant No.61971201)。
文摘High-resolution transmission electron microscopy(HRTEM)promises rapid atomic-scale dynamic structure imaging.Yet,the precision limitations of aberration parameters and the challenge of eliminating aberrations in Cs-corrected transmission electron microscopy constrain resolution.A machine learning algorithm is developed to determine the aberration parameters with higher precision from small,lattice-periodic crystal images.The proposed algorithm is then validated with simulated HRTEM images of graphene and applied to the experimental images of a molybdenum disulfide(MoS_(2))monolayer with 25 variables(14 aberrations)resolved in wide ranges.Using these measured parameters,the phases of the exit-wave functions are reconstructed for each image in a focal series of MoS_(2)monolayers.The images were acquired due to the unexpected movement of the specimen holder.Four-dimensional data extraction reveals time-varying atomic structures and ripple.In particular,the atomic evolution of the sulfur-vacancy point and line defects,as well as the edge structure near the amorphous,is visualized as the resolution has been improved from about 1.75?to 0.9 A.This method can help salvage important transmission electron microscope images and is beneficial for the images obtained from electron microscopes with average stability.
基金supported by National Natural Science Foundation of China No.62231012Natural Science Foundation for Outstanding Young Scholars of Heilongjiang Province under Grant YQ2020F001Heilongjiang Province Postdoctoral General Foundation under Grant AUGA4110004923.
文摘Low earth orbit(LEO)satellites with wide coverage can carry the mobile edge computing(MEC)servers with powerful computing capabilities to form the LEO satellite edge computing system,providing computing services for the global ground users.In this paper,the computation offloading problem and resource allocation problem are formulated as a mixed integer nonlinear program(MINLP)problem.This paper proposes a computation offloading algorithm based on deep deterministic policy gradient(DDPG)to obtain the user offloading decisions and user uplink transmission power.This paper uses the convex optimization algorithm based on Lagrange multiplier method to obtain the optimal MEC server resource allocation scheme.In addition,the expression of suboptimal user local CPU cycles is derived by relaxation method.Simulation results show that the proposed algorithm can achieve excellent convergence effect,and the proposed algorithm significantly reduces the system utility values at considerable time cost compared with other algorithms.
文摘Background and objective Atypical 'cardiac' chest pain (ACCP) is not usually caused by myocardial ischaemia. Current noninvasive investigations for these symptoms are not yet as accurate as invasive coronary angiography. The latest 64-row multi-detector computed tomography (MDCT) technology is non-invasive, has high specificity and negative predictive values for the detection of significant coronary disease. Our aim was to investigate if this modality can provide more information in the assessment of outpatients with ACCP in addition to established cardiovascular risk scores. Methods Seventy consecutive patients presenting to the outpatient clinic with ACCP underwent 64-row MDCT scan of the coronary arteries. They were categorized into low, medium or high risk groups based upon the Framingham and PROCAM scores. We defined a clinically abnormal MDCT scan as coronary stenosis =50% or calcium score >400 Agatston. Results Fifty-three (75.7%) patients did not have clinically abnormal scans. Framingham score classified 43 patients as low-risk while PROCAM classified 59 patients as low-risk. MDCT scans were abnormal for 18.6% and 22.0% of the respective low-risk group of patients. For patients with medium-to-high risk, 33.3% and 36.4% of Framingham and PROCAM patient groups respectively had abnormal MDCT scans. Conclusion MDCT adds valuable information in the assessment of patients with ACCP by identifying a significant proportion of patients categorized as low-risk to have underlying significant coronary stenosis and coronary calcification by established cardiovascular risk scores.
基金funded by Department of Medicine, University of Udine (Udine, Italy) in accordance with Good Publication Practice (GPP3) guidelines (http://www.ismpp.org/gpp3)
文摘AIM To compare high-resolution computed tomography(HRCT) findings between humoral primary immunodeficiencies(hPIDs) subtypes; to correlate these findings to pulmonary function tests(PFTs).METHODS We retrospectively identified 52 consecutive adult patients with hPIDs who underwent 64-row HRCT and PFTs at the time of diagnosis. On a per-patient basis, an experienced radiologist recorded airway abnormalities(bronchiectasis,airway wall thickening, mucus plugging, tree-in-bud, and air-trapping) and parenchymal-interstitial abnormalities(consolidations, ground-glass opacities,linear and/or irregular opacities, nodules, and bullae/cysts) found on HRCT.The chi-square test was performed to compare the prevalence of each abnormality among patients with different subtypes of hPIDs. Overall logistic regression analysis was performed to assess whether HRCT findings predicted obstructive and/or restrictive PFTs results(absent-to-mild vs moderate-tosevere).RESULTS Thirty-eight of the 52 patients with hPIDs showed common variable immunodeficiency disorders(CVID), while the remaining 14 had CVID-like conditions(i.e., 11 had isolated IgG subclass deficiencies and 3 had selective IgA deficiencies). The prevalence of most HRCT abnormalities was not significantly different between CVID and CVID-like patients(P > 0.05), except for linear and/or irregular opacities(prevalence of 31.6% in the CVID group and 0 in the CVID-like group; P = 0.0427). Airway wall thickening was the most frequent HRCT abnormality found in both CVID and CVID-like patients(71% of cases in both groups). The presence of tree-in-bud abnormalities was an independent predictor of moderate-to-severe obstructive defects at PFTs(Odds Ratio, OR, of 18.75, P < 0.05), while the presence of linear and/or irregular opacities was an independent predictor of restrictive defects at PFTs(OR = 13.00; P < 0.05).CONCLUSION CVID and CVID-like patients showed similar HRCT findings. Tree-in-bud and linear and/or irregular opacities predicted higher risks of, respectively,obstructive and restrictive defects at PFTs.
基金supported by the National Key Research and Development Program of China(No.2021YFB2900504).
文摘In 6th Generation Mobile Networks(6G),the Space-Integrated-Ground(SIG)Radio Access Network(RAN)promises seamless coverage and exceptionally high Quality of Service(QoS)for diverse services.However,achieving this necessitates effective management of computation and wireless resources tailored to the requirements of various services.The heterogeneity of computation resources and interference among shared wireless resources pose significant coordination and management challenges.To solve these problems,this work provides an overview of multi-dimensional resource management in 6G SIG RAN,including computation and wireless resource.Firstly it provides with a review of current investigations on computation and wireless resource management and an analysis of existing deficiencies and challenges.Then focusing on the provided challenges,the work proposes an MEC-based computation resource management scheme and a mixed numerology-based wireless resource management scheme.Furthermore,it outlines promising future technologies,including joint model-driven and data-driven resource management technology,and blockchain-based resource management technology within the 6G SIG network.The work also highlights remaining challenges,such as reducing communication costs associated with unstable ground-to-satellite links and overcoming barriers posed by spectrum isolation.Overall,this comprehensive approach aims to pave the way for efficient and effective resource management in future 6G networks.
文摘The Literary Lab at Stanford University is one of the birthplaces of digital humanities and has maintained significant influence in this field over the years.Professor Hui Haifeng has been engaged in research on digital humanities and computational criticism in recent years.During his visiting scholarship at Stanford University,he participated in the activities of the Literary Lab.Taking this opportunity,he interviewed Professor Mark Algee-Hewitt,the director of the Literary Lab,discussing important topics such as the current state and reception of DH(digital humanities)in the English Department,the operations of the Literary Lab,and the landscape of computational criticism.Mark Algee-Hewitt's research focuses on the eighteenth and early nineteenth centuries in England and Germany and seeks to combine literary criticism with digital and quantitative analyses of literary texts.In particular,he is interested in the history of aesthetic theory and the development and transmission of aesthetic and philosophical concepts during the Enlightenment and Romantic periods.He is also interested in the relationship between aesthetic theory and the poetry of the long eighteenth century.Although his primary background is English literature,he also has a degree in computer science.He believes that the influence of digital humanities within the humanities disciplines is growing increasingly significant.This impact is evident in both the attraction and assistance it offers to students,as well as in the new interpretations it brings to traditional literary studies.He argues that the key to effectively integrating digital humanities into the English Department is to focus on literary research questions,exploring how digital tools can raise new questions or provide new insights into traditional research.