Customer attrition in the banking industry occurs when consumers quit using the goods and services offered by the bank for some time and,after that,end their connection with the bank.Therefore,customer retention is es...Customer attrition in the banking industry occurs when consumers quit using the goods and services offered by the bank for some time and,after that,end their connection with the bank.Therefore,customer retention is essential in today’s extremely competitive banking market.Additionally,having a solid customer base helps attract new consumers by fostering confidence and a referral from a current clientele.These factors make reducing client attrition a crucial step that banks must pursue.In our research,we aim to examine bank data and forecast which users will most likely discontinue using the bank’s services and become paying customers.We use various machine learning algorithms to analyze the data and show comparative analysis on different evaluation metrics.In addition,we developed a Data Visualization RShiny app for data science and management regarding customer churn analysis.Analyzing this data will help the bank indicate the trend and then try to retain customers on the verge of attrition.展开更多
Many complex systems are frequently subject to the influence of uncertain disturbances,which can exert a profound effect on the critical transitions(CTs),potentially resulting in catastrophic consequences.Consequently...Many complex systems are frequently subject to the influence of uncertain disturbances,which can exert a profound effect on the critical transitions(CTs),potentially resulting in catastrophic consequences.Consequently,it is of uttermost importance to provide warnings for noise-induced CTs in various applications.Although capturing certain generic symptoms of transition behaviors from observational and simulated data poses a challenging problem,this work attempts to extract information regarding CTs from simulated data of a Gaussian white noise-induced tri-stable system.Using the extended dynamic mode decomposition(EDMD)algorithm,we initially obtain finite-dimensional approximations of both the stochastic Koopman operator and the generator.Subsequently,the drift parameters and the noise intensity within the system are identified from the simulated data.Utilizing the identified system,the parameter-dependent basin of the unsafe regime(PDBUR)is quantified,enabling data-driven early warning of Gaussian white noise-induced CTs.Finally,an error analysis is carried out to verify the effectiveness of the data-driven results.Our findings may serve as a paradigm for understanding and predicting noise-induced CTs in complex systems based on data.展开更多
Remote sensing plays a pivotal role in forest inventory by enabling efficient large-scale monitoring while minimizing fieldwork costs.However,missing values pose a critical challenge in remote sensing applications,as ...Remote sensing plays a pivotal role in forest inventory by enabling efficient large-scale monitoring while minimizing fieldwork costs.However,missing values pose a critical challenge in remote sensing applications,as ignoring or mishandling such data gaps can introduce systematic bias into the estimation of target variables for natural resource monitoring.This can lead to cascading errors that propagate through forest and ecosystem management decisions,ultimately hindering progress toward sustainable forest management,biodiversity conservation,and climate change mitigation strategies.This study aims to propose and demonstrate a procedure that employs hybrid estimators to address the limitations of missing remotely sensed data in forest inventory,using Landsat 7 ETM+SLC-off data as an archived source for forest resource monitoring as a case in point.We compared forest inventory estimates from the hybrid estimator with those from a conventional model-based(CMB)estimator using Sentinel-2 data without missing values.Monte Carlo simulations revealed three key findings:(1)The hybrid estimator,leveraging missing-data remote sensing represented by Landsat 7 ETM+SLCoff data,achieved a sampling precision of over 90%,meeting China's national standard for the National Forest Inventory(NFI);(2)The hybrid estimator demonstrated comparable efficiency to the CMB estimator;(3)The uncertainty associated with hybrid estimators was primarily dominated by model parameter estimation,which could be effectively mitigated by slightly increasing the training sample size or refining model specification.Overall,in forest inventory,the hybrid estimator can surmount the limitations posed by missing values in remotely sensed auxiliary data,effectively balancing cost-effectiveness and flexibility.展开更多
1.Introduction Artificial intelligence(AI)is rapidly reshaping geoscience,from Earth observation interpretation and hazard forecasting to subsurface characterisation and Earth system modelling(Kochupillai et al.,2022;...1.Introduction Artificial intelligence(AI)is rapidly reshaping geoscience,from Earth observation interpretation and hazard forecasting to subsurface characterisation and Earth system modelling(Kochupillai et al.,2022;Sun et al.,2024).These capabilities emerge at a time when geoscientific evidence is increasingly informing high-stakes decisions about climate adaptation,resource development,and disaster risk reduction(McGovern et al.,2022).展开更多
In trying to explain why Hong Kong of China ranks highest in life expectancy in the world,we review what various experts are hypothesizing,and how data science methods may be used to provide more evidence-based conclu...In trying to explain why Hong Kong of China ranks highest in life expectancy in the world,we review what various experts are hypothesizing,and how data science methods may be used to provide more evidence-based conclusions.While more data become available,we find some data analysis studies were too simplistic,while others too overwhelming in answering this challenging question.We find the approach that analyzes life expectancy related data(mortality causes and rate for different cohorts)inspiring,and use this approach to study a carefully selected set of targets for comparison.In discussing the factors that matter,we argue that it is more reasonable to try to identify a set of factors that together explain the phenomenon.展开更多
Health data and cutting-edge technologies empower medicine and improve healthcare.It has become even more true during the COVID-19 pandemic.Through coronavirus data sharing and worldwide collaboration,the speed of vac...Health data and cutting-edge technologies empower medicine and improve healthcare.It has become even more true during the COVID-19 pandemic.Through coronavirus data sharing and worldwide collaboration,the speed of vaccine development for COVID-19 is unprecedented.Digital and data technologies were quickly adopted during the pandemic,showing how those technologies can be harnessed to enhance public health and healthcare.A wide range of digital data sources are being utilized and visually presented to enhance the epidemiological surveillance of COVID-19.Digital contact tracing mobile apps have been adopted by many countries to control community transmission.Deep learning has been utilized to achieve various solutions for COVID-19 disruption,including outbreak prediction,virus spread tracking.展开更多
The data production elements are driving profound transformations in the real economy across production objects,methods,and tools,generating significant economic effects such as industrial structure upgrading.This pap...The data production elements are driving profound transformations in the real economy across production objects,methods,and tools,generating significant economic effects such as industrial structure upgrading.This paper aims to reveal the impact mechanism of the data elements on the“three transformations”(high-end,intelligent,and green)in the manufacturing sector,theoretically elucidating the intrinsic mechanisms by which the data elements influence these transformations.The study finds that the data elements significantly enhance the high-end,intelligent,and green levels of China's manufacturing industry.In terms of the pathways of impact,the data elements primarily influence the development of high-tech industries and overall green technological innovation,thereby affecting the high-end,intelligent,and green transformation of the industry.展开更多
Improving population health by creating more equitable health systems is a major focus of health policy and planning today.However,before we can achieve equity in health,we must first begin by leveraging all we have l...Improving population health by creating more equitable health systems is a major focus of health policy and planning today.However,before we can achieve equity in health,we must first begin by leveraging all we have learned,and are continuing to discover,about the many social,structural,and environmental determinants of health.We must fully consider the conditions in which people are born,grow,learn,work,play,and age.The study of social determinants of health has made tremendous strides in recent decades.At the same time,we have seen huge advances in how health data are collected,analyzed,and used to inform action in the health sector.It is time to merge these two fields,to harness the best from both and to improve decision-making to accelerate evidence-based action toward greater health equity.展开更多
Semantic communication(SemCom)aims to achieve high-fidelity information delivery under low communication consumption by only guaranteeing semantic accuracy.Nevertheless,semantic communication still suffers from unexpe...Semantic communication(SemCom)aims to achieve high-fidelity information delivery under low communication consumption by only guaranteeing semantic accuracy.Nevertheless,semantic communication still suffers from unexpected channel volatility and thus developing a re-transmission mechanism(e.g.,hybrid automatic repeat request[HARQ])becomes indispensable.In that regard,instead of discarding previously transmitted information,the incremental knowledge-based HARQ(IK-HARQ)is deemed as a more effective mechanism that could sufficiently utilize the information semantics.However,considering the possible existence of semantic ambiguity in image transmission,a simple bit-level cyclic redundancy check(CRC)might compromise the performance of IK-HARQ.Therefore,there emerges a strong incentive to revolutionize the CRC mechanism,thus more effectively reaping the benefits of both SemCom and HARQ.In this paper,built on top of swin transformer-based joint source-channel coding(JSCC)and IK-HARQ,we propose a semantic image transmission framework SC-TDA-HARQ.In particular,different from the conventional CRC,we introduce a topological data analysis(TDA)-based error detection method,which capably digs out the inner topological and geometric information of images,to capture semantic information and determine the necessity for re-transmission.Extensive numerical results validate the effectiveness and efficiency of the proposed SC-TDA-HARQ framework,especially under the limited bandwidth condition,and manifest the superiority of TDA-based error detection method in image transmission.展开更多
Cancer deaths and new cases worldwide are projected to rise by 47%by 2040,with transitioning countries experiencing an even higher increase of up to 95%.Tumor severity is profoundly influenced by the timing,accuracy,a...Cancer deaths and new cases worldwide are projected to rise by 47%by 2040,with transitioning countries experiencing an even higher increase of up to 95%.Tumor severity is profoundly influenced by the timing,accuracy,and stage of diagnosis,which directly impacts clinical decision-making.Various biological entities,including genes,proteins,mRNAs,miRNAs,and metabolites,contribute to cancer development.The emergence of multi-omics technologies has transformed cancer research by revealing molecular alterations across multiple biological layers.This integrative approach supports the notion that cancer is fundamentally driven by such alterations,enabling the discovery ofmolecular signatures for precision oncology.This reviewexplores the role of AI-drivenmulti-omics analyses in cancer medicine,emphasizing their potential to identify novel biomarkers and therapeutic targets,enhance understanding of Tumor biology,and address integration challenges in clinical workflows.Network biology analyzes identified ERBB2,KRAS,and TP53 as top hub genes in lung cancer based on Maximal Clique Centrality(MCC)scores.In contrast,TP53,ERBB2,ESR1,MYC,and BRCA1 emerged as central regulators in breast cancer,linked to cell proliferation,hormonal signaling,and genomic stability.The review also discusses how specific Artificial Intelligence(AI)algorithms can streamline the integration of heterogeneous datasets,facilitate the interpretation of the tumor microenvironment,and support data-driven clinical strategies.展开更多
BACKGROUND Hepatocellular carcinoma(HCC)remains a significant public health concern in South Korea even though the incidence rates are declining.While medical travel for cancer treatment is common,its patterns and inf...BACKGROUND Hepatocellular carcinoma(HCC)remains a significant public health concern in South Korea even though the incidence rates are declining.While medical travel for cancer treatment is common,its patterns and influencing factors for patients with HCC are unknown.AIM To assess medical travel patterns and determinants and their policy implications among patients with newly diagnosed HCC in South Korea.METHODS This retrospective cohort study used the National Health Insurance Service database to identify patients with newly diagnosed HCC from 2013 to 2021.Medical travel was defined as receiving initial treatment outside one’s residential region.Patient characteristics and regional trends were analyzed,and factors influencing medical travel were identified using logistic regression analysis.RESULTS Among 64808 patients 52.4%received treatment in the capital.This proportion increased to 67.4%when including the surrounding metropolitan area.Medical travel was significantly more common among younger and wealthier patients.Patients with greater comorbidity burden or liver cirrhosis were less likely to travel.While geographic distance influenced travel patterns,high-volume academic centers in the capital attracted patients nationwide regardless of proximity.CONCLUSION This nationwide study highlighted the centralization of HCC care in the capital.This observation indicates that regional cancer hubs should be strengthened and promoted for equitable healthcare access.展开更多
The widespread usage of rechargeable batteries in portable devices,electric vehicles,and energy storage systems has underscored the importance for accurately predicting their lifetimes.However,data scarcity often limi...The widespread usage of rechargeable batteries in portable devices,electric vehicles,and energy storage systems has underscored the importance for accurately predicting their lifetimes.However,data scarcity often limits the accuracy of prediction models,which is escalated by the incompletion of data induced by the issues such as sensor failures.To address these challenges,we propose a novel approach to accommodate data insufficiency through achieving external information from incomplete data samples,which are usually discarded in existing studies.In order to fully unleash the prediction power of incomplete data,we have investigated the Multiple Imputation by Chained Equations(MICE)method that diversifies the training data through exploring the potential data patterns.The experimental results demonstrate that the proposed method significantly outperforms the baselines in the most considered scenarios while reducing the prediction root mean square error(RMSE)by up to 18.9%.Furthermore,we have also observed that the penetration of incomplete data benefits the explainability of the prediction model through facilitating the feature selection.展开更多
Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensi...Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.展开更多
Electric Vehicle Charging Systems(EVCS)are increasingly vulnerable to cybersecurity threats as they integrate deeply into smart grids and Internet ofThings(IoT)environments,raising significant security challenges.Most...Electric Vehicle Charging Systems(EVCS)are increasingly vulnerable to cybersecurity threats as they integrate deeply into smart grids and Internet ofThings(IoT)environments,raising significant security challenges.Most existing research primarily emphasizes network-level anomaly detection,leaving critical vulnerabilities at the host level underexplored.This study introduces a novel forensic analysis framework leveraging host-level data,including system logs,kernel events,and Hardware Performance Counters(HPC),to detect and analyze sophisticated cyberattacks such as cryptojacking,Denial-of-Service(DoS),and reconnaissance activities targeting EVCS.Using comprehensive forensic analysis and machine learning models,the proposed framework significantly outperforms existing methods,achieving an accuracy of 98.81%.The findings offer insights into distinct behavioral signatures associated with specific cyber threats,enabling improved cybersecurity strategies and actionable recommendations for robust EVCS infrastructure protection.展开更多
Accurate capacity and State of Charge(SOC)estimation are crucial for ensuring the safety and longevity of lithium-ion batteries in electric vehicles.This study examines ten machine learning architectures,Including Dee...Accurate capacity and State of Charge(SOC)estimation are crucial for ensuring the safety and longevity of lithium-ion batteries in electric vehicles.This study examines ten machine learning architectures,Including Deep Belief Network(DBN),Bidirectional Recurrent Neural Network(BiDirRNN),Gated Recurrent Unit(GRU),and others using the NASA B0005 dataset of 591,458 instances.Results indicate that DBN excels in capacity estimation,achieving orders-of-magnitude lower error values and explaining over 99.97%of the predicted variable’s variance.When computational efficiency is paramount,the Deep Neural Network(DNN)offers a strong alternative,delivering near-competitive accuracy with significantly reduced prediction times.The GRU achieves the best overall performance for SOC estimation,attaining an R^(2) of 0.9999,while the BiDirRNN provides a marginally lower error at a slightly higher computational speed.In contrast,Convolutional Neural Networks(CNN)and Radial Basis Function Networks(RBFN)exhibit relatively high error rates,making them less viable for real-world battery management.Analyses of error distributions reveal that the top-performing models cluster most predictions within tight bounds,limiting the risk of overcharging or deep discharging.These findings highlight the trade-off between accuracy and computational overhead,offering valuable guidance for battery management system(BMS)designers seeking optimal performance under constrained resources.Future work may further explore advanced data augmentation and domain adaptation techniques to enhance these models’robustness in diverse operating conditions.展开更多
Metaheuristic optimization methods are iterative search processes that aim to efficiently solve complexoptimization problems. These basically find the solution space very efficiently, often without utilizing the gradi...Metaheuristic optimization methods are iterative search processes that aim to efficiently solve complexoptimization problems. These basically find the solution space very efficiently, often without utilizing the gradientinformation, and are inspired by the bio-inspired and socially motivated heuristics. Metaheuristic optimizationalgorithms are increasingly applied to complex feature selection problems in high-dimensional medical datasets.Among these, Teaching-Learning-Based optimization (TLBO) has proven effective for continuous design tasks bybalancing exploration and exploitation phases. However, its binary version (BTLBO) suffers from limited exploitationability, often converging prematurely or getting trapped in local optima, particularly when applied to discrete featureselection tasks. Previous studies reported that BTLBO yields lower classification accuracy and higher feature subsetvariance compared to other hybrid methods in benchmark tests, motivating the development of hybrid approaches.This study proposes a novel hybrid algorithm, BTLBO-Cheetah Optimizer (BTLBO-CO), which integrates the globalexploration strength of BTLBO with the local exploitation efficiency of the Cheetah Optimization (CO) algorithm. Theobjective is to enhance the feature selection process for cancer classification tasks involving high-dimensional data. Theproposed BTLBO-CO algorithm was evaluated on six benchmark cancer datasets: 11 tumors (T), Lung Cancer (LUC),Leukemia (LEU), Small Round Blue Cell Tumor or SRBCT (SR), Diffuse Large B-cell Lymphoma or DLBCL (DL), andProstate Tumor (PT).The results demonstrate superior classification accuracy across all six datasets, achieving 93.71%,96.12%, 98.13%, 97.11%, 98.44%, and 98.84%, respectively.These results validate the effectiveness of the hybrid approachin addressing diverse feature selection challenges using a Support Vector Machine (SVM) classifier.展开更多
This paper studies certain estimates for the lower bound of distance between unitary orbits of normal elements.We show that the distance between unitary orbits of normal elements of simple C^(*)-algebras of tracial ra...This paper studies certain estimates for the lower bound of distance between unitary orbits of normal elements.We show that the distance between unitary orbits of normal elements of simple C^(*)-algebras of tracial rank no more than k has a lower bound.Furthermore,if k≤1 and normal elements are commuting,then the lower bound will be better.Another result establishes a connection involving the spectrum distance operator Dc between a C^(*)-algebra of stable rank one C^(*)-algebra and its hereditary C^(*)-subalgebra.展开更多
In this paper,we establish and study a single-species logistic model with impulsive age-selective harvesting.First,we prove the ultimate boundedness of the solutions of the system.Then,we obtain conditions for the asy...In this paper,we establish and study a single-species logistic model with impulsive age-selective harvesting.First,we prove the ultimate boundedness of the solutions of the system.Then,we obtain conditions for the asymptotic stability of the trivial solution and the positive periodic solution.Finally,numerical simulations are presented to validate our results.Our results show that age-selective harvesting is more conducive to sustainable population survival than non-age-selective harvesting.展开更多
Predicting the behavior of renewable energy systems requires models capable of generating accurate forecasts from limited historical data,a challenge that becomes especially pronounced when commissioning new facil-iti...Predicting the behavior of renewable energy systems requires models capable of generating accurate forecasts from limited historical data,a challenge that becomes especially pronounced when commissioning new facil-ities where operational records are scarce.This review aims to synthesize recent progress in data-efficient deep learning approaches for addressing such“cold-start”forecasting problems.It primarily covers three interrelated domains—solar photovoltaic(PV),wind power,and electrical load forecasting—where data scarcity and operational variability are most critical,while also including representative studies on hydropower and carbon emission prediction to provide a broader systems perspective.To this end,we examined trends from over 150 predominantly peer-reviewed studies published between 2019 and mid-2025,highlighting advances in zero-shot and few-shot meta-learning frameworks that enable rapid model adaptation with minimal labeled data.Moreover,transfer learning approaches combined with spatiotemporal graph neural networks have been employed to transfer knowledge from existing energy assets to new,data-sparse environments,effectively capturing hidden dependencies among geographic features,meteorological dynamics,and grid structures.Synthetic data generation has further proven valuable for expanding training samples and mitigating overfitting in cold-start scenarios.In addition,large language models and explainable artificial intelligence(XAI)—notably conversational XAI systems—have been used to interpret and communicate complex model behaviors in accessible terms,fostering operator trust from the earliest deployment stages.By consolidating methodological advances,unresolved challenges,and open-source resources,this review provides a coherent overview of deep learning strategies that can shorten the data-sparse ramp-up period of new energy infrastructures and accelerate the transition toward resilient,low-carbon electricity grids.展开更多
文摘Customer attrition in the banking industry occurs when consumers quit using the goods and services offered by the bank for some time and,after that,end their connection with the bank.Therefore,customer retention is essential in today’s extremely competitive banking market.Additionally,having a solid customer base helps attract new consumers by fostering confidence and a referral from a current clientele.These factors make reducing client attrition a crucial step that banks must pursue.In our research,we aim to examine bank data and forecast which users will most likely discontinue using the bank’s services and become paying customers.We use various machine learning algorithms to analyze the data and show comparative analysis on different evaluation metrics.In addition,we developed a Data Visualization RShiny app for data science and management regarding customer churn analysis.Analyzing this data will help the bank indicate the trend and then try to retain customers on the verge of attrition.
基金Project supported by the National Natural Science Foundation of China(No.12402033)the National Natural Science Foundation for Distinguished Young Scholars of China(No.52225211)。
文摘Many complex systems are frequently subject to the influence of uncertain disturbances,which can exert a profound effect on the critical transitions(CTs),potentially resulting in catastrophic consequences.Consequently,it is of uttermost importance to provide warnings for noise-induced CTs in various applications.Although capturing certain generic symptoms of transition behaviors from observational and simulated data poses a challenging problem,this work attempts to extract information regarding CTs from simulated data of a Gaussian white noise-induced tri-stable system.Using the extended dynamic mode decomposition(EDMD)algorithm,we initially obtain finite-dimensional approximations of both the stochastic Koopman operator and the generator.Subsequently,the drift parameters and the noise intensity within the system are identified from the simulated data.Utilizing the identified system,the parameter-dependent basin of the unsafe regime(PDBUR)is quantified,enabling data-driven early warning of Gaussian white noise-induced CTs.Finally,an error analysis is carried out to verify the effectiveness of the data-driven results.Our findings may serve as a paradigm for understanding and predicting noise-induced CTs in complex systems based on data.
基金supported by the National Key R&D Program of China(No.2023YFF1304002-05)the National Social Science Fund of China(No.22BTJ005)the National Natural Science Foundation of China(No.32572049)。
文摘Remote sensing plays a pivotal role in forest inventory by enabling efficient large-scale monitoring while minimizing fieldwork costs.However,missing values pose a critical challenge in remote sensing applications,as ignoring or mishandling such data gaps can introduce systematic bias into the estimation of target variables for natural resource monitoring.This can lead to cascading errors that propagate through forest and ecosystem management decisions,ultimately hindering progress toward sustainable forest management,biodiversity conservation,and climate change mitigation strategies.This study aims to propose and demonstrate a procedure that employs hybrid estimators to address the limitations of missing remotely sensed data in forest inventory,using Landsat 7 ETM+SLC-off data as an archived source for forest resource monitoring as a case in point.We compared forest inventory estimates from the hybrid estimator with those from a conventional model-based(CMB)estimator using Sentinel-2 data without missing values.Monte Carlo simulations revealed three key findings:(1)The hybrid estimator,leveraging missing-data remote sensing represented by Landsat 7 ETM+SLCoff data,achieved a sampling precision of over 90%,meeting China's national standard for the National Forest Inventory(NFI);(2)The hybrid estimator demonstrated comparable efficiency to the CMB estimator;(3)The uncertainty associated with hybrid estimators was primarily dominated by model parameter estimation,which could be effectively mitigated by slightly increasing the training sample size or refining model specification.Overall,in forest inventory,the hybrid estimator can surmount the limitations posed by missing values in remotely sensed auxiliary data,effectively balancing cost-effectiveness and flexibility.
基金supported by the Natural Science Foundation of Jiangsu Province,China(Grant No.BK20240937)the Natural Science Foundation of Shandong Province(Grant No.ZR2021QE187)+2 种基金the Shandong Higher Education“Young Entrepreneurship Talents Introduction and Cultivation Program”Project(Grant No.ZXQT20221228001)the Natural Science Foundation of China(Grant No.42502273)the Science and Technology Innovation Program of Hunan Province(Grant No.2022RC4028).
文摘1.Introduction Artificial intelligence(AI)is rapidly reshaping geoscience,from Earth observation interpretation and hazard forecasting to subsurface characterisation and Earth system modelling(Kochupillai et al.,2022;Sun et al.,2024).These capabilities emerge at a time when geoscientific evidence is increasingly informing high-stakes decisions about climate adaptation,resource development,and disaster risk reduction(McGovern et al.,2022).
基金support of funding(No.UGC/IDS(R)11/21)from the Hong Kong SAR Government.
文摘In trying to explain why Hong Kong of China ranks highest in life expectancy in the world,we review what various experts are hypothesizing,and how data science methods may be used to provide more evidence-based conclusions.While more data become available,we find some data analysis studies were too simplistic,while others too overwhelming in answering this challenging question.We find the approach that analyzes life expectancy related data(mortality causes and rate for different cohorts)inspiring,and use this approach to study a carefully selected set of targets for comparison.In discussing the factors that matter,we argue that it is more reasonable to try to identify a set of factors that together explain the phenomenon.
文摘Health data and cutting-edge technologies empower medicine and improve healthcare.It has become even more true during the COVID-19 pandemic.Through coronavirus data sharing and worldwide collaboration,the speed of vaccine development for COVID-19 is unprecedented.Digital and data technologies were quickly adopted during the pandemic,showing how those technologies can be harnessed to enhance public health and healthcare.A wide range of digital data sources are being utilized and visually presented to enhance the epidemiological surveillance of COVID-19.Digital contact tracing mobile apps have been adopted by many countries to control community transmission.Deep learning has been utilized to achieve various solutions for COVID-19 disruption,including outbreak prediction,virus spread tracking.
文摘The data production elements are driving profound transformations in the real economy across production objects,methods,and tools,generating significant economic effects such as industrial structure upgrading.This paper aims to reveal the impact mechanism of the data elements on the“three transformations”(high-end,intelligent,and green)in the manufacturing sector,theoretically elucidating the intrinsic mechanisms by which the data elements influence these transformations.The study finds that the data elements significantly enhance the high-end,intelligent,and green levels of China's manufacturing industry.In terms of the pathways of impact,the data elements primarily influence the development of high-tech industries and overall green technological innovation,thereby affecting the high-end,intelligent,and green transformation of the industry.
文摘Improving population health by creating more equitable health systems is a major focus of health policy and planning today.However,before we can achieve equity in health,we must first begin by leveraging all we have learned,and are continuing to discover,about the many social,structural,and environmental determinants of health.We must fully consider the conditions in which people are born,grow,learn,work,play,and age.The study of social determinants of health has made tremendous strides in recent decades.At the same time,we have seen huge advances in how health data are collected,analyzed,and used to inform action in the health sector.It is time to merge these two fields,to harness the best from both and to improve decision-making to accelerate evidence-based action toward greater health equity.
基金supported in part by the National Key Research and Development Program of China under Grant 2024YFE0200600in part by the National Natural Science Foundation of China under Grant 62071425+3 种基金in part by the Zhejiang Key Research and Development Plan under Grant 2022C01093in part by the Zhejiang Provincial Natural Science Foundation of China under Grant LR23F010005in part by the National Key Laboratory of Wireless Communications Foundation under Grant 2023KP01601in part by the Big Data and Intelligent Computing Key Lab of CQUPT under Grant BDIC-2023-B-001.
文摘Semantic communication(SemCom)aims to achieve high-fidelity information delivery under low communication consumption by only guaranteeing semantic accuracy.Nevertheless,semantic communication still suffers from unexpected channel volatility and thus developing a re-transmission mechanism(e.g.,hybrid automatic repeat request[HARQ])becomes indispensable.In that regard,instead of discarding previously transmitted information,the incremental knowledge-based HARQ(IK-HARQ)is deemed as a more effective mechanism that could sufficiently utilize the information semantics.However,considering the possible existence of semantic ambiguity in image transmission,a simple bit-level cyclic redundancy check(CRC)might compromise the performance of IK-HARQ.Therefore,there emerges a strong incentive to revolutionize the CRC mechanism,thus more effectively reaping the benefits of both SemCom and HARQ.In this paper,built on top of swin transformer-based joint source-channel coding(JSCC)and IK-HARQ,we propose a semantic image transmission framework SC-TDA-HARQ.In particular,different from the conventional CRC,we introduce a topological data analysis(TDA)-based error detection method,which capably digs out the inner topological and geometric information of images,to capture semantic information and determine the necessity for re-transmission.Extensive numerical results validate the effectiveness and efficiency of the proposed SC-TDA-HARQ framework,especially under the limited bandwidth condition,and manifest the superiority of TDA-based error detection method in image transmission.
基金funded by KAU Endowment(WAQF)at King Abdulaziz University,Jeddah,Saudi Arabia.
文摘Cancer deaths and new cases worldwide are projected to rise by 47%by 2040,with transitioning countries experiencing an even higher increase of up to 95%.Tumor severity is profoundly influenced by the timing,accuracy,and stage of diagnosis,which directly impacts clinical decision-making.Various biological entities,including genes,proteins,mRNAs,miRNAs,and metabolites,contribute to cancer development.The emergence of multi-omics technologies has transformed cancer research by revealing molecular alterations across multiple biological layers.This integrative approach supports the notion that cancer is fundamentally driven by such alterations,enabling the discovery ofmolecular signatures for precision oncology.This reviewexplores the role of AI-drivenmulti-omics analyses in cancer medicine,emphasizing their potential to identify novel biomarkers and therapeutic targets,enhance understanding of Tumor biology,and address integration challenges in clinical workflows.Network biology analyzes identified ERBB2,KRAS,and TP53 as top hub genes in lung cancer based on Maximal Clique Centrality(MCC)scores.In contrast,TP53,ERBB2,ESR1,MYC,and BRCA1 emerged as central regulators in breast cancer,linked to cell proliferation,hormonal signaling,and genomic stability.The review also discusses how specific Artificial Intelligence(AI)algorithms can streamline the integration of heterogeneous datasets,facilitate the interpretation of the tumor microenvironment,and support data-driven clinical strategies.
基金Supported by Dong-A University Research Fund,No.20230598.
文摘BACKGROUND Hepatocellular carcinoma(HCC)remains a significant public health concern in South Korea even though the incidence rates are declining.While medical travel for cancer treatment is common,its patterns and influencing factors for patients with HCC are unknown.AIM To assess medical travel patterns and determinants and their policy implications among patients with newly diagnosed HCC in South Korea.METHODS This retrospective cohort study used the National Health Insurance Service database to identify patients with newly diagnosed HCC from 2013 to 2021.Medical travel was defined as receiving initial treatment outside one’s residential region.Patient characteristics and regional trends were analyzed,and factors influencing medical travel were identified using logistic regression analysis.RESULTS Among 64808 patients 52.4%received treatment in the capital.This proportion increased to 67.4%when including the surrounding metropolitan area.Medical travel was significantly more common among younger and wealthier patients.Patients with greater comorbidity burden or liver cirrhosis were less likely to travel.While geographic distance influenced travel patterns,high-volume academic centers in the capital attracted patients nationwide regardless of proximity.CONCLUSION This nationwide study highlighted the centralization of HCC care in the capital.This observation indicates that regional cancer hubs should be strengthened and promoted for equitable healthcare access.
文摘The widespread usage of rechargeable batteries in portable devices,electric vehicles,and energy storage systems has underscored the importance for accurately predicting their lifetimes.However,data scarcity often limits the accuracy of prediction models,which is escalated by the incompletion of data induced by the issues such as sensor failures.To address these challenges,we propose a novel approach to accommodate data insufficiency through achieving external information from incomplete data samples,which are usually discarded in existing studies.In order to fully unleash the prediction power of incomplete data,we have investigated the Multiple Imputation by Chained Equations(MICE)method that diversifies the training data through exploring the potential data patterns.The experimental results demonstrate that the proposed method significantly outperforms the baselines in the most considered scenarios while reducing the prediction root mean square error(RMSE)by up to 18.9%.Furthermore,we have also observed that the penetration of incomplete data benefits the explainability of the prediction model through facilitating the feature selection.
文摘Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.
文摘Electric Vehicle Charging Systems(EVCS)are increasingly vulnerable to cybersecurity threats as they integrate deeply into smart grids and Internet ofThings(IoT)environments,raising significant security challenges.Most existing research primarily emphasizes network-level anomaly detection,leaving critical vulnerabilities at the host level underexplored.This study introduces a novel forensic analysis framework leveraging host-level data,including system logs,kernel events,and Hardware Performance Counters(HPC),to detect and analyze sophisticated cyberattacks such as cryptojacking,Denial-of-Service(DoS),and reconnaissance activities targeting EVCS.Using comprehensive forensic analysis and machine learning models,the proposed framework significantly outperforms existing methods,achieving an accuracy of 98.81%.The findings offer insights into distinct behavioral signatures associated with specific cyber threats,enabling improved cybersecurity strategies and actionable recommendations for robust EVCS infrastructure protection.
文摘Accurate capacity and State of Charge(SOC)estimation are crucial for ensuring the safety and longevity of lithium-ion batteries in electric vehicles.This study examines ten machine learning architectures,Including Deep Belief Network(DBN),Bidirectional Recurrent Neural Network(BiDirRNN),Gated Recurrent Unit(GRU),and others using the NASA B0005 dataset of 591,458 instances.Results indicate that DBN excels in capacity estimation,achieving orders-of-magnitude lower error values and explaining over 99.97%of the predicted variable’s variance.When computational efficiency is paramount,the Deep Neural Network(DNN)offers a strong alternative,delivering near-competitive accuracy with significantly reduced prediction times.The GRU achieves the best overall performance for SOC estimation,attaining an R^(2) of 0.9999,while the BiDirRNN provides a marginally lower error at a slightly higher computational speed.In contrast,Convolutional Neural Networks(CNN)and Radial Basis Function Networks(RBFN)exhibit relatively high error rates,making them less viable for real-world battery management.Analyses of error distributions reveal that the top-performing models cluster most predictions within tight bounds,limiting the risk of overcharging or deep discharging.These findings highlight the trade-off between accuracy and computational overhead,offering valuable guidance for battery management system(BMS)designers seeking optimal performance under constrained resources.Future work may further explore advanced data augmentation and domain adaptation techniques to enhance these models’robustness in diverse operating conditions.
基金funded by the Deanship of Research andGraduate Studies at King Khalid University through the Large Research Project under grant number RGP2/417/46.
文摘Metaheuristic optimization methods are iterative search processes that aim to efficiently solve complexoptimization problems. These basically find the solution space very efficiently, often without utilizing the gradientinformation, and are inspired by the bio-inspired and socially motivated heuristics. Metaheuristic optimizationalgorithms are increasingly applied to complex feature selection problems in high-dimensional medical datasets.Among these, Teaching-Learning-Based optimization (TLBO) has proven effective for continuous design tasks bybalancing exploration and exploitation phases. However, its binary version (BTLBO) suffers from limited exploitationability, often converging prematurely or getting trapped in local optima, particularly when applied to discrete featureselection tasks. Previous studies reported that BTLBO yields lower classification accuracy and higher feature subsetvariance compared to other hybrid methods in benchmark tests, motivating the development of hybrid approaches.This study proposes a novel hybrid algorithm, BTLBO-Cheetah Optimizer (BTLBO-CO), which integrates the globalexploration strength of BTLBO with the local exploitation efficiency of the Cheetah Optimization (CO) algorithm. Theobjective is to enhance the feature selection process for cancer classification tasks involving high-dimensional data. Theproposed BTLBO-CO algorithm was evaluated on six benchmark cancer datasets: 11 tumors (T), Lung Cancer (LUC),Leukemia (LEU), Small Round Blue Cell Tumor or SRBCT (SR), Diffuse Large B-cell Lymphoma or DLBCL (DL), andProstate Tumor (PT).The results demonstrate superior classification accuracy across all six datasets, achieving 93.71%,96.12%, 98.13%, 97.11%, 98.44%, and 98.84%, respectively.These results validate the effectiveness of the hybrid approachin addressing diverse feature selection challenges using a Support Vector Machine (SVM) classifier.
基金Supported by Zhejiang Provincial Natural Science Foundation of China(No.ZCLQN25A0103)。
文摘This paper studies certain estimates for the lower bound of distance between unitary orbits of normal elements.We show that the distance between unitary orbits of normal elements of simple C^(*)-algebras of tracial rank no more than k has a lower bound.Furthermore,if k≤1 and normal elements are commuting,then the lower bound will be better.Another result establishes a connection involving the spectrum distance operator Dc between a C^(*)-algebra of stable rank one C^(*)-algebra and its hereditary C^(*)-subalgebra.
基金Supported by the National Natural Science Foundation of China(12261018)Universities Key Laboratory of Mathematical Modeling and Data Mining in Guizhou Province(2023013)。
文摘In this paper,we establish and study a single-species logistic model with impulsive age-selective harvesting.First,we prove the ultimate boundedness of the solutions of the system.Then,we obtain conditions for the asymptotic stability of the trivial solution and the positive periodic solution.Finally,numerical simulations are presented to validate our results.Our results show that age-selective harvesting is more conducive to sustainable population survival than non-age-selective harvesting.
文摘Predicting the behavior of renewable energy systems requires models capable of generating accurate forecasts from limited historical data,a challenge that becomes especially pronounced when commissioning new facil-ities where operational records are scarce.This review aims to synthesize recent progress in data-efficient deep learning approaches for addressing such“cold-start”forecasting problems.It primarily covers three interrelated domains—solar photovoltaic(PV),wind power,and electrical load forecasting—where data scarcity and operational variability are most critical,while also including representative studies on hydropower and carbon emission prediction to provide a broader systems perspective.To this end,we examined trends from over 150 predominantly peer-reviewed studies published between 2019 and mid-2025,highlighting advances in zero-shot and few-shot meta-learning frameworks that enable rapid model adaptation with minimal labeled data.Moreover,transfer learning approaches combined with spatiotemporal graph neural networks have been employed to transfer knowledge from existing energy assets to new,data-sparse environments,effectively capturing hidden dependencies among geographic features,meteorological dynamics,and grid structures.Synthetic data generation has further proven valuable for expanding training samples and mitigating overfitting in cold-start scenarios.In addition,large language models and explainable artificial intelligence(XAI)—notably conversational XAI systems—have been used to interpret and communicate complex model behaviors in accessible terms,fostering operator trust from the earliest deployment stages.By consolidating methodological advances,unresolved challenges,and open-source resources,this review provides a coherent overview of deep learning strategies that can shorten the data-sparse ramp-up period of new energy infrastructures and accelerate the transition toward resilient,low-carbon electricity grids.