Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from sei...Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from seismic networks,satellite observations,and geospatial repositories,creating the need for scalable infrastructures capable of integrating and analyzing such data to support intelligent decision-making.Data warehousing technologies provide a robust foundation for this purpose;however,existing earthquake-oriented data warehouses remain limited,often relying on simplified schemas,domain-specific analytics,or cataloguing efforts.This paper presents the design and implementation of a spatio-temporal data warehouse for seismic activity.The framework integrates spatial and temporal dimensions in a unified schema and introduces a novel array-based approach for managing many-to-many relationships between facts and dimensions without intermediate bridge tables.A comparative evaluation against a conventional bridge-table schema demonstrates that the array-based design improves fact-centric query performance,while the bridge-table schema remains advantageous for dimension-centric queries.To reconcile these trade-offs,a hybrid schema is proposed that retains both representations,ensuring balanced efficiency across heterogeneous workloads.The proposed framework demonstrates how spatio-temporal data warehousing can address schema complexity,improve query performance,and support multidimensional visualization.In doing so,it provides a foundation for integrating seismic analysis into broader big data-driven intelligent decision systems for disaster resilience,risk mitigation,and emergency management.展开更多
在小学高年级英语语篇教学中,存在学生思维浅表化、问题设计碎片化、旧版教材适配难这三个痛点。以译林版英语教材六年级上册Unit 4 Then and now中Story time的教学为例,教师立足教材文本,构建“课前定问—课初引链—课中解链—课后拓...在小学高年级英语语篇教学中,存在学生思维浅表化、问题设计碎片化、旧版教材适配难这三个痛点。以译林版英语教材六年级上册Unit 4 Then and now中Story time的教学为例,教师立足教材文本,构建“课前定问—课初引链—课中解链—课后拓链—全程评链”的五步闭环,用大问题拉主线、小问题搭台阶,能激活学生语篇学习内驱力,实现英语教学从“知识传递”到“素养培养”的转变。展开更多
With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service...With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service resources,uneven distribution,and prominent supply-demand contradictions have seriously affected service quality.Big data technology,with core advantages including data collection,analysis and mining,and accurate prediction,provides a new solution for the allocation of community elderly care service resources.This paper systematically studies the application value of big data technology in the allocation of community elderly care service resources from three aspects:resource allocation efficiency,service accuracy,and management intelligence.Combined with practical needs,it proposes optimal allocation strategies such as building a big data analysis platform and accurately grasping the elderly’s care needs,striving to provide operable path references for the construction of community elderly care service systems,promoting the early realization of the elderly care service goal of“adequate support and proper care for the elderly”,and boosting the high-quality development of China’s elderly care service industry.展开更多
The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physici...The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physician-centered perspective,these technologies can potentially be used to improve diagnostic precision,support individualized treatment planning,enhance longitudinal patient management,and accelerate both clinical and translational research.In this review,we synthesize the core AI methodologies most relevant to oncology-machine learning,deep learning,and large language models-and examine how they interact with established and emerging oncology data platforms.We further highlight practical use cases in clinical workflows and research pipelines,emphasizing opportunities for advancing precision cancer care while also addressing challenges associated with data heterogeneity,model generalizability,privacy protection,and real-world implementation.By underscoring the synergistic value of AI and big data,this review aims to inform the development of clinically meaningful,context-adapted strategies that promote translational innovation in both global and locally resourced healthcare environments.展开更多
With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues su...With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues such as computer memory capacity and distributed storage of data across different locations,and are unable to directly apply to large-scale data sets.Therefore,in the context of big data,designing efficient and theoretically guaranteed statistical learning and inference algorithms has become a key issue that the current field of statistics urgently needs to address.In this paper,the application status of statistical analysis methods in the big data environment was systematically reviewed,and its future development directions were analyzed to provide reference and support for the further development of theory and methods of the statistical analysis of big data.展开更多
大孔径CT基于X线断层扫描原理的超大孔径设计,能容纳肥胖患者以及携带医疗设备的特殊患者,其凭借高清晰成像为放疗计划的制定与实施提供了关键的解剖学依据,是现代放疗“精准化”转型的重要支撑[1]。我院于2017年引进飞利浦Brilliance B...大孔径CT基于X线断层扫描原理的超大孔径设计,能容纳肥胖患者以及携带医疗设备的特殊患者,其凭借高清晰成像为放疗计划的制定与实施提供了关键的解剖学依据,是现代放疗“精准化”转型的重要支撑[1]。我院于2017年引进飞利浦Brilliance Big Bore CT设备,截至2024年底,已累计完成5000余例肿瘤患者的放疗定位扫描。展开更多
The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threa...The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threats is both necessary and complex,yet these interconnected healthcare settings generate enormous amounts of heterogeneous data.Traditional Intrusion Detection Systems(IDS),which are generally centralized and machine learning-based,often fail to address the rapidly changing nature of cyberattacks and are challenged by ethical concerns related to patient data privacy.Moreover,traditional AI-driven IDS usually face challenges in handling large-scale,heterogeneous healthcare data while ensuring data privacy and operational efficiency.To address these issues,emerging technologies such as Big Data Analytics(BDA)and Federated Learning(FL)provide a hybrid framework for scalable,adaptive intrusion detection in IoT-driven healthcare systems.Big data techniques enable processing large-scale,highdimensional healthcare data,and FL can be used to train a model in a decentralized manner without transferring raw data,thereby maintaining privacy between institutions.This research proposes a privacy-preserving Federated Learning–based model that efficiently detects cyber threats in connected healthcare systems while ensuring distributed big data processing,privacy,and compliance with ethical regulations.To strengthen the reliability of the reported findings,the resultswere validated using cross-dataset testing and 95%confidence intervals derived frombootstrap analysis,confirming consistent performance across heterogeneous healthcare data distributions.This solution takes a significant step toward securing next-generation healthcare infrastructure by combining scalability,privacy,adaptability,and earlydetection capabilities.The proposed global model achieves a test accuracy of 99.93%±0.03(95%CI)and amiss-rate of only 0.07%±0.02,representing state-of-the-art performance in privacy-preserving intrusion detection.The proposed FL-driven IDS framework offers an efficient,privacy-preserving,and scalable solution for securing next-generation healthcare infrastructures by combining adaptability,early detection,and ethical data management.展开更多
The Wufeng–Longmaxi Formation derives its name from the Upper Ordovician Wufeng Formation and the Lower Silurian Longmaxi Formation,found in sequence in the Sichuan Basin.This formation hosts rich shale gas reservoir...The Wufeng–Longmaxi Formation derives its name from the Upper Ordovician Wufeng Formation and the Lower Silurian Longmaxi Formation,found in sequence in the Sichuan Basin.This formation hosts rich shale gas reservoirs,and its shale gas enrichment patterns are examined in this study using data from 1197 shale samples collected from 14 wells.Five basic and three key parameters,eight in all,are assessed for each sample.The five basic parameters include burial depth and the contents of four mineral types—quartz,clay,carbonate,and other minerals;the three key parameters,representing shale gas enrichment,are total organic carbon(TOC)content,porosity,and gas content.The SHapley Additive exPlanations(SHAP)analysis originated in game theory is used here in an interpretable machine learning framework,to address issues of heterogeneous data structure,noisy relationships,and multi-objective optimization.An evaluation of the ranking,contribution values,and conditions of changes for these parameters offers new quantitative insights into shale gas enrichment patterns.A quantitative analysis of the relationship between data-sets identifies the primary factors controlling TOC,porosity,and gas content of shale gas reservoirs.The results show that TOC and porosity jointly influence gas content;mineral content has a significant impact on both,TOC and porosity;and the burial depth governs porosity which,in turn,affects the conditions under which shale gas is preserved.Input parameter thresholds are also determined and provide a basis for the establishment of quantitative criteria to evaluate shale gas enrichment.The predictive accuracy of the model used in this study is significantly improved by the step-wise addition of two input parameters,namely TOC and porosity,separately and together.Thus,the game theory method in big data-driven analysis uses a combination of TOC and porosity to evaluate the gas content with encouraging results—suggesting that these are the key parameters that indicate source rock and reservoir properties.展开更多
The development of remote sensing has seen the creation of a global measurement infrastructure of sustainable development due to growing multipolar archives,rising revisit frequency,and the availability of cloud-acces...The development of remote sensing has seen the creation of a global measurement infrastructure of sustainable development due to growing multipolar archives,rising revisit frequency,and the availability of cloud-accessible platforms of Earth observation.This review summarizes how remote sensing big data is being organized into decision-grade sustainability intelligence,the new approaches to analytics,and how Sustainable Development Goals(SDGs)-oriented application pathways inter-relate action pathways that bridge observations with action.The terminologies like new data ecosystem,data readiness and interoperability,changing economics of scalable computation,and detailing the functions of diversity of modalities(optical,Synthetic Aperture Radar—SAR,thermal,Light Detection and Ranging—LiDAR,hyperspectral)have been defined.These themes of analytics,which are transforming the practice of operational analytics,are then condensed:foundations and self-supervised learning of transferable representations,multi-modal fusion to gap fill and richer inference,spatiotemporal intelligence to trend of early warning,physics-aware hybrid methods to enhance robustness and meaning under non-stationary conditions.Across the climate risk,food systems,water resources,sustainable cities,ecosystems and biodiversity,energy transitions,and health exposure pathways,the roles of Earth Observation(EO)products as direct measures and proxies,and concepts of validating,semantic comparability,and communicating uncertainties play a key role in EO products becoming credible when faced with high-stakes deployment decisions.Lastly,we chart world ways of implementation via monitoring services,early warning systems,and systems of multiple regimes,and previously underline cross-cutting priorities,scalable structures in validation,performance,so that domains of shift,agreeable governance,and Dual-use risk safeguards,and sustainable lifecycle support of EO services.These priorities form a realistic set of priorities on the alignment of remote sensing innovation with quantifiable SDGs progress.展开更多
On October 18,2017,the 19th National Congress Report called for the implementation of the Healthy China Strategy.The development of biomedical data plays a pivotal role in advancing this strategy.Since the 18th Nation...On October 18,2017,the 19th National Congress Report called for the implementation of the Healthy China Strategy.The development of biomedical data plays a pivotal role in advancing this strategy.Since the 18th National Congress of the Communist Party of China,China has vigorously promoted the integration and implementation of the Healthy China and Digital China strategies.The National Health Commission has prioritized the development of health and medical big data,issuing policies to promote standardized applica-tions and foster innovation in"Internet+Healthcare."Biomedical data has significantly contributed to preci-sion medicine,personalized health management,drug development,disease diagnosis,public health monitor-ing,and epidemic prediction capabilities.展开更多
The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This...The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This era integrates biotechnology,artificial intelligence(AI),and big data information technology.In contrast,China is still in a transition period between stages 2.0 and 3.0,which primarily relies on conventional selection and molecular breeding.In the context of increasingly complex international situations,accurately identifying core issues in China's seed industry innovation and seizing the frontier of international seed technology are strategically important.These efforts are essential for ensuring food security and revitalizing the seed industry.This paper systematically analyzes the characteristics of crop breeding data from artificial selection to intelligent design breeding.It explores the applications and development trends of AI and big data in modern crop breeding from several key perspectives.These include highthroughput phenotype acquisition and analysis,multiomics big data database and management system construction,AI-based multiomics integrated analysis,and the development of intelligent breeding software tools based on biological big data and AI technology.Based on an in-depth analysis of the current status and challenges of China's seed industry technology development,we propose strategic goals and key tasks for China's new generation of AI and big data-driven intelligent design breeding.These suggestions aim to accelerate the development of an intelligent-driven crop breeding engineering system that features large-scale gene mining,efficient gene manipulation,engineered variety design,and systematized biobreeding.This study provides a theoretical basis and practical guidance for the development of China's seed industry technology.展开更多
This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models ...This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models offer insights, they fall short in presenting a holistic view of complex urban challenges. System dynamics (SD) models that are often utilized to provide holistic, systematic understanding of a research subject, like the urban system, emerge as valuable tools, but data scarcity and theoretical inadequacy pose challenges. The research reviews relevant papers on recent SD model applications in urban sustainability since 2018, categorizing them based on nine key indicators. Among the reviewed papers, data limitations and model assumptions were identified as ma jor challenges in applying SD models to urban sustainability. This led to exploring the transformative potential of big data analytics, a rare approach in this field as identified by this study, to enhance SD models’ empirical foundation. Integrating big data could provide data-driven calibration, potentially improving predictive accuracy and reducing reliance on simplified assumptions. The paper concludes by advocating for new approaches that reduce assumptions and promote real-time applicable models, contributing to a comprehensive understanding of urban sustainability through the synergy of big data and SD models.展开更多
Deep-time Earth research plays a pivotal role in deciphering the rates,patterns,and mechanisms of Earth's evolutionary processes throughout geological history,providing essential scientific foundations for climate...Deep-time Earth research plays a pivotal role in deciphering the rates,patterns,and mechanisms of Earth's evolutionary processes throughout geological history,providing essential scientific foundations for climate prediction,natural resource exploration,and sustainable planetary stewardship.To advance Deep-time Earth research in the era of big data and artificial intelligence,the International Union of Geological Sciences initiated the“Deeptime Digital Earth International Big Science Program”(DDE)in 2019.At the core of this ambitious program lies the development of geoscience knowledge graphs,serving as a transformative knowledge infrastructure that enables the integration,sharing,mining,and analysis of heterogeneous geoscience big data.The DDE knowledge graph initiative has made significant strides in three critical dimensions:(1)establishing a unified knowledge structure across geoscience disciplines that ensures consistent representation of geological entities and their interrelationships through standardized ontologies and semantic frameworks;(2)developing a robust and scalable software infrastructure capable of supporting both expert-driven and machine-assisted knowledge engineering for large-scale graph construction and management;(3)implementing a comprehensive three-tiered architecture encompassing basic,discipline-specific,and application-oriented knowledge graphs,spanning approximately 20 geoscience disciplines.Through its open knowledge framework and international collaborative network,this initiative has fostered multinational research collaborations,establishing a robust foundation for next-generation geoscience research while propelling the discipline toward FAIR(Findable,Accessible,Interoperable,Reusable)data practices in deep-time Earth systems research.展开更多
Research into metamorphism plays a pivotal role in reconstructing the evolution of continent,particularly through the study of ancient rocks that are highly susceptible to metamorphic alterations due to multiple tecto...Research into metamorphism plays a pivotal role in reconstructing the evolution of continent,particularly through the study of ancient rocks that are highly susceptible to metamorphic alterations due to multiple tectonic activities.In the big data era,the establishment of new data platforms and the application of big data methods have become a focus for metamorphic rocks.Significant progress has been made in creating specialized databases,compiling comprehensive datasets,and utilizing data analytics to address complex scientific questions.However,many existing databases are inadequate in meeting the specific requirements of metamorphic research,resulting from a substantial amount of valuable data remaining uncollected.Therefore,constructing new databases that can cope with the development of the data era is necessary.This article provides an extensive review of existing databases related to metamorphic rocks and discusses data-driven studies in this.Accordingly,several crucial factors that need to be taken into consideration in the establishment of specialized metamorphic databases are identified,aiming to leverage data-driven applications to achieve broader scientific objectives in metamorphic research.展开更多
Viral infectious diseases,characterized by their intricate nature and wide-ranging diversity,pose substantial challenges in the domain of data management.The vast volume of data generated by these diseases,spanning fr...Viral infectious diseases,characterized by their intricate nature and wide-ranging diversity,pose substantial challenges in the domain of data management.The vast volume of data generated by these diseases,spanning from the molecular mechanisms within cells to large-scale epidemiological patterns,has surpassed the capabilities of traditional analytical methods.In the era of artificial intelligence(AI)and big data,there is an urgent necessity for the optimization of these analytical methods to more effectively handle and utilize the information.Despite the rapid accumulation of data associated with viral infections,the lack of a comprehensive framework for integrating,selecting,and analyzing these datasets has left numerous researchers uncertain about which data to select,how to access it,and how to utilize it most effectively in their research.This review endeavors to fill these gaps by exploring the multifaceted nature of viral infectious diseases and summarizing relevant data across multiple levels,from the molecular details of pathogens to broad epidemiological trends.The scope extends from the micro-scale to the macro-scale,encompassing pathogens,hosts,and vectors.In addition to data summarization,this review thoroughly investigates various dataset sources.It also traces the historical evolution of data collection in the field of viral infectious diseases,highlighting the progress achieved over time.Simultaneously,it evaluates the current limitations that impede data utilization.Furthermore,we propose strategies to surmount these challenges,focusing on the development and application of advanced computational techniques,AI-driven models,and enhanced data integration practices.By providing a comprehensive synthesis of existing knowledge,this review is designed to guide future research and contribute to more informed approaches in the surveillance,prevention,and control of viral infectious diseases,particularly within the context of the expanding big-data landscape.展开更多
1.Introduction Nasopharyngeal carcinoma(NPC)has an imbalanced geographical and ethnic distribution,with notably high incidence rates in Southeastern Asia and China.China accounted for 42.4%of the newly diagnosed cases...1.Introduction Nasopharyngeal carcinoma(NPC)has an imbalanced geographical and ethnic distribution,with notably high incidence rates in Southeastern Asia and China.China accounted for 42.4%of the newly diagnosed cases worldwide in 2022.1NPC is sensitive to irradiation,and radiotherapy is the mainstay curative treatment modality.2The widespread use of intensity-modulated radiation therapy(IMRT)and image-guided radiotherapy(IGRT)has achieved great advances in survival outcomes and toxicity profiles among NPC patients.1In radiotherapy of NPC,the tumor’s proximity to critical structures demands accuracy in tumor delineation in order to avoid radiation-induced toxicities.However,tumor target delineation for radiotherapy of NPC is labor-intensive and radiation oncologists’proficiency varied considerably.3In recent years,the advent of big data analytics and artificial intelligence(AI)has opened up new avenues for improving the precision and efficacy of radiotherapy and individualized treatment in NPC management.3-6In this article,we explored how big data,AI-assisted delineation,radiotherapy planning,and adaptive radiotherapy(ART)are transforming clinical decision-making in NPC treatment.We also provided an outlook on the historical development of AI and big data,their current dominance in oncological radiotherapy,and their projected impact on future clinical practice(Figure 1).展开更多
This study examines the Big Data Collection and Preprocessing course at Anhui Institute of Information Engineering,implementing a hybrid teaching reform using the Bosi Smart Learning Platform.The proposed hybrid model...This study examines the Big Data Collection and Preprocessing course at Anhui Institute of Information Engineering,implementing a hybrid teaching reform using the Bosi Smart Learning Platform.The proposed hybrid model follows a“three-stage”and“two-subject”framework,incorporating a structured design for teaching content and assessment methods before,during,and after class.Practical results indicate that this approach significantly enhances teaching effectiveness and improves students’learning autonomy.展开更多
Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of th...Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of these data has not been well stored,managed and mined.With the development of cloud computing technology,it provides a rare development opportunity for logging big data private cloud.The traditional petrophysical evaluation and interpretation model has encountered great challenges in the face of new evaluation objects.The solution research of logging big data distributed storage,processing and learning functions integrated in logging big data private cloud has not been carried out yet.To establish a distributed logging big-data private cloud platform centered on a unifi ed learning model,which achieves the distributed storage and processing of logging big data and facilitates the learning of novel knowledge patterns via the unifi ed logging learning model integrating physical simulation and data models in a large-scale functional space,thus resolving the geo-engineering evaluation problem of geothermal fi elds.Based on the research idea of“logging big data cloud platform-unifi ed logging learning model-large function space-knowledge learning&discovery-application”,the theoretical foundation of unified learning model,cloud platform architecture,data storage and learning algorithm,arithmetic power allocation and platform monitoring,platform stability,data security,etc.have been carried on analysis.The designed logging big data cloud platform realizes parallel distributed storage and processing of data and learning algorithms.The feasibility of constructing a well logging big data cloud platform based on a unifi ed learning model of physics and data is analyzed in terms of the structure,ecology,management and security of the cloud platform.The case study shows that the logging big data cloud platform has obvious technical advantages over traditional logging evaluation methods in terms of knowledge discovery method,data software and results sharing,accuracy,speed and complexity.展开更多
文摘Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from seismic networks,satellite observations,and geospatial repositories,creating the need for scalable infrastructures capable of integrating and analyzing such data to support intelligent decision-making.Data warehousing technologies provide a robust foundation for this purpose;however,existing earthquake-oriented data warehouses remain limited,often relying on simplified schemas,domain-specific analytics,or cataloguing efforts.This paper presents the design and implementation of a spatio-temporal data warehouse for seismic activity.The framework integrates spatial and temporal dimensions in a unified schema and introduces a novel array-based approach for managing many-to-many relationships between facts and dimensions without intermediate bridge tables.A comparative evaluation against a conventional bridge-table schema demonstrates that the array-based design improves fact-centric query performance,while the bridge-table schema remains advantageous for dimension-centric queries.To reconcile these trade-offs,a hybrid schema is proposed that retains both representations,ensuring balanced efficiency across heterogeneous workloads.The proposed framework demonstrates how spatio-temporal data warehousing can address schema complexity,improve query performance,and support multidimensional visualization.In doing so,it provides a foundation for integrating seismic analysis into broader big data-driven intelligent decision systems for disaster resilience,risk mitigation,and emergency management.
文摘在小学高年级英语语篇教学中,存在学生思维浅表化、问题设计碎片化、旧版教材适配难这三个痛点。以译林版英语教材六年级上册Unit 4 Then and now中Story time的教学为例,教师立足教材文本,构建“课前定问—课初引链—课中解链—课后拓链—全程评链”的五步闭环,用大问题拉主线、小问题搭台阶,能激活学生语篇学习内驱力,实现英语教学从“知识传递”到“素养培养”的转变。
文摘With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service resources,uneven distribution,and prominent supply-demand contradictions have seriously affected service quality.Big data technology,with core advantages including data collection,analysis and mining,and accurate prediction,provides a new solution for the allocation of community elderly care service resources.This paper systematically studies the application value of big data technology in the allocation of community elderly care service resources from three aspects:resource allocation efficiency,service accuracy,and management intelligence.Combined with practical needs,it proposes optimal allocation strategies such as building a big data analysis platform and accurately grasping the elderly’s care needs,striving to provide operable path references for the construction of community elderly care service systems,promoting the early realization of the elderly care service goal of“adequate support and proper care for the elderly”,and boosting the high-quality development of China’s elderly care service industry.
基金Hunan Provincial Natural Science Foundation of China,Grant/Award Numbers:2024JJ6289,2023JJ60464,2023JJ60334Changsha City Technology Program,Grant/Award Number:kq2403120+1 种基金Climb Plan of Hunan Cancer Hospital,Grant/Award Numbers:ZX2021005,QH2023006High-Level Talent Support Program of Hunan Cancer Hospital,Grant/Award Number:20250731-1050。
文摘The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physician-centered perspective,these technologies can potentially be used to improve diagnostic precision,support individualized treatment planning,enhance longitudinal patient management,and accelerate both clinical and translational research.In this review,we synthesize the core AI methodologies most relevant to oncology-machine learning,deep learning,and large language models-and examine how they interact with established and emerging oncology data platforms.We further highlight practical use cases in clinical workflows and research pipelines,emphasizing opportunities for advancing precision cancer care while also addressing challenges associated with data heterogeneity,model generalizability,privacy protection,and real-world implementation.By underscoring the synergistic value of AI and big data,this review aims to inform the development of clinically meaningful,context-adapted strategies that promote translational innovation in both global and locally resourced healthcare environments.
文摘With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues such as computer memory capacity and distributed storage of data across different locations,and are unable to directly apply to large-scale data sets.Therefore,in the context of big data,designing efficient and theoretically guaranteed statistical learning and inference algorithms has become a key issue that the current field of statistics urgently needs to address.In this paper,the application status of statistical analysis methods in the big data environment was systematically reviewed,and its future development directions were analyzed to provide reference and support for the further development of theory and methods of the statistical analysis of big data.
文摘大孔径CT基于X线断层扫描原理的超大孔径设计,能容纳肥胖患者以及携带医疗设备的特殊患者,其凭借高清晰成像为放疗计划的制定与实施提供了关键的解剖学依据,是现代放疗“精准化”转型的重要支撑[1]。我院于2017年引进飞利浦Brilliance Big Bore CT设备,截至2024年底,已累计完成5000余例肿瘤患者的放疗定位扫描。
文摘The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threats is both necessary and complex,yet these interconnected healthcare settings generate enormous amounts of heterogeneous data.Traditional Intrusion Detection Systems(IDS),which are generally centralized and machine learning-based,often fail to address the rapidly changing nature of cyberattacks and are challenged by ethical concerns related to patient data privacy.Moreover,traditional AI-driven IDS usually face challenges in handling large-scale,heterogeneous healthcare data while ensuring data privacy and operational efficiency.To address these issues,emerging technologies such as Big Data Analytics(BDA)and Federated Learning(FL)provide a hybrid framework for scalable,adaptive intrusion detection in IoT-driven healthcare systems.Big data techniques enable processing large-scale,highdimensional healthcare data,and FL can be used to train a model in a decentralized manner without transferring raw data,thereby maintaining privacy between institutions.This research proposes a privacy-preserving Federated Learning–based model that efficiently detects cyber threats in connected healthcare systems while ensuring distributed big data processing,privacy,and compliance with ethical regulations.To strengthen the reliability of the reported findings,the resultswere validated using cross-dataset testing and 95%confidence intervals derived frombootstrap analysis,confirming consistent performance across heterogeneous healthcare data distributions.This solution takes a significant step toward securing next-generation healthcare infrastructure by combining scalability,privacy,adaptability,and earlydetection capabilities.The proposed global model achieves a test accuracy of 99.93%±0.03(95%CI)and amiss-rate of only 0.07%±0.02,representing state-of-the-art performance in privacy-preserving intrusion detection.The proposed FL-driven IDS framework offers an efficient,privacy-preserving,and scalable solution for securing next-generation healthcare infrastructures by combining adaptability,early detection,and ethical data management.
基金funded by the Technical Development(Entrusted)Project of Science and Department of SINOPEC(Grant No.P23240-4)the National Natural Science Foundation of China(Grant Nos.42172165,42272143 and 2025ZD1403901-05)。
文摘The Wufeng–Longmaxi Formation derives its name from the Upper Ordovician Wufeng Formation and the Lower Silurian Longmaxi Formation,found in sequence in the Sichuan Basin.This formation hosts rich shale gas reservoirs,and its shale gas enrichment patterns are examined in this study using data from 1197 shale samples collected from 14 wells.Five basic and three key parameters,eight in all,are assessed for each sample.The five basic parameters include burial depth and the contents of four mineral types—quartz,clay,carbonate,and other minerals;the three key parameters,representing shale gas enrichment,are total organic carbon(TOC)content,porosity,and gas content.The SHapley Additive exPlanations(SHAP)analysis originated in game theory is used here in an interpretable machine learning framework,to address issues of heterogeneous data structure,noisy relationships,and multi-objective optimization.An evaluation of the ranking,contribution values,and conditions of changes for these parameters offers new quantitative insights into shale gas enrichment patterns.A quantitative analysis of the relationship between data-sets identifies the primary factors controlling TOC,porosity,and gas content of shale gas reservoirs.The results show that TOC and porosity jointly influence gas content;mineral content has a significant impact on both,TOC and porosity;and the burial depth governs porosity which,in turn,affects the conditions under which shale gas is preserved.Input parameter thresholds are also determined and provide a basis for the establishment of quantitative criteria to evaluate shale gas enrichment.The predictive accuracy of the model used in this study is significantly improved by the step-wise addition of two input parameters,namely TOC and porosity,separately and together.Thus,the game theory method in big data-driven analysis uses a combination of TOC and porosity to evaluate the gas content with encouraging results—suggesting that these are the key parameters that indicate source rock and reservoir properties.
文摘The development of remote sensing has seen the creation of a global measurement infrastructure of sustainable development due to growing multipolar archives,rising revisit frequency,and the availability of cloud-accessible platforms of Earth observation.This review summarizes how remote sensing big data is being organized into decision-grade sustainability intelligence,the new approaches to analytics,and how Sustainable Development Goals(SDGs)-oriented application pathways inter-relate action pathways that bridge observations with action.The terminologies like new data ecosystem,data readiness and interoperability,changing economics of scalable computation,and detailing the functions of diversity of modalities(optical,Synthetic Aperture Radar—SAR,thermal,Light Detection and Ranging—LiDAR,hyperspectral)have been defined.These themes of analytics,which are transforming the practice of operational analytics,are then condensed:foundations and self-supervised learning of transferable representations,multi-modal fusion to gap fill and richer inference,spatiotemporal intelligence to trend of early warning,physics-aware hybrid methods to enhance robustness and meaning under non-stationary conditions.Across the climate risk,food systems,water resources,sustainable cities,ecosystems and biodiversity,energy transitions,and health exposure pathways,the roles of Earth Observation(EO)products as direct measures and proxies,and concepts of validating,semantic comparability,and communicating uncertainties play a key role in EO products becoming credible when faced with high-stakes deployment decisions.Lastly,we chart world ways of implementation via monitoring services,early warning systems,and systems of multiple regimes,and previously underline cross-cutting priorities,scalable structures in validation,performance,so that domains of shift,agreeable governance,and Dual-use risk safeguards,and sustainable lifecycle support of EO services.These priorities form a realistic set of priorities on the alignment of remote sensing innovation with quantifiable SDGs progress.
文摘On October 18,2017,the 19th National Congress Report called for the implementation of the Healthy China Strategy.The development of biomedical data plays a pivotal role in advancing this strategy.Since the 18th National Congress of the Communist Party of China,China has vigorously promoted the integration and implementation of the Healthy China and Digital China strategies.The National Health Commission has prioritized the development of health and medical big data,issuing policies to promote standardized applica-tions and foster innovation in"Internet+Healthcare."Biomedical data has significantly contributed to preci-sion medicine,personalized health management,drug development,disease diagnosis,public health monitor-ing,and epidemic prediction capabilities.
基金partially supported by the Construction of Collaborative Innovation Center of Beijing Academy of Agricultural and Forestry Sciences(KJCX20240406)the Beijing Natural Science Foundation(JQ24037)+1 种基金the National Natural Science Foundation of China(32330075)the Earmarked Fund for China Agriculture Research System(CARS-02 and CARS-54)。
文摘The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This era integrates biotechnology,artificial intelligence(AI),and big data information technology.In contrast,China is still in a transition period between stages 2.0 and 3.0,which primarily relies on conventional selection and molecular breeding.In the context of increasingly complex international situations,accurately identifying core issues in China's seed industry innovation and seizing the frontier of international seed technology are strategically important.These efforts are essential for ensuring food security and revitalizing the seed industry.This paper systematically analyzes the characteristics of crop breeding data from artificial selection to intelligent design breeding.It explores the applications and development trends of AI and big data in modern crop breeding from several key perspectives.These include highthroughput phenotype acquisition and analysis,multiomics big data database and management system construction,AI-based multiomics integrated analysis,and the development of intelligent breeding software tools based on biological big data and AI technology.Based on an in-depth analysis of the current status and challenges of China's seed industry technology development,we propose strategic goals and key tasks for China's new generation of AI and big data-driven intelligent design breeding.These suggestions aim to accelerate the development of an intelligent-driven crop breeding engineering system that features large-scale gene mining,efficient gene manipulation,engineered variety design,and systematized biobreeding.This study provides a theoretical basis and practical guidance for the development of China's seed industry technology.
基金sponsored by the U.S.Department of Housing and Urban Development(Grant No.NJLTS0027-22)The opinions expressed in this study are the authors alone,and do not represent the U.S.Depart-ment of HUD’s opinions.
文摘This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models offer insights, they fall short in presenting a holistic view of complex urban challenges. System dynamics (SD) models that are often utilized to provide holistic, systematic understanding of a research subject, like the urban system, emerge as valuable tools, but data scarcity and theoretical inadequacy pose challenges. The research reviews relevant papers on recent SD model applications in urban sustainability since 2018, categorizing them based on nine key indicators. Among the reviewed papers, data limitations and model assumptions were identified as ma jor challenges in applying SD models to urban sustainability. This led to exploring the transformative potential of big data analytics, a rare approach in this field as identified by this study, to enhance SD models’ empirical foundation. Integrating big data could provide data-driven calibration, potentially improving predictive accuracy and reducing reliance on simplified assumptions. The paper concludes by advocating for new approaches that reduce assumptions and promote real-time applicable models, contributing to a comprehensive understanding of urban sustainability through the synergy of big data and SD models.
基金Strategic Priority Research Program of the Chinese Academy of Sciences,No.XDB0740000National Key Research and Development Program of China,No.2022YFB3904200,No.2022YFF0711601+1 种基金Key Project of Innovation LREIS,No.PI009National Natural Science Foundation of China,No.42471503。
文摘Deep-time Earth research plays a pivotal role in deciphering the rates,patterns,and mechanisms of Earth's evolutionary processes throughout geological history,providing essential scientific foundations for climate prediction,natural resource exploration,and sustainable planetary stewardship.To advance Deep-time Earth research in the era of big data and artificial intelligence,the International Union of Geological Sciences initiated the“Deeptime Digital Earth International Big Science Program”(DDE)in 2019.At the core of this ambitious program lies the development of geoscience knowledge graphs,serving as a transformative knowledge infrastructure that enables the integration,sharing,mining,and analysis of heterogeneous geoscience big data.The DDE knowledge graph initiative has made significant strides in three critical dimensions:(1)establishing a unified knowledge structure across geoscience disciplines that ensures consistent representation of geological entities and their interrelationships through standardized ontologies and semantic frameworks;(2)developing a robust and scalable software infrastructure capable of supporting both expert-driven and machine-assisted knowledge engineering for large-scale graph construction and management;(3)implementing a comprehensive three-tiered architecture encompassing basic,discipline-specific,and application-oriented knowledge graphs,spanning approximately 20 geoscience disciplines.Through its open knowledge framework and international collaborative network,this initiative has fostered multinational research collaborations,establishing a robust foundation for next-generation geoscience research while propelling the discipline toward FAIR(Findable,Accessible,Interoperable,Reusable)data practices in deep-time Earth systems research.
基金funded by the National Natural Science Foundation of China(No.42220104008)。
文摘Research into metamorphism plays a pivotal role in reconstructing the evolution of continent,particularly through the study of ancient rocks that are highly susceptible to metamorphic alterations due to multiple tectonic activities.In the big data era,the establishment of new data platforms and the application of big data methods have become a focus for metamorphic rocks.Significant progress has been made in creating specialized databases,compiling comprehensive datasets,and utilizing data analytics to address complex scientific questions.However,many existing databases are inadequate in meeting the specific requirements of metamorphic research,resulting from a substantial amount of valuable data remaining uncollected.Therefore,constructing new databases that can cope with the development of the data era is necessary.This article provides an extensive review of existing databases related to metamorphic rocks and discusses data-driven studies in this.Accordingly,several crucial factors that need to be taken into consideration in the establishment of specialized metamorphic databases are identified,aiming to leverage data-driven applications to achieve broader scientific objectives in metamorphic research.
基金supported by the National Natural Science Foundation of China(32370703)the CAMS Innovation Fund for Medical Sciences(CIFMS)(2022-I2M-1-021,2021-I2M-1-061)the Major Project of Guangzhou National Labora-tory(GZNL2024A01015).
文摘Viral infectious diseases,characterized by their intricate nature and wide-ranging diversity,pose substantial challenges in the domain of data management.The vast volume of data generated by these diseases,spanning from the molecular mechanisms within cells to large-scale epidemiological patterns,has surpassed the capabilities of traditional analytical methods.In the era of artificial intelligence(AI)and big data,there is an urgent necessity for the optimization of these analytical methods to more effectively handle and utilize the information.Despite the rapid accumulation of data associated with viral infections,the lack of a comprehensive framework for integrating,selecting,and analyzing these datasets has left numerous researchers uncertain about which data to select,how to access it,and how to utilize it most effectively in their research.This review endeavors to fill these gaps by exploring the multifaceted nature of viral infectious diseases and summarizing relevant data across multiple levels,from the molecular details of pathogens to broad epidemiological trends.The scope extends from the micro-scale to the macro-scale,encompassing pathogens,hosts,and vectors.In addition to data summarization,this review thoroughly investigates various dataset sources.It also traces the historical evolution of data collection in the field of viral infectious diseases,highlighting the progress achieved over time.Simultaneously,it evaluates the current limitations that impede data utilization.Furthermore,we propose strategies to surmount these challenges,focusing on the development and application of advanced computational techniques,AI-driven models,and enhanced data integration practices.By providing a comprehensive synthesis of existing knowledge,this review is designed to guide future research and contribute to more informed approaches in the surveillance,prevention,and control of viral infectious diseases,particularly within the context of the expanding big-data landscape.
基金supported in part by the National Natural Science Foundation of China(92259202)the Guangzhou Municipal Health Commission(2023P-GX02)+2 种基金the Science and Technology Projects in Guangzhou(2024B01J1301)the Fundamental Research Funds for the Central Universities,Sun Yat-sen University Clinical Research 5010 Program(2024001)the Cancer Innovative Research Program of Sun Yat-sen University Cancer Center(CIRP-SYSUCC-0010).
文摘1.Introduction Nasopharyngeal carcinoma(NPC)has an imbalanced geographical and ethnic distribution,with notably high incidence rates in Southeastern Asia and China.China accounted for 42.4%of the newly diagnosed cases worldwide in 2022.1NPC is sensitive to irradiation,and radiotherapy is the mainstay curative treatment modality.2The widespread use of intensity-modulated radiation therapy(IMRT)and image-guided radiotherapy(IGRT)has achieved great advances in survival outcomes and toxicity profiles among NPC patients.1In radiotherapy of NPC,the tumor’s proximity to critical structures demands accuracy in tumor delineation in order to avoid radiation-induced toxicities.However,tumor target delineation for radiotherapy of NPC is labor-intensive and radiation oncologists’proficiency varied considerably.3In recent years,the advent of big data analytics and artificial intelligence(AI)has opened up new avenues for improving the precision and efficacy of radiotherapy and individualized treatment in NPC management.3-6In this article,we explored how big data,AI-assisted delineation,radiotherapy planning,and adaptive radiotherapy(ART)are transforming clinical decision-making in NPC treatment.We also provided an outlook on the historical development of AI and big data,their current dominance in oncological radiotherapy,and their projected impact on future clinical practice(Figure 1).
基金2024 Anqing Normal University University-Level Key Project(ZK2024062D)。
文摘This study examines the Big Data Collection and Preprocessing course at Anhui Institute of Information Engineering,implementing a hybrid teaching reform using the Bosi Smart Learning Platform.The proposed hybrid model follows a“three-stage”and“two-subject”framework,incorporating a structured design for teaching content and assessment methods before,during,and after class.Practical results indicate that this approach significantly enhances teaching effectiveness and improves students’learning autonomy.
基金supported By Grant (PLN2022-14) of State Key Laboratory of Oil and Gas Reservoir Geology and Exploitation (Southwest Petroleum University)。
文摘Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of these data has not been well stored,managed and mined.With the development of cloud computing technology,it provides a rare development opportunity for logging big data private cloud.The traditional petrophysical evaluation and interpretation model has encountered great challenges in the face of new evaluation objects.The solution research of logging big data distributed storage,processing and learning functions integrated in logging big data private cloud has not been carried out yet.To establish a distributed logging big-data private cloud platform centered on a unifi ed learning model,which achieves the distributed storage and processing of logging big data and facilitates the learning of novel knowledge patterns via the unifi ed logging learning model integrating physical simulation and data models in a large-scale functional space,thus resolving the geo-engineering evaluation problem of geothermal fi elds.Based on the research idea of“logging big data cloud platform-unifi ed logging learning model-large function space-knowledge learning&discovery-application”,the theoretical foundation of unified learning model,cloud platform architecture,data storage and learning algorithm,arithmetic power allocation and platform monitoring,platform stability,data security,etc.have been carried on analysis.The designed logging big data cloud platform realizes parallel distributed storage and processing of data and learning algorithms.The feasibility of constructing a well logging big data cloud platform based on a unifi ed learning model of physics and data is analyzed in terms of the structure,ecology,management and security of the cloud platform.The case study shows that the logging big data cloud platform has obvious technical advantages over traditional logging evaluation methods in terms of knowledge discovery method,data software and results sharing,accuracy,speed and complexity.