Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from sei...Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from seismic networks,satellite observations,and geospatial repositories,creating the need for scalable infrastructures capable of integrating and analyzing such data to support intelligent decision-making.Data warehousing technologies provide a robust foundation for this purpose;however,existing earthquake-oriented data warehouses remain limited,often relying on simplified schemas,domain-specific analytics,or cataloguing efforts.This paper presents the design and implementation of a spatio-temporal data warehouse for seismic activity.The framework integrates spatial and temporal dimensions in a unified schema and introduces a novel array-based approach for managing many-to-many relationships between facts and dimensions without intermediate bridge tables.A comparative evaluation against a conventional bridge-table schema demonstrates that the array-based design improves fact-centric query performance,while the bridge-table schema remains advantageous for dimension-centric queries.To reconcile these trade-offs,a hybrid schema is proposed that retains both representations,ensuring balanced efficiency across heterogeneous workloads.The proposed framework demonstrates how spatio-temporal data warehousing can address schema complexity,improve query performance,and support multidimensional visualization.In doing so,it provides a foundation for integrating seismic analysis into broader big data-driven intelligent decision systems for disaster resilience,risk mitigation,and emergency management.展开更多
The goal of this study was to determine the spatiotemporal characteristics of mangrove distribution and fragmentation patterns from 1988 through 2019 in Dongzhaigang.Land cover datasets were generated for Dongzhaigang...The goal of this study was to determine the spatiotemporal characteristics of mangrove distribution and fragmentation patterns from 1988 through 2019 in Dongzhaigang.Land cover datasets were generated for Dongzhaigang for multiple years via a decision tree method based on a classification and regression tree(CART)algorithm using Landsat time series images.Spatiotemporal transform and fragmentation patterns of mangrove distribution were separately assessed with a transfer matrix of land cover types and a landscape pattern index.The classification method combined with multi-band images showed good accuracy,with overall accuracy higher than 90%.Mangrove areas in 1988,1999,2009,and 2019 were 2050,1875,1818,and 1750 ha,respectively,with decreases mainly due to conversion to aquaculture ponds and farmland.A mangrove growth index(MGI)was proposed,reflecting the water-mangrove relationship,showing positive mangrove growth from 1988–2009 and negative growth from 2009–2019.Study results indicated anthropogenic factors play a leading role in the extent and scale of mangrove effects over the past 30 years.According to the analysis results,corresponding management and protection measures are proposed to provide reference for the sustainable development of Dongzhaigang Mangrove Wetland ecosystem.展开更多
With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service...With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service resources,uneven distribution,and prominent supply-demand contradictions have seriously affected service quality.Big data technology,with core advantages including data collection,analysis and mining,and accurate prediction,provides a new solution for the allocation of community elderly care service resources.This paper systematically studies the application value of big data technology in the allocation of community elderly care service resources from three aspects:resource allocation efficiency,service accuracy,and management intelligence.Combined with practical needs,it proposes optimal allocation strategies such as building a big data analysis platform and accurately grasping the elderly’s care needs,striving to provide operable path references for the construction of community elderly care service systems,promoting the early realization of the elderly care service goal of“adequate support and proper care for the elderly”,and boosting the high-quality development of China’s elderly care service industry.展开更多
With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues su...With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues such as computer memory capacity and distributed storage of data across different locations,and are unable to directly apply to large-scale data sets.Therefore,in the context of big data,designing efficient and theoretically guaranteed statistical learning and inference algorithms has become a key issue that the current field of statistics urgently needs to address.In this paper,the application status of statistical analysis methods in the big data environment was systematically reviewed,and its future development directions were analyzed to provide reference and support for the further development of theory and methods of the statistical analysis of big data.展开更多
The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physici...The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physician-centered perspective,these technologies can potentially be used to improve diagnostic precision,support individualized treatment planning,enhance longitudinal patient management,and accelerate both clinical and translational research.In this review,we synthesize the core AI methodologies most relevant to oncology-machine learning,deep learning,and large language models-and examine how they interact with established and emerging oncology data platforms.We further highlight practical use cases in clinical workflows and research pipelines,emphasizing opportunities for advancing precision cancer care while also addressing challenges associated with data heterogeneity,model generalizability,privacy protection,and real-world implementation.By underscoring the synergistic value of AI and big data,this review aims to inform the development of clinically meaningful,context-adapted strategies that promote translational innovation in both global and locally resourced healthcare environments.展开更多
The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threa...The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threats is both necessary and complex,yet these interconnected healthcare settings generate enormous amounts of heterogeneous data.Traditional Intrusion Detection Systems(IDS),which are generally centralized and machine learning-based,often fail to address the rapidly changing nature of cyberattacks and are challenged by ethical concerns related to patient data privacy.Moreover,traditional AI-driven IDS usually face challenges in handling large-scale,heterogeneous healthcare data while ensuring data privacy and operational efficiency.To address these issues,emerging technologies such as Big Data Analytics(BDA)and Federated Learning(FL)provide a hybrid framework for scalable,adaptive intrusion detection in IoT-driven healthcare systems.Big data techniques enable processing large-scale,highdimensional healthcare data,and FL can be used to train a model in a decentralized manner without transferring raw data,thereby maintaining privacy between institutions.This research proposes a privacy-preserving Federated Learning–based model that efficiently detects cyber threats in connected healthcare systems while ensuring distributed big data processing,privacy,and compliance with ethical regulations.To strengthen the reliability of the reported findings,the resultswere validated using cross-dataset testing and 95%confidence intervals derived frombootstrap analysis,confirming consistent performance across heterogeneous healthcare data distributions.This solution takes a significant step toward securing next-generation healthcare infrastructure by combining scalability,privacy,adaptability,and earlydetection capabilities.The proposed global model achieves a test accuracy of 99.93%±0.03(95%CI)and amiss-rate of only 0.07%±0.02,representing state-of-the-art performance in privacy-preserving intrusion detection.The proposed FL-driven IDS framework offers an efficient,privacy-preserving,and scalable solution for securing next-generation healthcare infrastructures by combining adaptability,early detection,and ethical data management.展开更多
The development of remote sensing has seen the creation of a global measurement infrastructure of sustainable development due to growing multipolar archives,rising revisit frequency,and the availability of cloud-acces...The development of remote sensing has seen the creation of a global measurement infrastructure of sustainable development due to growing multipolar archives,rising revisit frequency,and the availability of cloud-accessible platforms of Earth observation.This review summarizes how remote sensing big data is being organized into decision-grade sustainability intelligence,the new approaches to analytics,and how Sustainable Development Goals(SDGs)-oriented application pathways inter-relate action pathways that bridge observations with action.The terminologies like new data ecosystem,data readiness and interoperability,changing economics of scalable computation,and detailing the functions of diversity of modalities(optical,Synthetic Aperture Radar—SAR,thermal,Light Detection and Ranging—LiDAR,hyperspectral)have been defined.These themes of analytics,which are transforming the practice of operational analytics,are then condensed:foundations and self-supervised learning of transferable representations,multi-modal fusion to gap fill and richer inference,spatiotemporal intelligence to trend of early warning,physics-aware hybrid methods to enhance robustness and meaning under non-stationary conditions.Across the climate risk,food systems,water resources,sustainable cities,ecosystems and biodiversity,energy transitions,and health exposure pathways,the roles of Earth Observation(EO)products as direct measures and proxies,and concepts of validating,semantic comparability,and communicating uncertainties play a key role in EO products becoming credible when faced with high-stakes deployment decisions.Lastly,we chart world ways of implementation via monitoring services,early warning systems,and systems of multiple regimes,and previously underline cross-cutting priorities,scalable structures in validation,performance,so that domains of shift,agreeable governance,and Dual-use risk safeguards,and sustainable lifecycle support of EO services.These priorities form a realistic set of priorities on the alignment of remote sensing innovation with quantifiable SDGs progress.展开更多
In 2007,China surpassed the USA to become the largest carbon emitter in the world.China has promised a 60%–65%reduction in carbon emissions per unit GDP by 2030,compared to the baseline of 2005.Therefore,it is import...In 2007,China surpassed the USA to become the largest carbon emitter in the world.China has promised a 60%–65%reduction in carbon emissions per unit GDP by 2030,compared to the baseline of 2005.Therefore,it is important to obtain accurate dynamic information on the spatial and temporal patterns of carbon emissions and carbon footprints to support formulating effective national carbon emission reduction policies.This study attempts to build a carbon emission panel data model that simulates carbon emissions in China from 2000–2013 using nighttime lighting data and carbon emission statistics data.By applying the Exploratory Spatial-Temporal Data Analysis(ESTDA)framework,this study conducted an analysis on the spatial patterns and dynamic spatial-temporal interactions of carbon footprints from 2001–2013.The improved Tapio decoupling model was adopted to investigate the levels of coupling or decoupling between the carbon emission load and economic growth in 336 prefecture-level units.The results show that,firstly,high accuracy was achieved by the model in simulating carbon emissions.Secondly,the total carbon footprints and carbon deficits across China increased with average annual growth rates of 4.82%and 5.72%,respectively.The overall carbon footprints and carbon deficits were larger in the North than that in the South.There were extremely significant spatial autocorrelation features in the carbon footprints of prefecture-level units.Thirdly,the relative lengths of the Local Indicators of Spatial Association(LISA)time paths were longer in the North than that in the South,and they increased from the coastal to the central and western regions.Lastly,the overall decoupling index was mainly a weak decoupling type,but the number of cities with this weak decoupling continued to decrease.The unsustainable development trend of China’s economic growth and carbon emission load will continue for some time.展开更多
The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This...The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This era integrates biotechnology,artificial intelligence(AI),and big data information technology.In contrast,China is still in a transition period between stages 2.0 and 3.0,which primarily relies on conventional selection and molecular breeding.In the context of increasingly complex international situations,accurately identifying core issues in China's seed industry innovation and seizing the frontier of international seed technology are strategically important.These efforts are essential for ensuring food security and revitalizing the seed industry.This paper systematically analyzes the characteristics of crop breeding data from artificial selection to intelligent design breeding.It explores the applications and development trends of AI and big data in modern crop breeding from several key perspectives.These include highthroughput phenotype acquisition and analysis,multiomics big data database and management system construction,AI-based multiomics integrated analysis,and the development of intelligent breeding software tools based on biological big data and AI technology.Based on an in-depth analysis of the current status and challenges of China's seed industry technology development,we propose strategic goals and key tasks for China's new generation of AI and big data-driven intelligent design breeding.These suggestions aim to accelerate the development of an intelligent-driven crop breeding engineering system that features large-scale gene mining,efficient gene manipulation,engineered variety design,and systematized biobreeding.This study provides a theoretical basis and practical guidance for the development of China's seed industry technology.展开更多
This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models ...This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models offer insights, they fall short in presenting a holistic view of complex urban challenges. System dynamics (SD) models that are often utilized to provide holistic, systematic understanding of a research subject, like the urban system, emerge as valuable tools, but data scarcity and theoretical inadequacy pose challenges. The research reviews relevant papers on recent SD model applications in urban sustainability since 2018, categorizing them based on nine key indicators. Among the reviewed papers, data limitations and model assumptions were identified as ma jor challenges in applying SD models to urban sustainability. This led to exploring the transformative potential of big data analytics, a rare approach in this field as identified by this study, to enhance SD models’ empirical foundation. Integrating big data could provide data-driven calibration, potentially improving predictive accuracy and reducing reliance on simplified assumptions. The paper concludes by advocating for new approaches that reduce assumptions and promote real-time applicable models, contributing to a comprehensive understanding of urban sustainability through the synergy of big data and SD models.展开更多
Viral infectious diseases,characterized by their intricate nature and wide-ranging diversity,pose substantial challenges in the domain of data management.The vast volume of data generated by these diseases,spanning fr...Viral infectious diseases,characterized by their intricate nature and wide-ranging diversity,pose substantial challenges in the domain of data management.The vast volume of data generated by these diseases,spanning from the molecular mechanisms within cells to large-scale epidemiological patterns,has surpassed the capabilities of traditional analytical methods.In the era of artificial intelligence(AI)and big data,there is an urgent necessity for the optimization of these analytical methods to more effectively handle and utilize the information.Despite the rapid accumulation of data associated with viral infections,the lack of a comprehensive framework for integrating,selecting,and analyzing these datasets has left numerous researchers uncertain about which data to select,how to access it,and how to utilize it most effectively in their research.This review endeavors to fill these gaps by exploring the multifaceted nature of viral infectious diseases and summarizing relevant data across multiple levels,from the molecular details of pathogens to broad epidemiological trends.The scope extends from the micro-scale to the macro-scale,encompassing pathogens,hosts,and vectors.In addition to data summarization,this review thoroughly investigates various dataset sources.It also traces the historical evolution of data collection in the field of viral infectious diseases,highlighting the progress achieved over time.Simultaneously,it evaluates the current limitations that impede data utilization.Furthermore,we propose strategies to surmount these challenges,focusing on the development and application of advanced computational techniques,AI-driven models,and enhanced data integration practices.By providing a comprehensive synthesis of existing knowledge,this review is designed to guide future research and contribute to more informed approaches in the surveillance,prevention,and control of viral infectious diseases,particularly within the context of the expanding big-data landscape.展开更多
On October 18,2017,the 19th National Congress Report called for the implementation of the Healthy China Strategy.The development of biomedical data plays a pivotal role in advancing this strategy.Since the 18th Nation...On October 18,2017,the 19th National Congress Report called for the implementation of the Healthy China Strategy.The development of biomedical data plays a pivotal role in advancing this strategy.Since the 18th National Congress of the Communist Party of China,China has vigorously promoted the integration and implementation of the Healthy China and Digital China strategies.The National Health Commission has prioritized the development of health and medical big data,issuing policies to promote standardized applica-tions and foster innovation in"Internet+Healthcare."Biomedical data has significantly contributed to preci-sion medicine,personalized health management,drug development,disease diagnosis,public health monitor-ing,and epidemic prediction capabilities.展开更多
Research into metamorphism plays a pivotal role in reconstructing the evolution of continent,particularly through the study of ancient rocks that are highly susceptible to metamorphic alterations due to multiple tecto...Research into metamorphism plays a pivotal role in reconstructing the evolution of continent,particularly through the study of ancient rocks that are highly susceptible to metamorphic alterations due to multiple tectonic activities.In the big data era,the establishment of new data platforms and the application of big data methods have become a focus for metamorphic rocks.Significant progress has been made in creating specialized databases,compiling comprehensive datasets,and utilizing data analytics to address complex scientific questions.However,many existing databases are inadequate in meeting the specific requirements of metamorphic research,resulting from a substantial amount of valuable data remaining uncollected.Therefore,constructing new databases that can cope with the development of the data era is necessary.This article provides an extensive review of existing databases related to metamorphic rocks and discusses data-driven studies in this.Accordingly,several crucial factors that need to be taken into consideration in the establishment of specialized metamorphic databases are identified,aiming to leverage data-driven applications to achieve broader scientific objectives in metamorphic research.展开更多
This study examines the Big Data Collection and Preprocessing course at Anhui Institute of Information Engineering,implementing a hybrid teaching reform using the Bosi Smart Learning Platform.The proposed hybrid model...This study examines the Big Data Collection and Preprocessing course at Anhui Institute of Information Engineering,implementing a hybrid teaching reform using the Bosi Smart Learning Platform.The proposed hybrid model follows a“three-stage”and“two-subject”framework,incorporating a structured design for teaching content and assessment methods before,during,and after class.Practical results indicate that this approach significantly enhances teaching effectiveness and improves students’learning autonomy.展开更多
Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of th...Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of these data has not been well stored,managed and mined.With the development of cloud computing technology,it provides a rare development opportunity for logging big data private cloud.The traditional petrophysical evaluation and interpretation model has encountered great challenges in the face of new evaluation objects.The solution research of logging big data distributed storage,processing and learning functions integrated in logging big data private cloud has not been carried out yet.To establish a distributed logging big-data private cloud platform centered on a unifi ed learning model,which achieves the distributed storage and processing of logging big data and facilitates the learning of novel knowledge patterns via the unifi ed logging learning model integrating physical simulation and data models in a large-scale functional space,thus resolving the geo-engineering evaluation problem of geothermal fi elds.Based on the research idea of“logging big data cloud platform-unifi ed logging learning model-large function space-knowledge learning&discovery-application”,the theoretical foundation of unified learning model,cloud platform architecture,data storage and learning algorithm,arithmetic power allocation and platform monitoring,platform stability,data security,etc.have been carried on analysis.The designed logging big data cloud platform realizes parallel distributed storage and processing of data and learning algorithms.The feasibility of constructing a well logging big data cloud platform based on a unifi ed learning model of physics and data is analyzed in terms of the structure,ecology,management and security of the cloud platform.The case study shows that the logging big data cloud platform has obvious technical advantages over traditional logging evaluation methods in terms of knowledge discovery method,data software and results sharing,accuracy,speed and complexity.展开更多
1.Introduction Nasopharyngeal carcinoma(NPC)has an imbalanced geographical and ethnic distribution,with notably high incidence rates in Southeastern Asia and China.China accounted for 42.4%of the newly diagnosed cases...1.Introduction Nasopharyngeal carcinoma(NPC)has an imbalanced geographical and ethnic distribution,with notably high incidence rates in Southeastern Asia and China.China accounted for 42.4%of the newly diagnosed cases worldwide in 2022.1NPC is sensitive to irradiation,and radiotherapy is the mainstay curative treatment modality.2The widespread use of intensity-modulated radiation therapy(IMRT)and image-guided radiotherapy(IGRT)has achieved great advances in survival outcomes and toxicity profiles among NPC patients.1In radiotherapy of NPC,the tumor’s proximity to critical structures demands accuracy in tumor delineation in order to avoid radiation-induced toxicities.However,tumor target delineation for radiotherapy of NPC is labor-intensive and radiation oncologists’proficiency varied considerably.3In recent years,the advent of big data analytics and artificial intelligence(AI)has opened up new avenues for improving the precision and efficacy of radiotherapy and individualized treatment in NPC management.3-6In this article,we explored how big data,AI-assisted delineation,radiotherapy planning,and adaptive radiotherapy(ART)are transforming clinical decision-making in NPC treatment.We also provided an outlook on the historical development of AI and big data,their current dominance in oncological radiotherapy,and their projected impact on future clinical practice(Figure 1).展开更多
With the rapid development of intelligent manufacturing,industrial big data play an increasingly crucial role in the digital transformation of enterprises.However,current industrial big data platforms still face chall...With the rapid development of intelligent manufacturing,industrial big data play an increasingly crucial role in the digital transformation of enterprises.However,current industrial big data platforms still face challenges in data acquisition,processing,and visualization,including data processing inefficiencies,suboptimal storage solutions,and insufficient visualization experiences,which are often exacerbated by inherent data quality issues such as noise and outliers.To address these problems,this study proposes an industrial big data processing framework based on Flink and builds a data presentation system by combining Grafana and ECharts.The system collects data through enterprise sensors,utilizes Kafka message queues for data buffering,and uses Flink for efficient real-time data processing,incorporating foundational data cleansing techniques and strategies for mitigating common noise and anomalies.For data storage,MySQL is employed for static data,and InfluxDB is used for real-time data to improve storage efficiency.In terms of data visualization,Grafana displays real-time data,whereas ECharts is used for static data,offering users an intuitive and comprehensive data display interface.This study aims to provide an efficient and customizable industrial big data solution,with an emphasis on improving data reliability for visualization,to help enterprises monitor equipment information in real time,obtain effective information,and accelerate their intelligent transformation process.展开更多
In the era of big data,the financial industry is undergoing profound changes.By integrating multiple data sources such as transaction records,customer interactions,market trends,and regulatory requirements,big data te...In the era of big data,the financial industry is undergoing profound changes.By integrating multiple data sources such as transaction records,customer interactions,market trends,and regulatory requirements,big data technology has significantly improved the decision-making efficiency,customer insight,and risk management capabilities of financial institutions.The financial industry has become a pioneer in the application of big data technology,which is widely used in scenarios such as fraud detection,risk management,customer service optimization,and smart transactions.However,financial data security management also faces many challenges,including data breaches,privacy protection,compliance requirements,the complexity of emerging technologies,and the balance between data access and security.This article explores the major challenges of financial data security management,coping strategies,and the evolution of the regulatory environment,and it looks ahead to future trends,highlighting the important role of artificial intelligence and machine learning in financial data security.展开更多
As industrial production progresses toward digitalization,massive amounts of data have been collected,transmitted,and stored,with characteristics of large-scale,high-dimensional,heterogeneous,and spatiotemporal dynami...As industrial production progresses toward digitalization,massive amounts of data have been collected,transmitted,and stored,with characteristics of large-scale,high-dimensional,heterogeneous,and spatiotemporal dynamics.The high complexity of industrial big data poses challenges for the practical decision-making of domain experts,leading to ever-increasing needs for integrating computational intelligence with human perception into traditional data analysis.Industrial big data visualization integrates theoretical methods and practical technologies from multiple disciplines,including data mining,information visualization,computer graphics,and human-computer interaction,providing a highly effective manner for understanding and exploring the complex industrial processes.This review summarizes the state-of-the-art approaches,characterizes them with six visualization methods,and categorizes them based on analytical tasks and applications.Furthermore,key research challenges and potential future directions are identified.展开更多
In this paper,the application of agricultural big data in agricultural economic management is deeply explored,and its potential in promoting profit growth and innovation is analyzed.However,challenges persist in data ...In this paper,the application of agricultural big data in agricultural economic management is deeply explored,and its potential in promoting profit growth and innovation is analyzed.However,challenges persist in data collection and integration,limitations of analytical technologies,talent development,team building,and policy support when applying agricultural big data.Effective application strategies are proposed,including data-driven precision agriculture practices,construction of data integration and management platforms,data security and privacy protection strategies,as well as long-term planning and development strategies for agricultural big data,to maximize its impact on agricultural economic management.Future advancements require collaborative efforts in technological innovation,talent cultivation,and policy support,to realize the extensive application of agricultural big data in agricultural economic management and ensure sustainable industrial development.展开更多
文摘Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from seismic networks,satellite observations,and geospatial repositories,creating the need for scalable infrastructures capable of integrating and analyzing such data to support intelligent decision-making.Data warehousing technologies provide a robust foundation for this purpose;however,existing earthquake-oriented data warehouses remain limited,often relying on simplified schemas,domain-specific analytics,or cataloguing efforts.This paper presents the design and implementation of a spatio-temporal data warehouse for seismic activity.The framework integrates spatial and temporal dimensions in a unified schema and introduces a novel array-based approach for managing many-to-many relationships between facts and dimensions without intermediate bridge tables.A comparative evaluation against a conventional bridge-table schema demonstrates that the array-based design improves fact-centric query performance,while the bridge-table schema remains advantageous for dimension-centric queries.To reconcile these trade-offs,a hybrid schema is proposed that retains both representations,ensuring balanced efficiency across heterogeneous workloads.The proposed framework demonstrates how spatio-temporal data warehousing can address schema complexity,improve query performance,and support multidimensional visualization.In doing so,it provides a foundation for integrating seismic analysis into broader big data-driven intelligent decision systems for disaster resilience,risk mitigation,and emergency management.
基金financially supported by the National Natural Science Foundation of China(Nos.U2244225 and 42020104005)the Ministry of Education of China(111 Project)the Fundamental Research Funds for the Central Universities,China University of Geosciences(Wuhan)and China Geological Survey(No.DD20211391)。
文摘The goal of this study was to determine the spatiotemporal characteristics of mangrove distribution and fragmentation patterns from 1988 through 2019 in Dongzhaigang.Land cover datasets were generated for Dongzhaigang for multiple years via a decision tree method based on a classification and regression tree(CART)algorithm using Landsat time series images.Spatiotemporal transform and fragmentation patterns of mangrove distribution were separately assessed with a transfer matrix of land cover types and a landscape pattern index.The classification method combined with multi-band images showed good accuracy,with overall accuracy higher than 90%.Mangrove areas in 1988,1999,2009,and 2019 were 2050,1875,1818,and 1750 ha,respectively,with decreases mainly due to conversion to aquaculture ponds and farmland.A mangrove growth index(MGI)was proposed,reflecting the water-mangrove relationship,showing positive mangrove growth from 1988–2009 and negative growth from 2009–2019.Study results indicated anthropogenic factors play a leading role in the extent and scale of mangrove effects over the past 30 years.According to the analysis results,corresponding management and protection measures are proposed to provide reference for the sustainable development of Dongzhaigang Mangrove Wetland ecosystem.
文摘With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service resources,uneven distribution,and prominent supply-demand contradictions have seriously affected service quality.Big data technology,with core advantages including data collection,analysis and mining,and accurate prediction,provides a new solution for the allocation of community elderly care service resources.This paper systematically studies the application value of big data technology in the allocation of community elderly care service resources from three aspects:resource allocation efficiency,service accuracy,and management intelligence.Combined with practical needs,it proposes optimal allocation strategies such as building a big data analysis platform and accurately grasping the elderly’s care needs,striving to provide operable path references for the construction of community elderly care service systems,promoting the early realization of the elderly care service goal of“adequate support and proper care for the elderly”,and boosting the high-quality development of China’s elderly care service industry.
文摘With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues such as computer memory capacity and distributed storage of data across different locations,and are unable to directly apply to large-scale data sets.Therefore,in the context of big data,designing efficient and theoretically guaranteed statistical learning and inference algorithms has become a key issue that the current field of statistics urgently needs to address.In this paper,the application status of statistical analysis methods in the big data environment was systematically reviewed,and its future development directions were analyzed to provide reference and support for the further development of theory and methods of the statistical analysis of big data.
基金Hunan Provincial Natural Science Foundation of China,Grant/Award Numbers:2024JJ6289,2023JJ60464,2023JJ60334Changsha City Technology Program,Grant/Award Number:kq2403120+1 种基金Climb Plan of Hunan Cancer Hospital,Grant/Award Numbers:ZX2021005,QH2023006High-Level Talent Support Program of Hunan Cancer Hospital,Grant/Award Number:20250731-1050。
文摘The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physician-centered perspective,these technologies can potentially be used to improve diagnostic precision,support individualized treatment planning,enhance longitudinal patient management,and accelerate both clinical and translational research.In this review,we synthesize the core AI methodologies most relevant to oncology-machine learning,deep learning,and large language models-and examine how they interact with established and emerging oncology data platforms.We further highlight practical use cases in clinical workflows and research pipelines,emphasizing opportunities for advancing precision cancer care while also addressing challenges associated with data heterogeneity,model generalizability,privacy protection,and real-world implementation.By underscoring the synergistic value of AI and big data,this review aims to inform the development of clinically meaningful,context-adapted strategies that promote translational innovation in both global and locally resourced healthcare environments.
文摘The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threats is both necessary and complex,yet these interconnected healthcare settings generate enormous amounts of heterogeneous data.Traditional Intrusion Detection Systems(IDS),which are generally centralized and machine learning-based,often fail to address the rapidly changing nature of cyberattacks and are challenged by ethical concerns related to patient data privacy.Moreover,traditional AI-driven IDS usually face challenges in handling large-scale,heterogeneous healthcare data while ensuring data privacy and operational efficiency.To address these issues,emerging technologies such as Big Data Analytics(BDA)and Federated Learning(FL)provide a hybrid framework for scalable,adaptive intrusion detection in IoT-driven healthcare systems.Big data techniques enable processing large-scale,highdimensional healthcare data,and FL can be used to train a model in a decentralized manner without transferring raw data,thereby maintaining privacy between institutions.This research proposes a privacy-preserving Federated Learning–based model that efficiently detects cyber threats in connected healthcare systems while ensuring distributed big data processing,privacy,and compliance with ethical regulations.To strengthen the reliability of the reported findings,the resultswere validated using cross-dataset testing and 95%confidence intervals derived frombootstrap analysis,confirming consistent performance across heterogeneous healthcare data distributions.This solution takes a significant step toward securing next-generation healthcare infrastructure by combining scalability,privacy,adaptability,and earlydetection capabilities.The proposed global model achieves a test accuracy of 99.93%±0.03(95%CI)and amiss-rate of only 0.07%±0.02,representing state-of-the-art performance in privacy-preserving intrusion detection.The proposed FL-driven IDS framework offers an efficient,privacy-preserving,and scalable solution for securing next-generation healthcare infrastructures by combining adaptability,early detection,and ethical data management.
文摘The development of remote sensing has seen the creation of a global measurement infrastructure of sustainable development due to growing multipolar archives,rising revisit frequency,and the availability of cloud-accessible platforms of Earth observation.This review summarizes how remote sensing big data is being organized into decision-grade sustainability intelligence,the new approaches to analytics,and how Sustainable Development Goals(SDGs)-oriented application pathways inter-relate action pathways that bridge observations with action.The terminologies like new data ecosystem,data readiness and interoperability,changing economics of scalable computation,and detailing the functions of diversity of modalities(optical,Synthetic Aperture Radar—SAR,thermal,Light Detection and Ranging—LiDAR,hyperspectral)have been defined.These themes of analytics,which are transforming the practice of operational analytics,are then condensed:foundations and self-supervised learning of transferable representations,multi-modal fusion to gap fill and richer inference,spatiotemporal intelligence to trend of early warning,physics-aware hybrid methods to enhance robustness and meaning under non-stationary conditions.Across the climate risk,food systems,water resources,sustainable cities,ecosystems and biodiversity,energy transitions,and health exposure pathways,the roles of Earth Observation(EO)products as direct measures and proxies,and concepts of validating,semantic comparability,and communicating uncertainties play a key role in EO products becoming credible when faced with high-stakes deployment decisions.Lastly,we chart world ways of implementation via monitoring services,early warning systems,and systems of multiple regimes,and previously underline cross-cutting priorities,scalable structures in validation,performance,so that domains of shift,agreeable governance,and Dual-use risk safeguards,and sustainable lifecycle support of EO services.These priorities form a realistic set of priorities on the alignment of remote sensing innovation with quantifiable SDGs progress.
基金National Natural Science Foundation of China Youth Science Foundation ProjectNo.41701170+1 种基金National Natural Science Foundation of China,No.41661025,No.42071216Fundamental Research Funds for the Central Universities,No.18LZUJBWZY068。
文摘In 2007,China surpassed the USA to become the largest carbon emitter in the world.China has promised a 60%–65%reduction in carbon emissions per unit GDP by 2030,compared to the baseline of 2005.Therefore,it is important to obtain accurate dynamic information on the spatial and temporal patterns of carbon emissions and carbon footprints to support formulating effective national carbon emission reduction policies.This study attempts to build a carbon emission panel data model that simulates carbon emissions in China from 2000–2013 using nighttime lighting data and carbon emission statistics data.By applying the Exploratory Spatial-Temporal Data Analysis(ESTDA)framework,this study conducted an analysis on the spatial patterns and dynamic spatial-temporal interactions of carbon footprints from 2001–2013.The improved Tapio decoupling model was adopted to investigate the levels of coupling or decoupling between the carbon emission load and economic growth in 336 prefecture-level units.The results show that,firstly,high accuracy was achieved by the model in simulating carbon emissions.Secondly,the total carbon footprints and carbon deficits across China increased with average annual growth rates of 4.82%and 5.72%,respectively.The overall carbon footprints and carbon deficits were larger in the North than that in the South.There were extremely significant spatial autocorrelation features in the carbon footprints of prefecture-level units.Thirdly,the relative lengths of the Local Indicators of Spatial Association(LISA)time paths were longer in the North than that in the South,and they increased from the coastal to the central and western regions.Lastly,the overall decoupling index was mainly a weak decoupling type,but the number of cities with this weak decoupling continued to decrease.The unsustainable development trend of China’s economic growth and carbon emission load will continue for some time.
基金partially supported by the Construction of Collaborative Innovation Center of Beijing Academy of Agricultural and Forestry Sciences(KJCX20240406)the Beijing Natural Science Foundation(JQ24037)+1 种基金the National Natural Science Foundation of China(32330075)the Earmarked Fund for China Agriculture Research System(CARS-02 and CARS-54)。
文摘The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This era integrates biotechnology,artificial intelligence(AI),and big data information technology.In contrast,China is still in a transition period between stages 2.0 and 3.0,which primarily relies on conventional selection and molecular breeding.In the context of increasingly complex international situations,accurately identifying core issues in China's seed industry innovation and seizing the frontier of international seed technology are strategically important.These efforts are essential for ensuring food security and revitalizing the seed industry.This paper systematically analyzes the characteristics of crop breeding data from artificial selection to intelligent design breeding.It explores the applications and development trends of AI and big data in modern crop breeding from several key perspectives.These include highthroughput phenotype acquisition and analysis,multiomics big data database and management system construction,AI-based multiomics integrated analysis,and the development of intelligent breeding software tools based on biological big data and AI technology.Based on an in-depth analysis of the current status and challenges of China's seed industry technology development,we propose strategic goals and key tasks for China's new generation of AI and big data-driven intelligent design breeding.These suggestions aim to accelerate the development of an intelligent-driven crop breeding engineering system that features large-scale gene mining,efficient gene manipulation,engineered variety design,and systematized biobreeding.This study provides a theoretical basis and practical guidance for the development of China's seed industry technology.
基金sponsored by the U.S.Department of Housing and Urban Development(Grant No.NJLTS0027-22)The opinions expressed in this study are the authors alone,and do not represent the U.S.Depart-ment of HUD’s opinions.
文摘This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models offer insights, they fall short in presenting a holistic view of complex urban challenges. System dynamics (SD) models that are often utilized to provide holistic, systematic understanding of a research subject, like the urban system, emerge as valuable tools, but data scarcity and theoretical inadequacy pose challenges. The research reviews relevant papers on recent SD model applications in urban sustainability since 2018, categorizing them based on nine key indicators. Among the reviewed papers, data limitations and model assumptions were identified as ma jor challenges in applying SD models to urban sustainability. This led to exploring the transformative potential of big data analytics, a rare approach in this field as identified by this study, to enhance SD models’ empirical foundation. Integrating big data could provide data-driven calibration, potentially improving predictive accuracy and reducing reliance on simplified assumptions. The paper concludes by advocating for new approaches that reduce assumptions and promote real-time applicable models, contributing to a comprehensive understanding of urban sustainability through the synergy of big data and SD models.
基金supported by the National Natural Science Foundation of China(32370703)the CAMS Innovation Fund for Medical Sciences(CIFMS)(2022-I2M-1-021,2021-I2M-1-061)the Major Project of Guangzhou National Labora-tory(GZNL2024A01015).
文摘Viral infectious diseases,characterized by their intricate nature and wide-ranging diversity,pose substantial challenges in the domain of data management.The vast volume of data generated by these diseases,spanning from the molecular mechanisms within cells to large-scale epidemiological patterns,has surpassed the capabilities of traditional analytical methods.In the era of artificial intelligence(AI)and big data,there is an urgent necessity for the optimization of these analytical methods to more effectively handle and utilize the information.Despite the rapid accumulation of data associated with viral infections,the lack of a comprehensive framework for integrating,selecting,and analyzing these datasets has left numerous researchers uncertain about which data to select,how to access it,and how to utilize it most effectively in their research.This review endeavors to fill these gaps by exploring the multifaceted nature of viral infectious diseases and summarizing relevant data across multiple levels,from the molecular details of pathogens to broad epidemiological trends.The scope extends from the micro-scale to the macro-scale,encompassing pathogens,hosts,and vectors.In addition to data summarization,this review thoroughly investigates various dataset sources.It also traces the historical evolution of data collection in the field of viral infectious diseases,highlighting the progress achieved over time.Simultaneously,it evaluates the current limitations that impede data utilization.Furthermore,we propose strategies to surmount these challenges,focusing on the development and application of advanced computational techniques,AI-driven models,and enhanced data integration practices.By providing a comprehensive synthesis of existing knowledge,this review is designed to guide future research and contribute to more informed approaches in the surveillance,prevention,and control of viral infectious diseases,particularly within the context of the expanding big-data landscape.
文摘On October 18,2017,the 19th National Congress Report called for the implementation of the Healthy China Strategy.The development of biomedical data plays a pivotal role in advancing this strategy.Since the 18th National Congress of the Communist Party of China,China has vigorously promoted the integration and implementation of the Healthy China and Digital China strategies.The National Health Commission has prioritized the development of health and medical big data,issuing policies to promote standardized applica-tions and foster innovation in"Internet+Healthcare."Biomedical data has significantly contributed to preci-sion medicine,personalized health management,drug development,disease diagnosis,public health monitor-ing,and epidemic prediction capabilities.
基金funded by the National Natural Science Foundation of China(No.42220104008)。
文摘Research into metamorphism plays a pivotal role in reconstructing the evolution of continent,particularly through the study of ancient rocks that are highly susceptible to metamorphic alterations due to multiple tectonic activities.In the big data era,the establishment of new data platforms and the application of big data methods have become a focus for metamorphic rocks.Significant progress has been made in creating specialized databases,compiling comprehensive datasets,and utilizing data analytics to address complex scientific questions.However,many existing databases are inadequate in meeting the specific requirements of metamorphic research,resulting from a substantial amount of valuable data remaining uncollected.Therefore,constructing new databases that can cope with the development of the data era is necessary.This article provides an extensive review of existing databases related to metamorphic rocks and discusses data-driven studies in this.Accordingly,several crucial factors that need to be taken into consideration in the establishment of specialized metamorphic databases are identified,aiming to leverage data-driven applications to achieve broader scientific objectives in metamorphic research.
基金2024 Anqing Normal University University-Level Key Project(ZK2024062D)。
文摘This study examines the Big Data Collection and Preprocessing course at Anhui Institute of Information Engineering,implementing a hybrid teaching reform using the Bosi Smart Learning Platform.The proposed hybrid model follows a“three-stage”and“two-subject”framework,incorporating a structured design for teaching content and assessment methods before,during,and after class.Practical results indicate that this approach significantly enhances teaching effectiveness and improves students’learning autonomy.
基金supported By Grant (PLN2022-14) of State Key Laboratory of Oil and Gas Reservoir Geology and Exploitation (Southwest Petroleum University)。
文摘Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of these data has not been well stored,managed and mined.With the development of cloud computing technology,it provides a rare development opportunity for logging big data private cloud.The traditional petrophysical evaluation and interpretation model has encountered great challenges in the face of new evaluation objects.The solution research of logging big data distributed storage,processing and learning functions integrated in logging big data private cloud has not been carried out yet.To establish a distributed logging big-data private cloud platform centered on a unifi ed learning model,which achieves the distributed storage and processing of logging big data and facilitates the learning of novel knowledge patterns via the unifi ed logging learning model integrating physical simulation and data models in a large-scale functional space,thus resolving the geo-engineering evaluation problem of geothermal fi elds.Based on the research idea of“logging big data cloud platform-unifi ed logging learning model-large function space-knowledge learning&discovery-application”,the theoretical foundation of unified learning model,cloud platform architecture,data storage and learning algorithm,arithmetic power allocation and platform monitoring,platform stability,data security,etc.have been carried on analysis.The designed logging big data cloud platform realizes parallel distributed storage and processing of data and learning algorithms.The feasibility of constructing a well logging big data cloud platform based on a unifi ed learning model of physics and data is analyzed in terms of the structure,ecology,management and security of the cloud platform.The case study shows that the logging big data cloud platform has obvious technical advantages over traditional logging evaluation methods in terms of knowledge discovery method,data software and results sharing,accuracy,speed and complexity.
基金supported in part by the National Natural Science Foundation of China(92259202)the Guangzhou Municipal Health Commission(2023P-GX02)+2 种基金the Science and Technology Projects in Guangzhou(2024B01J1301)the Fundamental Research Funds for the Central Universities,Sun Yat-sen University Clinical Research 5010 Program(2024001)the Cancer Innovative Research Program of Sun Yat-sen University Cancer Center(CIRP-SYSUCC-0010).
文摘1.Introduction Nasopharyngeal carcinoma(NPC)has an imbalanced geographical and ethnic distribution,with notably high incidence rates in Southeastern Asia and China.China accounted for 42.4%of the newly diagnosed cases worldwide in 2022.1NPC is sensitive to irradiation,and radiotherapy is the mainstay curative treatment modality.2The widespread use of intensity-modulated radiation therapy(IMRT)and image-guided radiotherapy(IGRT)has achieved great advances in survival outcomes and toxicity profiles among NPC patients.1In radiotherapy of NPC,the tumor’s proximity to critical structures demands accuracy in tumor delineation in order to avoid radiation-induced toxicities.However,tumor target delineation for radiotherapy of NPC is labor-intensive and radiation oncologists’proficiency varied considerably.3In recent years,the advent of big data analytics and artificial intelligence(AI)has opened up new avenues for improving the precision and efficacy of radiotherapy and individualized treatment in NPC management.3-6In this article,we explored how big data,AI-assisted delineation,radiotherapy planning,and adaptive radiotherapy(ART)are transforming clinical decision-making in NPC treatment.We also provided an outlook on the historical development of AI and big data,their current dominance in oncological radiotherapy,and their projected impact on future clinical practice(Figure 1).
文摘With the rapid development of intelligent manufacturing,industrial big data play an increasingly crucial role in the digital transformation of enterprises.However,current industrial big data platforms still face challenges in data acquisition,processing,and visualization,including data processing inefficiencies,suboptimal storage solutions,and insufficient visualization experiences,which are often exacerbated by inherent data quality issues such as noise and outliers.To address these problems,this study proposes an industrial big data processing framework based on Flink and builds a data presentation system by combining Grafana and ECharts.The system collects data through enterprise sensors,utilizes Kafka message queues for data buffering,and uses Flink for efficient real-time data processing,incorporating foundational data cleansing techniques and strategies for mitigating common noise and anomalies.For data storage,MySQL is employed for static data,and InfluxDB is used for real-time data to improve storage efficiency.In terms of data visualization,Grafana displays real-time data,whereas ECharts is used for static data,offering users an intuitive and comprehensive data display interface.This study aims to provide an efficient and customizable industrial big data solution,with an emphasis on improving data reliability for visualization,to help enterprises monitor equipment information in real time,obtain effective information,and accelerate their intelligent transformation process.
基金Exploration and Practice of the Application of Blockchain Technology to the Cultivation of Compound Talents under the Background of Free Trade Port(HKJG2023-18)。
文摘In the era of big data,the financial industry is undergoing profound changes.By integrating multiple data sources such as transaction records,customer interactions,market trends,and regulatory requirements,big data technology has significantly improved the decision-making efficiency,customer insight,and risk management capabilities of financial institutions.The financial industry has become a pioneer in the application of big data technology,which is widely used in scenarios such as fraud detection,risk management,customer service optimization,and smart transactions.However,financial data security management also faces many challenges,including data breaches,privacy protection,compliance requirements,the complexity of emerging technologies,and the balance between data access and security.This article explores the major challenges of financial data security management,coping strategies,and the evolution of the regulatory environment,and it looks ahead to future trends,highlighting the important role of artificial intelligence and machine learning in financial data security.
基金supported in part by the National Key Research and Development Plan Project(2022YFB3304700)in part by the Xinliao Talent Program of Liaoning Province(XLYC2202002).
文摘As industrial production progresses toward digitalization,massive amounts of data have been collected,transmitted,and stored,with characteristics of large-scale,high-dimensional,heterogeneous,and spatiotemporal dynamics.The high complexity of industrial big data poses challenges for the practical decision-making of domain experts,leading to ever-increasing needs for integrating computational intelligence with human perception into traditional data analysis.Industrial big data visualization integrates theoretical methods and practical technologies from multiple disciplines,including data mining,information visualization,computer graphics,and human-computer interaction,providing a highly effective manner for understanding and exploring the complex industrial processes.This review summarizes the state-of-the-art approaches,characterizes them with six visualization methods,and categorizes them based on analytical tasks and applications.Furthermore,key research challenges and potential future directions are identified.
基金Supported by Research and Application of Soil Collection Software and Soil Ecological Big Data Platform in Guangxi Woodland(GUILINKEYAN[2022ZC]44)Construction of Soil Information Database and Visualization System for Artificial Forests in Central Guangxi(2023GXZCLK62).
文摘In this paper,the application of agricultural big data in agricultural economic management is deeply explored,and its potential in promoting profit growth and innovation is analyzed.However,challenges persist in data collection and integration,limitations of analytical technologies,talent development,team building,and policy support when applying agricultural big data.Effective application strategies are proposed,including data-driven precision agriculture practices,construction of data integration and management platforms,data security and privacy protection strategies,as well as long-term planning and development strategies for agricultural big data,to maximize its impact on agricultural economic management.Future advancements require collaborative efforts in technological innovation,talent cultivation,and policy support,to realize the extensive application of agricultural big data in agricultural economic management and ensure sustainable industrial development.