Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from sei...Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from seismic networks,satellite observations,and geospatial repositories,creating the need for scalable infrastructures capable of integrating and analyzing such data to support intelligent decision-making.Data warehousing technologies provide a robust foundation for this purpose;however,existing earthquake-oriented data warehouses remain limited,often relying on simplified schemas,domain-specific analytics,or cataloguing efforts.This paper presents the design and implementation of a spatio-temporal data warehouse for seismic activity.The framework integrates spatial and temporal dimensions in a unified schema and introduces a novel array-based approach for managing many-to-many relationships between facts and dimensions without intermediate bridge tables.A comparative evaluation against a conventional bridge-table schema demonstrates that the array-based design improves fact-centric query performance,while the bridge-table schema remains advantageous for dimension-centric queries.To reconcile these trade-offs,a hybrid schema is proposed that retains both representations,ensuring balanced efficiency across heterogeneous workloads.The proposed framework demonstrates how spatio-temporal data warehousing can address schema complexity,improve query performance,and support multidimensional visualization.In doing so,it provides a foundation for integrating seismic analysis into broader big data-driven intelligent decision systems for disaster resilience,risk mitigation,and emergency management.展开更多
The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threa...The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threats is both necessary and complex,yet these interconnected healthcare settings generate enormous amounts of heterogeneous data.Traditional Intrusion Detection Systems(IDS),which are generally centralized and machine learning-based,often fail to address the rapidly changing nature of cyberattacks and are challenged by ethical concerns related to patient data privacy.Moreover,traditional AI-driven IDS usually face challenges in handling large-scale,heterogeneous healthcare data while ensuring data privacy and operational efficiency.To address these issues,emerging technologies such as Big Data Analytics(BDA)and Federated Learning(FL)provide a hybrid framework for scalable,adaptive intrusion detection in IoT-driven healthcare systems.Big data techniques enable processing large-scale,highdimensional healthcare data,and FL can be used to train a model in a decentralized manner without transferring raw data,thereby maintaining privacy between institutions.This research proposes a privacy-preserving Federated Learning–based model that efficiently detects cyber threats in connected healthcare systems while ensuring distributed big data processing,privacy,and compliance with ethical regulations.To strengthen the reliability of the reported findings,the resultswere validated using cross-dataset testing and 95%confidence intervals derived frombootstrap analysis,confirming consistent performance across heterogeneous healthcare data distributions.This solution takes a significant step toward securing next-generation healthcare infrastructure by combining scalability,privacy,adaptability,and earlydetection capabilities.The proposed global model achieves a test accuracy of 99.93%±0.03(95%CI)and amiss-rate of only 0.07%±0.02,representing state-of-the-art performance in privacy-preserving intrusion detection.The proposed FL-driven IDS framework offers an efficient,privacy-preserving,and scalable solution for securing next-generation healthcare infrastructures by combining adaptability,early detection,and ethical data management.展开更多
The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This...The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This era integrates biotechnology,artificial intelligence(AI),and big data information technology.In contrast,China is still in a transition period between stages 2.0 and 3.0,which primarily relies on conventional selection and molecular breeding.In the context of increasingly complex international situations,accurately identifying core issues in China's seed industry innovation and seizing the frontier of international seed technology are strategically important.These efforts are essential for ensuring food security and revitalizing the seed industry.This paper systematically analyzes the characteristics of crop breeding data from artificial selection to intelligent design breeding.It explores the applications and development trends of AI and big data in modern crop breeding from several key perspectives.These include highthroughput phenotype acquisition and analysis,multiomics big data database and management system construction,AI-based multiomics integrated analysis,and the development of intelligent breeding software tools based on biological big data and AI technology.Based on an in-depth analysis of the current status and challenges of China's seed industry technology development,we propose strategic goals and key tasks for China's new generation of AI and big data-driven intelligent design breeding.These suggestions aim to accelerate the development of an intelligent-driven crop breeding engineering system that features large-scale gene mining,efficient gene manipulation,engineered variety design,and systematized biobreeding.This study provides a theoretical basis and practical guidance for the development of China's seed industry technology.展开更多
With the rapid development of the global economy, maritime transportation has become much more convenient due to large capacities and low freight. However, this means the sea lanes are becoming more and more crowded,l...With the rapid development of the global economy, maritime transportation has become much more convenient due to large capacities and low freight. However, this means the sea lanes are becoming more and more crowded,leading to high probabilities of marine accidents in complex maritime environments. According to relevant historical statistics, a large number of accidents have happened in water areas that lack high precision navigation data, which can be utilized to enhance navigation safety. The purpose of this work was to carry out ship route planning automatically, by mining historical big automatic identification system(AIS) data. It is well-known that experiential navigation information hidden in maritime big data could be automatically extracted using advanced data mining techniques;assisting in the generation of safe and reliable ship planning routes for complex maritime environments. In this paper, a novel method is proposed to construct a big data-driven framework for generating ship planning routes automatically, under varying navigation conditions. The method performs density-based spatial clustering of applications with noise first on a large number of ship trajectories to form different trajectory vector clusters. Then, it iteratively calculates its centerline in the trajectory vector cluster, and constructs the waterway network from the node-arc topology relationship among these centerlines. The generation of shipping route could be based on the waterway network and conducted by rasterizing the marine environment risks for the sea area not covered by the waterway network. Numerous experiments have been conducted on different AIS data sets in different water areas, and the experimental results have demonstrated the effectiveness of the framework of the ship route planning proposed in this paper.展开更多
The Wufeng–Longmaxi Formation derives its name from the Upper Ordovician Wufeng Formation and the Lower Silurian Longmaxi Formation,found in sequence in the Sichuan Basin.This formation hosts rich shale gas reservoir...The Wufeng–Longmaxi Formation derives its name from the Upper Ordovician Wufeng Formation and the Lower Silurian Longmaxi Formation,found in sequence in the Sichuan Basin.This formation hosts rich shale gas reservoirs,and its shale gas enrichment patterns are examined in this study using data from 1197 shale samples collected from 14 wells.Five basic and three key parameters,eight in all,are assessed for each sample.The five basic parameters include burial depth and the contents of four mineral types—quartz,clay,carbonate,and other minerals;the three key parameters,representing shale gas enrichment,are total organic carbon(TOC)content,porosity,and gas content.The SHapley Additive exPlanations(SHAP)analysis originated in game theory is used here in an interpretable machine learning framework,to address issues of heterogeneous data structure,noisy relationships,and multi-objective optimization.An evaluation of the ranking,contribution values,and conditions of changes for these parameters offers new quantitative insights into shale gas enrichment patterns.A quantitative analysis of the relationship between data-sets identifies the primary factors controlling TOC,porosity,and gas content of shale gas reservoirs.The results show that TOC and porosity jointly influence gas content;mineral content has a significant impact on both,TOC and porosity;and the burial depth governs porosity which,in turn,affects the conditions under which shale gas is preserved.Input parameter thresholds are also determined and provide a basis for the establishment of quantitative criteria to evaluate shale gas enrichment.The predictive accuracy of the model used in this study is significantly improved by the step-wise addition of two input parameters,namely TOC and porosity,separately and together.Thus,the game theory method in big data-driven analysis uses a combination of TOC and porosity to evaluate the gas content with encouraging results—suggesting that these are the key parameters that indicate source rock and reservoir properties.展开更多
Artificial intelligence(AI)has reshaped the subject of product innovation and triggered transformations in product innovation strategies and processes.This study proposes a subject-strategy-process(SSP)framework for b...Artificial intelligence(AI)has reshaped the subject of product innovation and triggered transformations in product innovation strategies and processes.This study proposes a subject-strategy-process(SSP)framework for business intelligence(BI)for big data-driven product innovation through logical deduction,drawing on the theory of big data cooperative assets and an adaptive innovation perspective on enterprise-user interaction.The aim is to explore new mechanisms through which AI influences product innovation in manufacturing.This study indicates three aspects.Firstly,the two-way involvement of humans and AI forms a dual feedback-enhancement mechanism of factor combination and knowledge accumulation.This mechanism drives structural changes in innovation subjects and forms a new foundation for strategic and process transformations in product innovation.Secondly,the alignment between an enterprise’s cognitive strategy about AI,competitive strategy,organizational culture,business model,and ecosystem jointly shapes the integrated application of AI in innovation processes.Thirdly,the new features of the big data-driven product innovation process include full-process diffusion from the fuzzy front end,nonlinear iteration of demand-solution pairs,and generative self-testing in intelligent manufacturing.Taken together,the study demonstrates that the SSP framework is well-suited to analyzing the new mechanisms of BI for big data-driven product innovation,which offers a fresh lens for examining the relationship between AI and product innovation.展开更多
This study integrates multiple sources of data(transaction data,policy text,public opinion data)with visualization techniques(such as heat maps,time-series trend charts,3D building brochures)to construct an analysis f...This study integrates multiple sources of data(transaction data,policy text,public opinion data)with visualization techniques(such as heat maps,time-series trend charts,3D building brochures)to construct an analysis framework for the Chengdu real estate market.By using the Adaptive Neuro-Fuzzy Inference System(ANFIS)prediction model,spatial GIS(Geographic Information System analysis)analysis,and interactive dashboards,this study reveals market differentiation,policy impacts,and changes in demand structure,thereby providing decision support for the government,enterprises,and homebuyers.展开更多
To address the issue of instability or even imbalance in the orientation and attitude control of quadrotor unmanned aerial vehicles(QUAVs)under random disturbances,this paper proposes a distributed antidisturbance dat...To address the issue of instability or even imbalance in the orientation and attitude control of quadrotor unmanned aerial vehicles(QUAVs)under random disturbances,this paper proposes a distributed antidisturbance data-driven event-triggered fusion control method,which achieves efficient fault diagnosis while suppressing random disturbances and mitigating communication conflicts within the QUAV swarm.First,the impact of random disturbances on the UAV swarm is analyzed,and a model for orientation and attitude control of QUAVs under stochastic perturbations is established,with the disturbance gain threshold determined.Second,a fault diagnosis system based on a high-gain observer is designed,constructing a fault gain criterion by integrating orientation and attitude information from QUAVs.Subsequently,a model-free dynamic linearization-based data modeling(MFDLDM)framework is developed using model-free adaptive control,which efficiently fits the nonlinear control model of the QUAV swarm while reducing temporal constraints on control data.On this basis,this paper constructs a distributed data-driven event-triggered controller based on the staggered communication mechanism,which consists of an equivalent QUAV controller and an event-triggered controller,and is able to reduce the communication conflicts while suppressing the influence of random interference.Finally,by incorporating random disturbances into the controller,comparative experiments and physical validations are conducted on the QUAV platforms,fully demonstrating the strong adaptability and robustness of the proposed distributed event-triggered fault-tolerant control system.展开更多
Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations a...Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.展开更多
在小学高年级英语语篇教学中,存在学生思维浅表化、问题设计碎片化、旧版教材适配难这三个痛点。以译林版英语教材六年级上册Unit 4 Then and now中Story time的教学为例,教师立足教材文本,构建“课前定问—课初引链—课中解链—课后拓...在小学高年级英语语篇教学中,存在学生思维浅表化、问题设计碎片化、旧版教材适配难这三个痛点。以译林版英语教材六年级上册Unit 4 Then and now中Story time的教学为例,教师立足教材文本,构建“课前定问—课初引链—课中解链—课后拓链—全程评链”的五步闭环,用大问题拉主线、小问题搭台阶,能激活学生语篇学习内驱力,实现英语教学从“知识传递”到“素养培养”的转变。展开更多
With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service...With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service resources,uneven distribution,and prominent supply-demand contradictions have seriously affected service quality.Big data technology,with core advantages including data collection,analysis and mining,and accurate prediction,provides a new solution for the allocation of community elderly care service resources.This paper systematically studies the application value of big data technology in the allocation of community elderly care service resources from three aspects:resource allocation efficiency,service accuracy,and management intelligence.Combined with practical needs,it proposes optimal allocation strategies such as building a big data analysis platform and accurately grasping the elderly’s care needs,striving to provide operable path references for the construction of community elderly care service systems,promoting the early realization of the elderly care service goal of“adequate support and proper care for the elderly”,and boosting the high-quality development of China’s elderly care service industry.展开更多
大孔径CT基于X线断层扫描原理的超大孔径设计,能容纳肥胖患者以及携带医疗设备的特殊患者,其凭借高清晰成像为放疗计划的制定与实施提供了关键的解剖学依据,是现代放疗“精准化”转型的重要支撑[1]。我院于2017年引进飞利浦Brilliance B...大孔径CT基于X线断层扫描原理的超大孔径设计,能容纳肥胖患者以及携带医疗设备的特殊患者,其凭借高清晰成像为放疗计划的制定与实施提供了关键的解剖学依据,是现代放疗“精准化”转型的重要支撑[1]。我院于2017年引进飞利浦Brilliance Big Bore CT设备,截至2024年底,已累计完成5000余例肿瘤患者的放疗定位扫描。展开更多
Carbonated silicate melts transport oxidized carbon from the deep to shallow mantle,and thus play an important role in the deep carbon cycle.However,it is unclear how carbonated silicate melts are formed.Here,we repor...Carbonated silicate melts transport oxidized carbon from the deep to shallow mantle,and thus play an important role in the deep carbon cycle.However,it is unclear how carbonated silicate melts are formed.Here,we report whole-rock major and trace element,and Mg-Zn-Fe isotope,and in situ olivine and clinopyroxene Mg isotope data for peridotites and pyroxenites in Cenozoic alkaline basalts from Shulan and Yitong,Northeast China,to reveal the formation mechanisms of carbonated silicate melts.As residues of the primitive mantle,most of the Shulan and Yitong(spinel)harzburgite and lherzolite have relatively lower δ26Mg and higher δ66Zn than the mantle values,and show LREE-enriched patterns.These features indicate that these peridotites have been modified by carbonated silicate melt,transforming them into carbonated peridotites.The Shulan websterites exhibit metasomatic textures,and most of their whole-rock and in situ δ26Mg values are lower than the mantle values,which also supports the presence of a carbonated lithospheric mantle.The Yitong cumulate orthopyroxenites,representing silica-rich melts,have lower δ26Mg and higher δ66Zn values than mantle peridotites,implying that recycled carbonate has been involved in these silica-rich melts.The Yitong cumulate wehrlites and(olivine)websterites display major element compositions similar to those of carbonated peridotite melts,coupled with low whole-rock and clinopyroxene δ26Mg values,high δ66Zn and δ57Fe values,indicating the occurrence of carbonated silicate melts.The transition from orthopyroxenites to(olivine)websterites is marked by the decreasing of their whole-rock δ26Mg,δ57Fe values and Zn/Fe ratios and increasing of their δ66Zn values,along with decreasing of their crystallization temperatures,suggesting that silica-rich melts were gradually transformed into carbonated silicate melts via continuous interaction with carbonated peridotite.Our findings demonstrate that silica-rich melts from the stagnant slab can extract carbon from pre-existing carbonated mantle within the big mantle wedge.展开更多
This study introduces an innovative“Big Model”strategy to enhance Bridge Structural Health Monitoring(SHM)using a Convolutional Neural Network(CNN),time-frequency analysis,and fine element analysis.Leveraging ensemb...This study introduces an innovative“Big Model”strategy to enhance Bridge Structural Health Monitoring(SHM)using a Convolutional Neural Network(CNN),time-frequency analysis,and fine element analysis.Leveraging ensemble methods,collaborative learning,and distributed computing,the approach effectively manages the complexity and scale of large-scale bridge data.The CNN employs transfer learning,fine-tuning,and continuous monitoring to optimize models for adaptive and accurate structural health assessments,focusing on extracting meaningful features through time-frequency analysis.By integrating Finite Element Analysis,time-frequency analysis,and CNNs,the strategy provides a comprehensive understanding of bridge health.Utilizing diverse sensor data,sophisticated feature extraction,and advanced CNN architecture,the model is optimized through rigorous preprocessing and hyperparameter tuning.This approach significantly enhances the ability to make accurate predictions,monitor structural health,and support proactive maintenance practices,thereby ensuring the safety and longevity of critical infrastructure.展开更多
This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models ...This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models offer insights, they fall short in presenting a holistic view of complex urban challenges. System dynamics (SD) models that are often utilized to provide holistic, systematic understanding of a research subject, like the urban system, emerge as valuable tools, but data scarcity and theoretical inadequacy pose challenges. The research reviews relevant papers on recent SD model applications in urban sustainability since 2018, categorizing them based on nine key indicators. Among the reviewed papers, data limitations and model assumptions were identified as ma jor challenges in applying SD models to urban sustainability. This led to exploring the transformative potential of big data analytics, a rare approach in this field as identified by this study, to enhance SD models’ empirical foundation. Integrating big data could provide data-driven calibration, potentially improving predictive accuracy and reducing reliance on simplified assumptions. The paper concludes by advocating for new approaches that reduce assumptions and promote real-time applicable models, contributing to a comprehensive understanding of urban sustainability through the synergy of big data and SD models.展开更多
Research into metamorphism plays a pivotal role in reconstructing the evolution of continent,particularly through the study of ancient rocks that are highly susceptible to metamorphic alterations due to multiple tecto...Research into metamorphism plays a pivotal role in reconstructing the evolution of continent,particularly through the study of ancient rocks that are highly susceptible to metamorphic alterations due to multiple tectonic activities.In the big data era,the establishment of new data platforms and the application of big data methods have become a focus for metamorphic rocks.Significant progress has been made in creating specialized databases,compiling comprehensive datasets,and utilizing data analytics to address complex scientific questions.However,many existing databases are inadequate in meeting the specific requirements of metamorphic research,resulting from a substantial amount of valuable data remaining uncollected.Therefore,constructing new databases that can cope with the development of the data era is necessary.This article provides an extensive review of existing databases related to metamorphic rocks and discusses data-driven studies in this.Accordingly,several crucial factors that need to be taken into consideration in the establishment of specialized metamorphic databases are identified,aiming to leverage data-driven applications to achieve broader scientific objectives in metamorphic research.展开更多
Integrating machine learning and data mining is crucial for processing big data and extracting valuable insights to enhance decision-making.However,imbalanced target variables within big data present technical challen...Integrating machine learning and data mining is crucial for processing big data and extracting valuable insights to enhance decision-making.However,imbalanced target variables within big data present technical challenges that hinder the performance of supervised learning classifiers on key evaluation metrics,limiting their overall effectiveness.This study presents a comprehensive review of both common and recently developed Supervised Learning Classifiers(SLCs)and evaluates their performance in data-driven decision-making.The evaluation uses various metrics,with a particular focus on the Harmonic Mean Score(F-1 score)on an imbalanced real-world bank target marketing dataset.The findings indicate that grid-search random forest and random-search random forest excel in Precision and area under the curve,while Extreme Gradient Boosting(XGBoost)outperforms other traditional classifiers in terms of F-1 score.Employing oversampling methods to address the imbalanced data shows significant performance improvement in XGBoost,delivering superior results across all metrics,particularly when using the SMOTE variant known as the BorderlineSMOTE2 technique.The study concludes several key factors for effectively addressing the challenges of supervised learning with imbalanced datasets.These factors include the importance of selecting appropriate datasets for training and testing,choosing the right classifiers,employing effective techniques for processing and handling imbalanced datasets,and identifying suitable metrics for performance evaluation.Additionally,factors also entail the utilisation of effective exploratory data analysis in conjunction with visualisation techniques to yield insights conducive to data-driven decision-making.展开更多
Mitigating vortex-induced vibrations(VIV)in flexible risers represents a critical concern in offshore oil and gas production,considering its potential impact on operational safety and efficiency.The accurate predictio...Mitigating vortex-induced vibrations(VIV)in flexible risers represents a critical concern in offshore oil and gas production,considering its potential impact on operational safety and efficiency.The accurate prediction of displacement and position of VIV in flexible risers remains challenging under actual marine conditions.This study presents a data-driven model for riser displacement prediction that corresponds to field conditions.Experimental data analysis reveals that the XGBoost algorithm predicts the maximum displacement and position with superior accuracy compared with Support vector regression(SVR),considering both computational efficiency and precision.Platform displacement in the Y-direction demonstrates a significant positive correlation with both axial depth and maximum displacement magnitude.The fourth point displacement exhibits the highest contribution to model prediction outcomes,showing a positive influence on maximum displacement while negatively affecting the axial depth of maximum displacement.Platform displacement in the X-and Y-directions exhibits competitive effects on both the riser’s maximum displacement and its axial depth.Through the implementation of XGBoost algorithm and SHapley Additive exPlanation(SHAP)analysis,the model effectively estimates the riser’s maximum displacement and its precise location.This data-driven approach achieves predictions using minimal,readily available data points,enhancing its practical field applications and demonstrating clear relevance to academic and professional communities.展开更多
Based on the educational evaluation reform,this study explores the construction of an evidence-based value-added evaluation system based on data-driven,aiming to solve the limitations of traditional evaluation methods...Based on the educational evaluation reform,this study explores the construction of an evidence-based value-added evaluation system based on data-driven,aiming to solve the limitations of traditional evaluation methods.The research adopts the method of combining theoretical analysis and practical application,and designs the evidence-based value-added evaluation framework,which includes the core elements of a multi-source heterogeneous data acquisition and processing system,a value-added evaluation agent based on a large model,and an evaluation implementation and application mechanism.Through empirical research verification,the evaluation system has remarkable effects in improving learning participation,promoting ability development,and supporting teaching decision-making,and provides a theoretical reference and practical path for educational evaluation reform in the new era.The research shows that the evidence-based value-added evaluation system based on data-driven can reflect students’actual progress more fairly and objectively by accurately measuring the difference in starting point and development range of students,and provide strong support for the realization of high-quality education development.展开更多
文摘Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from seismic networks,satellite observations,and geospatial repositories,creating the need for scalable infrastructures capable of integrating and analyzing such data to support intelligent decision-making.Data warehousing technologies provide a robust foundation for this purpose;however,existing earthquake-oriented data warehouses remain limited,often relying on simplified schemas,domain-specific analytics,or cataloguing efforts.This paper presents the design and implementation of a spatio-temporal data warehouse for seismic activity.The framework integrates spatial and temporal dimensions in a unified schema and introduces a novel array-based approach for managing many-to-many relationships between facts and dimensions without intermediate bridge tables.A comparative evaluation against a conventional bridge-table schema demonstrates that the array-based design improves fact-centric query performance,while the bridge-table schema remains advantageous for dimension-centric queries.To reconcile these trade-offs,a hybrid schema is proposed that retains both representations,ensuring balanced efficiency across heterogeneous workloads.The proposed framework demonstrates how spatio-temporal data warehousing can address schema complexity,improve query performance,and support multidimensional visualization.In doing so,it provides a foundation for integrating seismic analysis into broader big data-driven intelligent decision systems for disaster resilience,risk mitigation,and emergency management.
文摘The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threats is both necessary and complex,yet these interconnected healthcare settings generate enormous amounts of heterogeneous data.Traditional Intrusion Detection Systems(IDS),which are generally centralized and machine learning-based,often fail to address the rapidly changing nature of cyberattacks and are challenged by ethical concerns related to patient data privacy.Moreover,traditional AI-driven IDS usually face challenges in handling large-scale,heterogeneous healthcare data while ensuring data privacy and operational efficiency.To address these issues,emerging technologies such as Big Data Analytics(BDA)and Federated Learning(FL)provide a hybrid framework for scalable,adaptive intrusion detection in IoT-driven healthcare systems.Big data techniques enable processing large-scale,highdimensional healthcare data,and FL can be used to train a model in a decentralized manner without transferring raw data,thereby maintaining privacy between institutions.This research proposes a privacy-preserving Federated Learning–based model that efficiently detects cyber threats in connected healthcare systems while ensuring distributed big data processing,privacy,and compliance with ethical regulations.To strengthen the reliability of the reported findings,the resultswere validated using cross-dataset testing and 95%confidence intervals derived frombootstrap analysis,confirming consistent performance across heterogeneous healthcare data distributions.This solution takes a significant step toward securing next-generation healthcare infrastructure by combining scalability,privacy,adaptability,and earlydetection capabilities.The proposed global model achieves a test accuracy of 99.93%±0.03(95%CI)and amiss-rate of only 0.07%±0.02,representing state-of-the-art performance in privacy-preserving intrusion detection.The proposed FL-driven IDS framework offers an efficient,privacy-preserving,and scalable solution for securing next-generation healthcare infrastructures by combining adaptability,early detection,and ethical data management.
基金partially supported by the Construction of Collaborative Innovation Center of Beijing Academy of Agricultural and Forestry Sciences(KJCX20240406)the Beijing Natural Science Foundation(JQ24037)+1 种基金the National Natural Science Foundation of China(32330075)the Earmarked Fund for China Agriculture Research System(CARS-02 and CARS-54)。
文摘The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This era integrates biotechnology,artificial intelligence(AI),and big data information technology.In contrast,China is still in a transition period between stages 2.0 and 3.0,which primarily relies on conventional selection and molecular breeding.In the context of increasingly complex international situations,accurately identifying core issues in China's seed industry innovation and seizing the frontier of international seed technology are strategically important.These efforts are essential for ensuring food security and revitalizing the seed industry.This paper systematically analyzes the characteristics of crop breeding data from artificial selection to intelligent design breeding.It explores the applications and development trends of AI and big data in modern crop breeding from several key perspectives.These include highthroughput phenotype acquisition and analysis,multiomics big data database and management system construction,AI-based multiomics integrated analysis,and the development of intelligent breeding software tools based on biological big data and AI technology.Based on an in-depth analysis of the current status and challenges of China's seed industry technology development,we propose strategic goals and key tasks for China's new generation of AI and big data-driven intelligent design breeding.These suggestions aim to accelerate the development of an intelligent-driven crop breeding engineering system that features large-scale gene mining,efficient gene manipulation,engineered variety design,and systematized biobreeding.This study provides a theoretical basis and practical guidance for the development of China's seed industry technology.
文摘With the rapid development of the global economy, maritime transportation has become much more convenient due to large capacities and low freight. However, this means the sea lanes are becoming more and more crowded,leading to high probabilities of marine accidents in complex maritime environments. According to relevant historical statistics, a large number of accidents have happened in water areas that lack high precision navigation data, which can be utilized to enhance navigation safety. The purpose of this work was to carry out ship route planning automatically, by mining historical big automatic identification system(AIS) data. It is well-known that experiential navigation information hidden in maritime big data could be automatically extracted using advanced data mining techniques;assisting in the generation of safe and reliable ship planning routes for complex maritime environments. In this paper, a novel method is proposed to construct a big data-driven framework for generating ship planning routes automatically, under varying navigation conditions. The method performs density-based spatial clustering of applications with noise first on a large number of ship trajectories to form different trajectory vector clusters. Then, it iteratively calculates its centerline in the trajectory vector cluster, and constructs the waterway network from the node-arc topology relationship among these centerlines. The generation of shipping route could be based on the waterway network and conducted by rasterizing the marine environment risks for the sea area not covered by the waterway network. Numerous experiments have been conducted on different AIS data sets in different water areas, and the experimental results have demonstrated the effectiveness of the framework of the ship route planning proposed in this paper.
基金funded by the Technical Development(Entrusted)Project of Science and Department of SINOPEC(Grant No.P23240-4)the National Natural Science Foundation of China(Grant Nos.42172165,42272143 and 2025ZD1403901-05)。
文摘The Wufeng–Longmaxi Formation derives its name from the Upper Ordovician Wufeng Formation and the Lower Silurian Longmaxi Formation,found in sequence in the Sichuan Basin.This formation hosts rich shale gas reservoirs,and its shale gas enrichment patterns are examined in this study using data from 1197 shale samples collected from 14 wells.Five basic and three key parameters,eight in all,are assessed for each sample.The five basic parameters include burial depth and the contents of four mineral types—quartz,clay,carbonate,and other minerals;the three key parameters,representing shale gas enrichment,are total organic carbon(TOC)content,porosity,and gas content.The SHapley Additive exPlanations(SHAP)analysis originated in game theory is used here in an interpretable machine learning framework,to address issues of heterogeneous data structure,noisy relationships,and multi-objective optimization.An evaluation of the ranking,contribution values,and conditions of changes for these parameters offers new quantitative insights into shale gas enrichment patterns.A quantitative analysis of the relationship between data-sets identifies the primary factors controlling TOC,porosity,and gas content of shale gas reservoirs.The results show that TOC and porosity jointly influence gas content;mineral content has a significant impact on both,TOC and porosity;and the burial depth governs porosity which,in turn,affects the conditions under which shale gas is preserved.Input parameter thresholds are also determined and provide a basis for the establishment of quantitative criteria to evaluate shale gas enrichment.The predictive accuracy of the model used in this study is significantly improved by the step-wise addition of two input parameters,namely TOC and porosity,separately and together.Thus,the game theory method in big data-driven analysis uses a combination of TOC and porosity to evaluate the gas content with encouraging results—suggesting that these are the key parameters that indicate source rock and reservoir properties.
基金supported by the key project of the National Natural Science Foundation of China“Research on the Theory,Methods,and Applications of Innovation via Enterprise-User Interaction Driven by Big Data in the Internet Environment”(No.71832014)the key project of the National Natural Science Foundation of China“Research on the Digital Transformation and Adaptive Management Changes of Manufacturing Enterprises”(No.72032009)the major project of the National Social Science Fund of China“Research on the Impact of Artificial Intelligence on the Transformation and Upgrading of the Manufacturing Industry and Its Governance System”(No.23&DA091).
文摘Artificial intelligence(AI)has reshaped the subject of product innovation and triggered transformations in product innovation strategies and processes.This study proposes a subject-strategy-process(SSP)framework for business intelligence(BI)for big data-driven product innovation through logical deduction,drawing on the theory of big data cooperative assets and an adaptive innovation perspective on enterprise-user interaction.The aim is to explore new mechanisms through which AI influences product innovation in manufacturing.This study indicates three aspects.Firstly,the two-way involvement of humans and AI forms a dual feedback-enhancement mechanism of factor combination and knowledge accumulation.This mechanism drives structural changes in innovation subjects and forms a new foundation for strategic and process transformations in product innovation.Secondly,the alignment between an enterprise’s cognitive strategy about AI,competitive strategy,organizational culture,business model,and ecosystem jointly shapes the integrated application of AI in innovation processes.Thirdly,the new features of the big data-driven product innovation process include full-process diffusion from the fuzzy front end,nonlinear iteration of demand-solution pairs,and generative self-testing in intelligent manufacturing.Taken together,the study demonstrates that the SSP framework is well-suited to analyzing the new mechanisms of BI for big data-driven product innovation,which offers a fresh lens for examining the relationship between AI and product innovation.
基金Chengdu City Philosophy and Social Sciences Research Center“artificial intelligence+urban communication”theory and Application Research Center Project“Chengdu real estate vertical market public opinion data visualization research”(Project No.RZCC2025017).
文摘This study integrates multiple sources of data(transaction data,policy text,public opinion data)with visualization techniques(such as heat maps,time-series trend charts,3D building brochures)to construct an analysis framework for the Chengdu real estate market.By using the Adaptive Neuro-Fuzzy Inference System(ANFIS)prediction model,spatial GIS(Geographic Information System analysis)analysis,and interactive dashboards,this study reveals market differentiation,policy impacts,and changes in demand structure,thereby providing decision support for the government,enterprises,and homebuyers.
基金supported in part by the National Natural Science Foundation of China,Grant/Award Number:62003267the Key Research and Development Program of Shaanxi Province,Grant/Award Number:2023-GHZD-33Open Project of the State Key Laboratory of Intelligent Game,Grant/Award Number:ZBKF-23-05。
文摘To address the issue of instability or even imbalance in the orientation and attitude control of quadrotor unmanned aerial vehicles(QUAVs)under random disturbances,this paper proposes a distributed antidisturbance data-driven event-triggered fusion control method,which achieves efficient fault diagnosis while suppressing random disturbances and mitigating communication conflicts within the QUAV swarm.First,the impact of random disturbances on the UAV swarm is analyzed,and a model for orientation and attitude control of QUAVs under stochastic perturbations is established,with the disturbance gain threshold determined.Second,a fault diagnosis system based on a high-gain observer is designed,constructing a fault gain criterion by integrating orientation and attitude information from QUAVs.Subsequently,a model-free dynamic linearization-based data modeling(MFDLDM)framework is developed using model-free adaptive control,which efficiently fits the nonlinear control model of the QUAV swarm while reducing temporal constraints on control data.On this basis,this paper constructs a distributed data-driven event-triggered controller based on the staggered communication mechanism,which consists of an equivalent QUAV controller and an event-triggered controller,and is able to reduce the communication conflicts while suppressing the influence of random interference.Finally,by incorporating random disturbances into the controller,comparative experiments and physical validations are conducted on the QUAV platforms,fully demonstrating the strong adaptability and robustness of the proposed distributed event-triggered fault-tolerant control system.
基金supported by the National Key R&D Program of China(Grant No.2023YFC3209504)Natural Science Foundation of Wuhan(Grant No.2024040801020271)the Fundamental Research Funds for Central Public Welfare Research Institutes(Grant No.CKSF2025718/YT).
文摘Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.
文摘在小学高年级英语语篇教学中,存在学生思维浅表化、问题设计碎片化、旧版教材适配难这三个痛点。以译林版英语教材六年级上册Unit 4 Then and now中Story time的教学为例,教师立足教材文本,构建“课前定问—课初引链—课中解链—课后拓链—全程评链”的五步闭环,用大问题拉主线、小问题搭台阶,能激活学生语篇学习内驱力,实现英语教学从“知识传递”到“素养培养”的转变。
文摘With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service resources,uneven distribution,and prominent supply-demand contradictions have seriously affected service quality.Big data technology,with core advantages including data collection,analysis and mining,and accurate prediction,provides a new solution for the allocation of community elderly care service resources.This paper systematically studies the application value of big data technology in the allocation of community elderly care service resources from three aspects:resource allocation efficiency,service accuracy,and management intelligence.Combined with practical needs,it proposes optimal allocation strategies such as building a big data analysis platform and accurately grasping the elderly’s care needs,striving to provide operable path references for the construction of community elderly care service systems,promoting the early realization of the elderly care service goal of“adequate support and proper care for the elderly”,and boosting the high-quality development of China’s elderly care service industry.
文摘大孔径CT基于X线断层扫描原理的超大孔径设计,能容纳肥胖患者以及携带医疗设备的特殊患者,其凭借高清晰成像为放疗计划的制定与实施提供了关键的解剖学依据,是现代放疗“精准化”转型的重要支撑[1]。我院于2017年引进飞利浦Brilliance Big Bore CT设备,截至2024年底,已累计完成5000余例肿瘤患者的放疗定位扫描。
基金financially supported by the National Natural Science Foundation of China(Grants Nos.42572064 and 42422203)the Program for Jilin University Science and Technology Innovative Research Team(Grant No.2021-TD-05)+1 种基金the Fundamental Research Funds for the Central Universities(Grant No.45124031D045)the Graduate Innovation Fund of Jilin University.
文摘Carbonated silicate melts transport oxidized carbon from the deep to shallow mantle,and thus play an important role in the deep carbon cycle.However,it is unclear how carbonated silicate melts are formed.Here,we report whole-rock major and trace element,and Mg-Zn-Fe isotope,and in situ olivine and clinopyroxene Mg isotope data for peridotites and pyroxenites in Cenozoic alkaline basalts from Shulan and Yitong,Northeast China,to reveal the formation mechanisms of carbonated silicate melts.As residues of the primitive mantle,most of the Shulan and Yitong(spinel)harzburgite and lherzolite have relatively lower δ26Mg and higher δ66Zn than the mantle values,and show LREE-enriched patterns.These features indicate that these peridotites have been modified by carbonated silicate melt,transforming them into carbonated peridotites.The Shulan websterites exhibit metasomatic textures,and most of their whole-rock and in situ δ26Mg values are lower than the mantle values,which also supports the presence of a carbonated lithospheric mantle.The Yitong cumulate orthopyroxenites,representing silica-rich melts,have lower δ26Mg and higher δ66Zn values than mantle peridotites,implying that recycled carbonate has been involved in these silica-rich melts.The Yitong cumulate wehrlites and(olivine)websterites display major element compositions similar to those of carbonated peridotite melts,coupled with low whole-rock and clinopyroxene δ26Mg values,high δ66Zn and δ57Fe values,indicating the occurrence of carbonated silicate melts.The transition from orthopyroxenites to(olivine)websterites is marked by the decreasing of their whole-rock δ26Mg,δ57Fe values and Zn/Fe ratios and increasing of their δ66Zn values,along with decreasing of their crystallization temperatures,suggesting that silica-rich melts were gradually transformed into carbonated silicate melts via continuous interaction with carbonated peridotite.Our findings demonstrate that silica-rich melts from the stagnant slab can extract carbon from pre-existing carbonated mantle within the big mantle wedge.
文摘This study introduces an innovative“Big Model”strategy to enhance Bridge Structural Health Monitoring(SHM)using a Convolutional Neural Network(CNN),time-frequency analysis,and fine element analysis.Leveraging ensemble methods,collaborative learning,and distributed computing,the approach effectively manages the complexity and scale of large-scale bridge data.The CNN employs transfer learning,fine-tuning,and continuous monitoring to optimize models for adaptive and accurate structural health assessments,focusing on extracting meaningful features through time-frequency analysis.By integrating Finite Element Analysis,time-frequency analysis,and CNNs,the strategy provides a comprehensive understanding of bridge health.Utilizing diverse sensor data,sophisticated feature extraction,and advanced CNN architecture,the model is optimized through rigorous preprocessing and hyperparameter tuning.This approach significantly enhances the ability to make accurate predictions,monitor structural health,and support proactive maintenance practices,thereby ensuring the safety and longevity of critical infrastructure.
基金sponsored by the U.S.Department of Housing and Urban Development(Grant No.NJLTS0027-22)The opinions expressed in this study are the authors alone,and do not represent the U.S.Depart-ment of HUD’s opinions.
文摘This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models offer insights, they fall short in presenting a holistic view of complex urban challenges. System dynamics (SD) models that are often utilized to provide holistic, systematic understanding of a research subject, like the urban system, emerge as valuable tools, but data scarcity and theoretical inadequacy pose challenges. The research reviews relevant papers on recent SD model applications in urban sustainability since 2018, categorizing them based on nine key indicators. Among the reviewed papers, data limitations and model assumptions were identified as ma jor challenges in applying SD models to urban sustainability. This led to exploring the transformative potential of big data analytics, a rare approach in this field as identified by this study, to enhance SD models’ empirical foundation. Integrating big data could provide data-driven calibration, potentially improving predictive accuracy and reducing reliance on simplified assumptions. The paper concludes by advocating for new approaches that reduce assumptions and promote real-time applicable models, contributing to a comprehensive understanding of urban sustainability through the synergy of big data and SD models.
基金funded by the National Natural Science Foundation of China(No.42220104008)。
文摘Research into metamorphism plays a pivotal role in reconstructing the evolution of continent,particularly through the study of ancient rocks that are highly susceptible to metamorphic alterations due to multiple tectonic activities.In the big data era,the establishment of new data platforms and the application of big data methods have become a focus for metamorphic rocks.Significant progress has been made in creating specialized databases,compiling comprehensive datasets,and utilizing data analytics to address complex scientific questions.However,many existing databases are inadequate in meeting the specific requirements of metamorphic research,resulting from a substantial amount of valuable data remaining uncollected.Therefore,constructing new databases that can cope with the development of the data era is necessary.This article provides an extensive review of existing databases related to metamorphic rocks and discusses data-driven studies in this.Accordingly,several crucial factors that need to be taken into consideration in the establishment of specialized metamorphic databases are identified,aiming to leverage data-driven applications to achieve broader scientific objectives in metamorphic research.
基金support from the Cyber Technology Institute(CTI)at the School of Computer Science and Informatics,De Montfort University,United Kingdom,along with financial assistance from Universiti Tun Hussein Onn Malaysia and the UTHM Publisher’s office through publication fund E15216.
文摘Integrating machine learning and data mining is crucial for processing big data and extracting valuable insights to enhance decision-making.However,imbalanced target variables within big data present technical challenges that hinder the performance of supervised learning classifiers on key evaluation metrics,limiting their overall effectiveness.This study presents a comprehensive review of both common and recently developed Supervised Learning Classifiers(SLCs)and evaluates their performance in data-driven decision-making.The evaluation uses various metrics,with a particular focus on the Harmonic Mean Score(F-1 score)on an imbalanced real-world bank target marketing dataset.The findings indicate that grid-search random forest and random-search random forest excel in Precision and area under the curve,while Extreme Gradient Boosting(XGBoost)outperforms other traditional classifiers in terms of F-1 score.Employing oversampling methods to address the imbalanced data shows significant performance improvement in XGBoost,delivering superior results across all metrics,particularly when using the SMOTE variant known as the BorderlineSMOTE2 technique.The study concludes several key factors for effectively addressing the challenges of supervised learning with imbalanced datasets.These factors include the importance of selecting appropriate datasets for training and testing,choosing the right classifiers,employing effective techniques for processing and handling imbalanced datasets,and identifying suitable metrics for performance evaluation.Additionally,factors also entail the utilisation of effective exploratory data analysis in conjunction with visualisation techniques to yield insights conducive to data-driven decision-making.
基金The research work was financially supported by the National Natural Science Foundation of China(Grant Nos.51979238 and 52301338)the Sichuan Science and Technology Program(Grant Nos.2023NSFSC1953 and 2023ZYD0140).
文摘Mitigating vortex-induced vibrations(VIV)in flexible risers represents a critical concern in offshore oil and gas production,considering its potential impact on operational safety and efficiency.The accurate prediction of displacement and position of VIV in flexible risers remains challenging under actual marine conditions.This study presents a data-driven model for riser displacement prediction that corresponds to field conditions.Experimental data analysis reveals that the XGBoost algorithm predicts the maximum displacement and position with superior accuracy compared with Support vector regression(SVR),considering both computational efficiency and precision.Platform displacement in the Y-direction demonstrates a significant positive correlation with both axial depth and maximum displacement magnitude.The fourth point displacement exhibits the highest contribution to model prediction outcomes,showing a positive influence on maximum displacement while negatively affecting the axial depth of maximum displacement.Platform displacement in the X-and Y-directions exhibits competitive effects on both the riser’s maximum displacement and its axial depth.Through the implementation of XGBoost algorithm and SHapley Additive exPlanation(SHAP)analysis,the model effectively estimates the riser’s maximum displacement and its precise location.This data-driven approach achieves predictions using minimal,readily available data points,enhancing its practical field applications and demonstrating clear relevance to academic and professional communities.
基金This paper is the research result of“Research on Innovation of Evidence-Based Teaching Paradigm in Vocational Education under the Background of New Quality Productivity”(2024JXQ176)the Shandong Province Artificial Intelligence Education Research Project(SDDJ202501035),which explores the application of artificial intelligence big models in student value-added evaluation from an evidence-based perspective。
文摘Based on the educational evaluation reform,this study explores the construction of an evidence-based value-added evaluation system based on data-driven,aiming to solve the limitations of traditional evaluation methods.The research adopts the method of combining theoretical analysis and practical application,and designs the evidence-based value-added evaluation framework,which includes the core elements of a multi-source heterogeneous data acquisition and processing system,a value-added evaluation agent based on a large model,and an evaluation implementation and application mechanism.Through empirical research verification,the evaluation system has remarkable effects in improving learning participation,promoting ability development,and supporting teaching decision-making,and provides a theoretical reference and practical path for educational evaluation reform in the new era.The research shows that the evidence-based value-added evaluation system based on data-driven can reflect students’actual progress more fairly and objectively by accurately measuring the difference in starting point and development range of students,and provide strong support for the realization of high-quality education development.