The Wufeng–Longmaxi Formation derives its name from the Upper Ordovician Wufeng Formation and the Lower Silurian Longmaxi Formation,found in sequence in the Sichuan Basin.This formation hosts rich shale gas reservoir...The Wufeng–Longmaxi Formation derives its name from the Upper Ordovician Wufeng Formation and the Lower Silurian Longmaxi Formation,found in sequence in the Sichuan Basin.This formation hosts rich shale gas reservoirs,and its shale gas enrichment patterns are examined in this study using data from 1197 shale samples collected from 14 wells.Five basic and three key parameters,eight in all,are assessed for each sample.The five basic parameters include burial depth and the contents of four mineral types—quartz,clay,carbonate,and other minerals;the three key parameters,representing shale gas enrichment,are total organic carbon(TOC)content,porosity,and gas content.The SHapley Additive exPlanations(SHAP)analysis originated in game theory is used here in an interpretable machine learning framework,to address issues of heterogeneous data structure,noisy relationships,and multi-objective optimization.An evaluation of the ranking,contribution values,and conditions of changes for these parameters offers new quantitative insights into shale gas enrichment patterns.A quantitative analysis of the relationship between data-sets identifies the primary factors controlling TOC,porosity,and gas content of shale gas reservoirs.The results show that TOC and porosity jointly influence gas content;mineral content has a significant impact on both,TOC and porosity;and the burial depth governs porosity which,in turn,affects the conditions under which shale gas is preserved.Input parameter thresholds are also determined and provide a basis for the establishment of quantitative criteria to evaluate shale gas enrichment.The predictive accuracy of the model used in this study is significantly improved by the step-wise addition of two input parameters,namely TOC and porosity,separately and together.Thus,the game theory method in big data-driven analysis uses a combination of TOC and porosity to evaluate the gas content with encouraging results—suggesting that these are the key parameters that indicate source rock and reservoir properties.展开更多
Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from sei...Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from seismic networks,satellite observations,and geospatial repositories,creating the need for scalable infrastructures capable of integrating and analyzing such data to support intelligent decision-making.Data warehousing technologies provide a robust foundation for this purpose;however,existing earthquake-oriented data warehouses remain limited,often relying on simplified schemas,domain-specific analytics,or cataloguing efforts.This paper presents the design and implementation of a spatio-temporal data warehouse for seismic activity.The framework integrates spatial and temporal dimensions in a unified schema and introduces a novel array-based approach for managing many-to-many relationships between facts and dimensions without intermediate bridge tables.A comparative evaluation against a conventional bridge-table schema demonstrates that the array-based design improves fact-centric query performance,while the bridge-table schema remains advantageous for dimension-centric queries.To reconcile these trade-offs,a hybrid schema is proposed that retains both representations,ensuring balanced efficiency across heterogeneous workloads.The proposed framework demonstrates how spatio-temporal data warehousing can address schema complexity,improve query performance,and support multidimensional visualization.In doing so,it provides a foundation for integrating seismic analysis into broader big data-driven intelligent decision systems for disaster resilience,risk mitigation,and emergency management.展开更多
The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threa...The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threats is both necessary and complex,yet these interconnected healthcare settings generate enormous amounts of heterogeneous data.Traditional Intrusion Detection Systems(IDS),which are generally centralized and machine learning-based,often fail to address the rapidly changing nature of cyberattacks and are challenged by ethical concerns related to patient data privacy.Moreover,traditional AI-driven IDS usually face challenges in handling large-scale,heterogeneous healthcare data while ensuring data privacy and operational efficiency.To address these issues,emerging technologies such as Big Data Analytics(BDA)and Federated Learning(FL)provide a hybrid framework for scalable,adaptive intrusion detection in IoT-driven healthcare systems.Big data techniques enable processing large-scale,highdimensional healthcare data,and FL can be used to train a model in a decentralized manner without transferring raw data,thereby maintaining privacy between institutions.This research proposes a privacy-preserving Federated Learning–based model that efficiently detects cyber threats in connected healthcare systems while ensuring distributed big data processing,privacy,and compliance with ethical regulations.To strengthen the reliability of the reported findings,the resultswere validated using cross-dataset testing and 95%confidence intervals derived frombootstrap analysis,confirming consistent performance across heterogeneous healthcare data distributions.This solution takes a significant step toward securing next-generation healthcare infrastructure by combining scalability,privacy,adaptability,and earlydetection capabilities.The proposed global model achieves a test accuracy of 99.93%±0.03(95%CI)and amiss-rate of only 0.07%±0.02,representing state-of-the-art performance in privacy-preserving intrusion detection.The proposed FL-driven IDS framework offers an efficient,privacy-preserving,and scalable solution for securing next-generation healthcare infrastructures by combining adaptability,early detection,and ethical data management.展开更多
The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This...The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This era integrates biotechnology,artificial intelligence(AI),and big data information technology.In contrast,China is still in a transition period between stages 2.0 and 3.0,which primarily relies on conventional selection and molecular breeding.In the context of increasingly complex international situations,accurately identifying core issues in China's seed industry innovation and seizing the frontier of international seed technology are strategically important.These efforts are essential for ensuring food security and revitalizing the seed industry.This paper systematically analyzes the characteristics of crop breeding data from artificial selection to intelligent design breeding.It explores the applications and development trends of AI and big data in modern crop breeding from several key perspectives.These include highthroughput phenotype acquisition and analysis,multiomics big data database and management system construction,AI-based multiomics integrated analysis,and the development of intelligent breeding software tools based on biological big data and AI technology.Based on an in-depth analysis of the current status and challenges of China's seed industry technology development,we propose strategic goals and key tasks for China's new generation of AI and big data-driven intelligent design breeding.These suggestions aim to accelerate the development of an intelligent-driven crop breeding engineering system that features large-scale gene mining,efficient gene manipulation,engineered variety design,and systematized biobreeding.This study provides a theoretical basis and practical guidance for the development of China's seed industry technology.展开更多
With the rapid development of the global economy, maritime transportation has become much more convenient due to large capacities and low freight. However, this means the sea lanes are becoming more and more crowded,l...With the rapid development of the global economy, maritime transportation has become much more convenient due to large capacities and low freight. However, this means the sea lanes are becoming more and more crowded,leading to high probabilities of marine accidents in complex maritime environments. According to relevant historical statistics, a large number of accidents have happened in water areas that lack high precision navigation data, which can be utilized to enhance navigation safety. The purpose of this work was to carry out ship route planning automatically, by mining historical big automatic identification system(AIS) data. It is well-known that experiential navigation information hidden in maritime big data could be automatically extracted using advanced data mining techniques;assisting in the generation of safe and reliable ship planning routes for complex maritime environments. In this paper, a novel method is proposed to construct a big data-driven framework for generating ship planning routes automatically, under varying navigation conditions. The method performs density-based spatial clustering of applications with noise first on a large number of ship trajectories to form different trajectory vector clusters. Then, it iteratively calculates its centerline in the trajectory vector cluster, and constructs the waterway network from the node-arc topology relationship among these centerlines. The generation of shipping route could be based on the waterway network and conducted by rasterizing the marine environment risks for the sea area not covered by the waterway network. Numerous experiments have been conducted on different AIS data sets in different water areas, and the experimental results have demonstrated the effectiveness of the framework of the ship route planning proposed in this paper.展开更多
Big data provide valuable insights by offering diverse information and sophisticated analysis through advanced algorithms.However,its huge volume,variety,and speed present significant challenges for effective computin...Big data provide valuable insights by offering diverse information and sophisticated analysis through advanced algorithms.However,its huge volume,variety,and speed present significant challenges for effective computing.To address these,this study applies a Multi-Criteria Decision-Making(MCDM)framework to manage spatial big data,specifically in new green applications.The paper introduces a robust MCDM framework using big data,designed to address renewable energy challenges within the environmental sector.This framework systematically prioritizes and evaluates large environmental datasets,incorporating economic,environmental,and social factors.This framework is especially efficient and reliable for green energy initiatives.Moreover,a pre-processing step extracts key features to enable high-performance efficient analysis and visualization.Results show that the framework improves accuracy by 18%compared to conventional single-criterion data analysis approaches in a large-scale case study and provides system managers with an interactive 3D visualization tool to enhance decision making process in big data environmental management.展开更多
This paper solves the problem of model-free dual-arm space robot maneuvering after non-cooperative target capture under high control quality requirements.The explicit system model is unavailable,and the maneuvering mi...This paper solves the problem of model-free dual-arm space robot maneuvering after non-cooperative target capture under high control quality requirements.The explicit system model is unavailable,and the maneuvering mission is disturbed by the measurement noise and the target adversarial behavior.To address these problems,a model-free Combined Adaptive-length Datadriven Predictive Controller(CADPC)is proposed.It consists of a separated subsystem identification method and a combined predictive control strategy.The subsystem identification method is composed of an adaptive data length,thereby reducing sensitivity to undetermined measurement noises and disturbances.Based on the subsystem identification,the combined predictive controller is established,reducing calculating resource.The stability of the CADPC is rigorously proven using the Input-to-State Stable(ISS)theorem and the small-gain theorem.Simulations demonstrate that CADPC effectively handles the model-free space robot post operation in the presence of significant disturbances,state measurement noise,and control input errors.It achieves improved steady-state accuracy,reduced steady-state control consumption,and minimized control input chattering.展开更多
Artificial intelligence(AI)has reshaped the subject of product innovation and triggered transformations in product innovation strategies and processes.This study proposes a subject-strategy-process(SSP)framework for b...Artificial intelligence(AI)has reshaped the subject of product innovation and triggered transformations in product innovation strategies and processes.This study proposes a subject-strategy-process(SSP)framework for business intelligence(BI)for big data-driven product innovation through logical deduction,drawing on the theory of big data cooperative assets and an adaptive innovation perspective on enterprise-user interaction.The aim is to explore new mechanisms through which AI influences product innovation in manufacturing.This study indicates three aspects.Firstly,the two-way involvement of humans and AI forms a dual feedback-enhancement mechanism of factor combination and knowledge accumulation.This mechanism drives structural changes in innovation subjects and forms a new foundation for strategic and process transformations in product innovation.Secondly,the alignment between an enterprise’s cognitive strategy about AI,competitive strategy,organizational culture,business model,and ecosystem jointly shapes the integrated application of AI in innovation processes.Thirdly,the new features of the big data-driven product innovation process include full-process diffusion from the fuzzy front end,nonlinear iteration of demand-solution pairs,and generative self-testing in intelligent manufacturing.Taken together,the study demonstrates that the SSP framework is well-suited to analyzing the new mechanisms of BI for big data-driven product innovation,which offers a fresh lens for examining the relationship between AI and product innovation.展开更多
This study integrates multiple sources of data(transaction data,policy text,public opinion data)with visualization techniques(such as heat maps,time-series trend charts,3D building brochures)to construct an analysis f...This study integrates multiple sources of data(transaction data,policy text,public opinion data)with visualization techniques(such as heat maps,time-series trend charts,3D building brochures)to construct an analysis framework for the Chengdu real estate market.By using the Adaptive Neuro-Fuzzy Inference System(ANFIS)prediction model,spatial GIS(Geographic Information System analysis)analysis,and interactive dashboards,this study reveals market differentiation,policy impacts,and changes in demand structure,thereby providing decision support for the government,enterprises,and homebuyers.展开更多
To address the issue of instability or even imbalance in the orientation and attitude control of quadrotor unmanned aerial vehicles(QUAVs)under random disturbances,this paper proposes a distributed antidisturbance dat...To address the issue of instability or even imbalance in the orientation and attitude control of quadrotor unmanned aerial vehicles(QUAVs)under random disturbances,this paper proposes a distributed antidisturbance data-driven event-triggered fusion control method,which achieves efficient fault diagnosis while suppressing random disturbances and mitigating communication conflicts within the QUAV swarm.First,the impact of random disturbances on the UAV swarm is analyzed,and a model for orientation and attitude control of QUAVs under stochastic perturbations is established,with the disturbance gain threshold determined.Second,a fault diagnosis system based on a high-gain observer is designed,constructing a fault gain criterion by integrating orientation and attitude information from QUAVs.Subsequently,a model-free dynamic linearization-based data modeling(MFDLDM)framework is developed using model-free adaptive control,which efficiently fits the nonlinear control model of the QUAV swarm while reducing temporal constraints on control data.On this basis,this paper constructs a distributed data-driven event-triggered controller based on the staggered communication mechanism,which consists of an equivalent QUAV controller and an event-triggered controller,and is able to reduce the communication conflicts while suppressing the influence of random interference.Finally,by incorporating random disturbances into the controller,comparative experiments and physical validations are conducted on the QUAV platforms,fully demonstrating the strong adaptability and robustness of the proposed distributed event-triggered fault-tolerant control system.展开更多
Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations a...Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.展开更多
Storm-enhanced density(SED)and the tongue of ionization(TOI)are key ionospheric storm-time structures whose rapid evolution and fine-scale variability remain challenging to capture with conventional empirical high-lat...Storm-enhanced density(SED)and the tongue of ionization(TOI)are key ionospheric storm-time structures whose rapid evolution and fine-scale variability remain challenging to capture with conventional empirical high-latitude drivers.In this study,we examine the May 10–11,2024,superstorm using the Thermosphere–Ionosphere–Electrodynamics General Circulation Model(TIEGCM)with observation-constrained high-latitude forcing.Auroral precipitation parameters(energy flux and mean energy)are assimilated from a Defense Meteorological Satellite Program(DMSP)Special Sensor Ultraviolet Spectrographic Imager(SSUSI)using a multi-resolution Gaussian process(Lattice Kriging)approach,whereas high-latitude convection potentials are derived by assimilating Super Dual Auroral Radar Network(SuperDARN)observations with the Thomas and Shepherd(2018)model(TS18).For comparison,an additional simulation is performed using empirical models for both convection and auroral forcing.The results show that during the main phase of the May 10 storm,the data-driven simulation provides a more realistic depiction of the SED source region than does the empirical model run by capturing its rapid intensification more clearly and reproducing its spatial location and structural features with higher fidelity.These improvements lead to a more accurate representation of its poleward extension into the polar cap that develops into the TOI.Above the ionospheric F2 peak over the SED source region,SuperDARN-constrained potentials generate stronger and more localized E×B drifts that dominate plasma uplift and drive its transport into the polar cap,although neutral winds and downward ambipolar diffusion partially offset these effects.Below the F2 peak,neutral winds and photochemical processes play a major role in shaping the spatial extent and intensity of the SED and TOI.These results highlight the role of observation-constrained high-latitude drivers in representing ionosphere–thermosphere responses during extreme storms and suggest their relevance for improving physical interpretation and model performance.展开更多
在小学高年级英语语篇教学中,存在学生思维浅表化、问题设计碎片化、旧版教材适配难这三个痛点。以译林版英语教材六年级上册Unit 4 Then and now中Story time的教学为例,教师立足教材文本,构建“课前定问—课初引链—课中解链—课后拓...在小学高年级英语语篇教学中,存在学生思维浅表化、问题设计碎片化、旧版教材适配难这三个痛点。以译林版英语教材六年级上册Unit 4 Then and now中Story time的教学为例,教师立足教材文本,构建“课前定问—课初引链—课中解链—课后拓链—全程评链”的五步闭环,用大问题拉主线、小问题搭台阶,能激活学生语篇学习内驱力,实现英语教学从“知识传递”到“素养培养”的转变。展开更多
With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service...With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service resources,uneven distribution,and prominent supply-demand contradictions have seriously affected service quality.Big data technology,with core advantages including data collection,analysis and mining,and accurate prediction,provides a new solution for the allocation of community elderly care service resources.This paper systematically studies the application value of big data technology in the allocation of community elderly care service resources from three aspects:resource allocation efficiency,service accuracy,and management intelligence.Combined with practical needs,it proposes optimal allocation strategies such as building a big data analysis platform and accurately grasping the elderly’s care needs,striving to provide operable path references for the construction of community elderly care service systems,promoting the early realization of the elderly care service goal of“adequate support and proper care for the elderly”,and boosting the high-quality development of China’s elderly care service industry.展开更多
With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues su...With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues such as computer memory capacity and distributed storage of data across different locations,and are unable to directly apply to large-scale data sets.Therefore,in the context of big data,designing efficient and theoretically guaranteed statistical learning and inference algorithms has become a key issue that the current field of statistics urgently needs to address.In this paper,the application status of statistical analysis methods in the big data environment was systematically reviewed,and its future development directions were analyzed to provide reference and support for the further development of theory and methods of the statistical analysis of big data.展开更多
The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physici...The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physician-centered perspective,these technologies can potentially be used to improve diagnostic precision,support individualized treatment planning,enhance longitudinal patient management,and accelerate both clinical and translational research.In this review,we synthesize the core AI methodologies most relevant to oncology-machine learning,deep learning,and large language models-and examine how they interact with established and emerging oncology data platforms.We further highlight practical use cases in clinical workflows and research pipelines,emphasizing opportunities for advancing precision cancer care while also addressing challenges associated with data heterogeneity,model generalizability,privacy protection,and real-world implementation.By underscoring the synergistic value of AI and big data,this review aims to inform the development of clinically meaningful,context-adapted strategies that promote translational innovation in both global and locally resourced healthcare environments.展开更多
大孔径CT基于X线断层扫描原理的超大孔径设计,能容纳肥胖患者以及携带医疗设备的特殊患者,其凭借高清晰成像为放疗计划的制定与实施提供了关键的解剖学依据,是现代放疗“精准化”转型的重要支撑[1]。我院于2017年引进飞利浦Brilliance B...大孔径CT基于X线断层扫描原理的超大孔径设计,能容纳肥胖患者以及携带医疗设备的特殊患者,其凭借高清晰成像为放疗计划的制定与实施提供了关键的解剖学依据,是现代放疗“精准化”转型的重要支撑[1]。我院于2017年引进飞利浦Brilliance Big Bore CT设备,截至2024年底,已累计完成5000余例肿瘤患者的放疗定位扫描。展开更多
The key challenge in the preparation of perovskite solar cells is to enhance the reproducibility of PSC manufacturing,particularly by better controlling multiple high-dimensional process parameters.This study proposes...The key challenge in the preparation of perovskite solar cells is to enhance the reproducibility of PSC manufacturing,particularly by better controlling multiple high-dimensional process parameters.This study proposes a machine learning(ML)approach to efficiently predict and analyze perovskite film fabrication processes.By evaluating five classic ML algorithms on 130 experimental data sets from blade-coating parameters,the Random Forest(RF)model was identified as the most effective,enabling rapid prediction of over 100,000 parameter sets in just 10 min-equivalent to 3 years of manual experimentation.The RF model demonstrated strong predictive accuracy,with an R^(2) close to 0.8.This approach led to the identification of optimal process parameter combinations,significantly improving the reproducibility of PSCs and reducing performance variance by approximately threefold,thereby advancing the development of scalable manufacturing processes.展开更多
To address the severe challenges of PM_(2.5) and ozone co-control during the"14^(th) Five-Year Plan"period and to enhance the precision and intelligence level of air environment governance,it is imperative t...To address the severe challenges of PM_(2.5) and ozone co-control during the"14^(th) Five-Year Plan"period and to enhance the precision and intelligence level of air environment governance,it is imperative to build an efficient comprehensive management platform for regional air quality.In this paper,the specific practice in Zibo City,Shandong Province is as an example to systematically analyze the top-level design,technical implementation,and innovative application of a comprehensive management platform for regional air quality integrating"perception monitoring,data fusion,research judgment of early warnings,analysis of sources,collaborative dispatching,and evaluation assessment".Through the construction of an"sky-air-ground"integrated three-dimensional monitoring network,the platform integrates multi-source heterogeneous environmental data,and employs big data,cloud computing,artificial intelligence,CALPUFF/CMAQ,and other numerical model technologies to achieve comprehensive perception,precise prediction,intelligent source tracing,and closed-loop management of air pollution.The platform innovatively establishes a full-process closed-loop management mechanism of"data-early warning-disposition-evaluation",and achieves a fundamental transformation from passive response to active anticipation and from experience-based judgment to data driving in environmental supervision.The application results show that this platform significantly improves the scientific decision-making ability and collaborative execution efficiency of air pollution governance in Zibo City,providing a replicable and scalable comprehensive solution for similar industrial cities to achieve the continuous improvement of air quality.展开更多
基金funded by the Technical Development(Entrusted)Project of Science and Department of SINOPEC(Grant No.P23240-4)the National Natural Science Foundation of China(Grant Nos.42172165,42272143 and 2025ZD1403901-05)。
文摘The Wufeng–Longmaxi Formation derives its name from the Upper Ordovician Wufeng Formation and the Lower Silurian Longmaxi Formation,found in sequence in the Sichuan Basin.This formation hosts rich shale gas reservoirs,and its shale gas enrichment patterns are examined in this study using data from 1197 shale samples collected from 14 wells.Five basic and three key parameters,eight in all,are assessed for each sample.The five basic parameters include burial depth and the contents of four mineral types—quartz,clay,carbonate,and other minerals;the three key parameters,representing shale gas enrichment,are total organic carbon(TOC)content,porosity,and gas content.The SHapley Additive exPlanations(SHAP)analysis originated in game theory is used here in an interpretable machine learning framework,to address issues of heterogeneous data structure,noisy relationships,and multi-objective optimization.An evaluation of the ranking,contribution values,and conditions of changes for these parameters offers new quantitative insights into shale gas enrichment patterns.A quantitative analysis of the relationship between data-sets identifies the primary factors controlling TOC,porosity,and gas content of shale gas reservoirs.The results show that TOC and porosity jointly influence gas content;mineral content has a significant impact on both,TOC and porosity;and the burial depth governs porosity which,in turn,affects the conditions under which shale gas is preserved.Input parameter thresholds are also determined and provide a basis for the establishment of quantitative criteria to evaluate shale gas enrichment.The predictive accuracy of the model used in this study is significantly improved by the step-wise addition of two input parameters,namely TOC and porosity,separately and together.Thus,the game theory method in big data-driven analysis uses a combination of TOC and porosity to evaluate the gas content with encouraging results—suggesting that these are the key parameters that indicate source rock and reservoir properties.
文摘Earthquakes are highly destructive spatio-temporal phenomena whose analysis is essential for disaster preparedness and risk mitigation.Modern seismological research produces vast volumes of heterogeneous data from seismic networks,satellite observations,and geospatial repositories,creating the need for scalable infrastructures capable of integrating and analyzing such data to support intelligent decision-making.Data warehousing technologies provide a robust foundation for this purpose;however,existing earthquake-oriented data warehouses remain limited,often relying on simplified schemas,domain-specific analytics,or cataloguing efforts.This paper presents the design and implementation of a spatio-temporal data warehouse for seismic activity.The framework integrates spatial and temporal dimensions in a unified schema and introduces a novel array-based approach for managing many-to-many relationships between facts and dimensions without intermediate bridge tables.A comparative evaluation against a conventional bridge-table schema demonstrates that the array-based design improves fact-centric query performance,while the bridge-table schema remains advantageous for dimension-centric queries.To reconcile these trade-offs,a hybrid schema is proposed that retains both representations,ensuring balanced efficiency across heterogeneous workloads.The proposed framework demonstrates how spatio-temporal data warehousing can address schema complexity,improve query performance,and support multidimensional visualization.In doing so,it provides a foundation for integrating seismic analysis into broader big data-driven intelligent decision systems for disaster resilience,risk mitigation,and emergency management.
文摘The increasing number of interconnected devices and the incorporation of smart technology into contemporary healthcare systems have significantly raised the attack surface of cyber threats.The early detection of threats is both necessary and complex,yet these interconnected healthcare settings generate enormous amounts of heterogeneous data.Traditional Intrusion Detection Systems(IDS),which are generally centralized and machine learning-based,often fail to address the rapidly changing nature of cyberattacks and are challenged by ethical concerns related to patient data privacy.Moreover,traditional AI-driven IDS usually face challenges in handling large-scale,heterogeneous healthcare data while ensuring data privacy and operational efficiency.To address these issues,emerging technologies such as Big Data Analytics(BDA)and Federated Learning(FL)provide a hybrid framework for scalable,adaptive intrusion detection in IoT-driven healthcare systems.Big data techniques enable processing large-scale,highdimensional healthcare data,and FL can be used to train a model in a decentralized manner without transferring raw data,thereby maintaining privacy between institutions.This research proposes a privacy-preserving Federated Learning–based model that efficiently detects cyber threats in connected healthcare systems while ensuring distributed big data processing,privacy,and compliance with ethical regulations.To strengthen the reliability of the reported findings,the resultswere validated using cross-dataset testing and 95%confidence intervals derived frombootstrap analysis,confirming consistent performance across heterogeneous healthcare data distributions.This solution takes a significant step toward securing next-generation healthcare infrastructure by combining scalability,privacy,adaptability,and earlydetection capabilities.The proposed global model achieves a test accuracy of 99.93%±0.03(95%CI)and amiss-rate of only 0.07%±0.02,representing state-of-the-art performance in privacy-preserving intrusion detection.The proposed FL-driven IDS framework offers an efficient,privacy-preserving,and scalable solution for securing next-generation healthcare infrastructures by combining adaptability,early detection,and ethical data management.
基金partially supported by the Construction of Collaborative Innovation Center of Beijing Academy of Agricultural and Forestry Sciences(KJCX20240406)the Beijing Natural Science Foundation(JQ24037)+1 种基金the National Natural Science Foundation of China(32330075)the Earmarked Fund for China Agriculture Research System(CARS-02 and CARS-54)。
文摘The security of the seed industry is crucial for ensuring national food security.Currently,developed countries in Europe and America,along with international seed industry giants,have entered the Breeding 4.0 era.This era integrates biotechnology,artificial intelligence(AI),and big data information technology.In contrast,China is still in a transition period between stages 2.0 and 3.0,which primarily relies on conventional selection and molecular breeding.In the context of increasingly complex international situations,accurately identifying core issues in China's seed industry innovation and seizing the frontier of international seed technology are strategically important.These efforts are essential for ensuring food security and revitalizing the seed industry.This paper systematically analyzes the characteristics of crop breeding data from artificial selection to intelligent design breeding.It explores the applications and development trends of AI and big data in modern crop breeding from several key perspectives.These include highthroughput phenotype acquisition and analysis,multiomics big data database and management system construction,AI-based multiomics integrated analysis,and the development of intelligent breeding software tools based on biological big data and AI technology.Based on an in-depth analysis of the current status and challenges of China's seed industry technology development,we propose strategic goals and key tasks for China's new generation of AI and big data-driven intelligent design breeding.These suggestions aim to accelerate the development of an intelligent-driven crop breeding engineering system that features large-scale gene mining,efficient gene manipulation,engineered variety design,and systematized biobreeding.This study provides a theoretical basis and practical guidance for the development of China's seed industry technology.
文摘With the rapid development of the global economy, maritime transportation has become much more convenient due to large capacities and low freight. However, this means the sea lanes are becoming more and more crowded,leading to high probabilities of marine accidents in complex maritime environments. According to relevant historical statistics, a large number of accidents have happened in water areas that lack high precision navigation data, which can be utilized to enhance navigation safety. The purpose of this work was to carry out ship route planning automatically, by mining historical big automatic identification system(AIS) data. It is well-known that experiential navigation information hidden in maritime big data could be automatically extracted using advanced data mining techniques;assisting in the generation of safe and reliable ship planning routes for complex maritime environments. In this paper, a novel method is proposed to construct a big data-driven framework for generating ship planning routes automatically, under varying navigation conditions. The method performs density-based spatial clustering of applications with noise first on a large number of ship trajectories to form different trajectory vector clusters. Then, it iteratively calculates its centerline in the trajectory vector cluster, and constructs the waterway network from the node-arc topology relationship among these centerlines. The generation of shipping route could be based on the waterway network and conducted by rasterizing the marine environment risks for the sea area not covered by the waterway network. Numerous experiments have been conducted on different AIS data sets in different water areas, and the experimental results have demonstrated the effectiveness of the framework of the ship route planning proposed in this paper.
基金supported by the Guangdong Major Project of Basic and Applied Basic Research(No.2023B0303000009).
文摘Big data provide valuable insights by offering diverse information and sophisticated analysis through advanced algorithms.However,its huge volume,variety,and speed present significant challenges for effective computing.To address these,this study applies a Multi-Criteria Decision-Making(MCDM)framework to manage spatial big data,specifically in new green applications.The paper introduces a robust MCDM framework using big data,designed to address renewable energy challenges within the environmental sector.This framework systematically prioritizes and evaluates large environmental datasets,incorporating economic,environmental,and social factors.This framework is especially efficient and reliable for green energy initiatives.Moreover,a pre-processing step extracts key features to enable high-performance efficient analysis and visualization.Results show that the framework improves accuracy by 18%compared to conventional single-criterion data analysis approaches in a large-scale case study and provides system managers with an interactive 3D visualization tool to enhance decision making process in big data environmental management.
基金supported by the National Natural Science Foundation of China(No.12372045)the National Key Research and the Development Program of China(Nos.2023YFC2205900,2023YFC2205901)。
文摘This paper solves the problem of model-free dual-arm space robot maneuvering after non-cooperative target capture under high control quality requirements.The explicit system model is unavailable,and the maneuvering mission is disturbed by the measurement noise and the target adversarial behavior.To address these problems,a model-free Combined Adaptive-length Datadriven Predictive Controller(CADPC)is proposed.It consists of a separated subsystem identification method and a combined predictive control strategy.The subsystem identification method is composed of an adaptive data length,thereby reducing sensitivity to undetermined measurement noises and disturbances.Based on the subsystem identification,the combined predictive controller is established,reducing calculating resource.The stability of the CADPC is rigorously proven using the Input-to-State Stable(ISS)theorem and the small-gain theorem.Simulations demonstrate that CADPC effectively handles the model-free space robot post operation in the presence of significant disturbances,state measurement noise,and control input errors.It achieves improved steady-state accuracy,reduced steady-state control consumption,and minimized control input chattering.
基金supported by the key project of the National Natural Science Foundation of China“Research on the Theory,Methods,and Applications of Innovation via Enterprise-User Interaction Driven by Big Data in the Internet Environment”(No.71832014)the key project of the National Natural Science Foundation of China“Research on the Digital Transformation and Adaptive Management Changes of Manufacturing Enterprises”(No.72032009)the major project of the National Social Science Fund of China“Research on the Impact of Artificial Intelligence on the Transformation and Upgrading of the Manufacturing Industry and Its Governance System”(No.23&DA091).
文摘Artificial intelligence(AI)has reshaped the subject of product innovation and triggered transformations in product innovation strategies and processes.This study proposes a subject-strategy-process(SSP)framework for business intelligence(BI)for big data-driven product innovation through logical deduction,drawing on the theory of big data cooperative assets and an adaptive innovation perspective on enterprise-user interaction.The aim is to explore new mechanisms through which AI influences product innovation in manufacturing.This study indicates three aspects.Firstly,the two-way involvement of humans and AI forms a dual feedback-enhancement mechanism of factor combination and knowledge accumulation.This mechanism drives structural changes in innovation subjects and forms a new foundation for strategic and process transformations in product innovation.Secondly,the alignment between an enterprise’s cognitive strategy about AI,competitive strategy,organizational culture,business model,and ecosystem jointly shapes the integrated application of AI in innovation processes.Thirdly,the new features of the big data-driven product innovation process include full-process diffusion from the fuzzy front end,nonlinear iteration of demand-solution pairs,and generative self-testing in intelligent manufacturing.Taken together,the study demonstrates that the SSP framework is well-suited to analyzing the new mechanisms of BI for big data-driven product innovation,which offers a fresh lens for examining the relationship between AI and product innovation.
基金Chengdu City Philosophy and Social Sciences Research Center“artificial intelligence+urban communication”theory and Application Research Center Project“Chengdu real estate vertical market public opinion data visualization research”(Project No.RZCC2025017).
文摘This study integrates multiple sources of data(transaction data,policy text,public opinion data)with visualization techniques(such as heat maps,time-series trend charts,3D building brochures)to construct an analysis framework for the Chengdu real estate market.By using the Adaptive Neuro-Fuzzy Inference System(ANFIS)prediction model,spatial GIS(Geographic Information System analysis)analysis,and interactive dashboards,this study reveals market differentiation,policy impacts,and changes in demand structure,thereby providing decision support for the government,enterprises,and homebuyers.
基金supported in part by the National Natural Science Foundation of China,Grant/Award Number:62003267the Key Research and Development Program of Shaanxi Province,Grant/Award Number:2023-GHZD-33Open Project of the State Key Laboratory of Intelligent Game,Grant/Award Number:ZBKF-23-05。
文摘To address the issue of instability or even imbalance in the orientation and attitude control of quadrotor unmanned aerial vehicles(QUAVs)under random disturbances,this paper proposes a distributed antidisturbance data-driven event-triggered fusion control method,which achieves efficient fault diagnosis while suppressing random disturbances and mitigating communication conflicts within the QUAV swarm.First,the impact of random disturbances on the UAV swarm is analyzed,and a model for orientation and attitude control of QUAVs under stochastic perturbations is established,with the disturbance gain threshold determined.Second,a fault diagnosis system based on a high-gain observer is designed,constructing a fault gain criterion by integrating orientation and attitude information from QUAVs.Subsequently,a model-free dynamic linearization-based data modeling(MFDLDM)framework is developed using model-free adaptive control,which efficiently fits the nonlinear control model of the QUAV swarm while reducing temporal constraints on control data.On this basis,this paper constructs a distributed data-driven event-triggered controller based on the staggered communication mechanism,which consists of an equivalent QUAV controller and an event-triggered controller,and is able to reduce the communication conflicts while suppressing the influence of random interference.Finally,by incorporating random disturbances into the controller,comparative experiments and physical validations are conducted on the QUAV platforms,fully demonstrating the strong adaptability and robustness of the proposed distributed event-triggered fault-tolerant control system.
基金supported by the National Key R&D Program of China(Grant No.2023YFC3209504)Natural Science Foundation of Wuhan(Grant No.2024040801020271)the Fundamental Research Funds for Central Public Welfare Research Institutes(Grant No.CKSF2025718/YT).
文摘Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.
基金The Shandong Provincial Natural Science Foundation(Grant No.ZR2022JQ18)supported this worksupported by the National Natural Science Foundation of China(NNFSC)Youth Program(Grant No.42304168)+1 种基金supported by the National Key R&D Program of China(Grant No.2022YFF0504400)the NNSFC(Grant Nos.42188101 and 42174210)。
文摘Storm-enhanced density(SED)and the tongue of ionization(TOI)are key ionospheric storm-time structures whose rapid evolution and fine-scale variability remain challenging to capture with conventional empirical high-latitude drivers.In this study,we examine the May 10–11,2024,superstorm using the Thermosphere–Ionosphere–Electrodynamics General Circulation Model(TIEGCM)with observation-constrained high-latitude forcing.Auroral precipitation parameters(energy flux and mean energy)are assimilated from a Defense Meteorological Satellite Program(DMSP)Special Sensor Ultraviolet Spectrographic Imager(SSUSI)using a multi-resolution Gaussian process(Lattice Kriging)approach,whereas high-latitude convection potentials are derived by assimilating Super Dual Auroral Radar Network(SuperDARN)observations with the Thomas and Shepherd(2018)model(TS18).For comparison,an additional simulation is performed using empirical models for both convection and auroral forcing.The results show that during the main phase of the May 10 storm,the data-driven simulation provides a more realistic depiction of the SED source region than does the empirical model run by capturing its rapid intensification more clearly and reproducing its spatial location and structural features with higher fidelity.These improvements lead to a more accurate representation of its poleward extension into the polar cap that develops into the TOI.Above the ionospheric F2 peak over the SED source region,SuperDARN-constrained potentials generate stronger and more localized E×B drifts that dominate plasma uplift and drive its transport into the polar cap,although neutral winds and downward ambipolar diffusion partially offset these effects.Below the F2 peak,neutral winds and photochemical processes play a major role in shaping the spatial extent and intensity of the SED and TOI.These results highlight the role of observation-constrained high-latitude drivers in representing ionosphere–thermosphere responses during extreme storms and suggest their relevance for improving physical interpretation and model performance.
文摘在小学高年级英语语篇教学中,存在学生思维浅表化、问题设计碎片化、旧版教材适配难这三个痛点。以译林版英语教材六年级上册Unit 4 Then and now中Story time的教学为例,教师立足教材文本,构建“课前定问—课初引链—课中解链—课后拓链—全程评链”的五步闭环,用大问题拉主线、小问题搭台阶,能激活学生语篇学习内驱力,实现英语教学从“知识传递”到“素养培养”的转变。
文摘With the accelerating aging process of China’s population,the demand for community elderly care services has shown diversified and personalized characteristics.However,problems such as insufficient total care service resources,uneven distribution,and prominent supply-demand contradictions have seriously affected service quality.Big data technology,with core advantages including data collection,analysis and mining,and accurate prediction,provides a new solution for the allocation of community elderly care service resources.This paper systematically studies the application value of big data technology in the allocation of community elderly care service resources from three aspects:resource allocation efficiency,service accuracy,and management intelligence.Combined with practical needs,it proposes optimal allocation strategies such as building a big data analysis platform and accurately grasping the elderly’s care needs,striving to provide operable path references for the construction of community elderly care service systems,promoting the early realization of the elderly care service goal of“adequate support and proper care for the elderly”,and boosting the high-quality development of China’s elderly care service industry.
文摘With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues such as computer memory capacity and distributed storage of data across different locations,and are unable to directly apply to large-scale data sets.Therefore,in the context of big data,designing efficient and theoretically guaranteed statistical learning and inference algorithms has become a key issue that the current field of statistics urgently needs to address.In this paper,the application status of statistical analysis methods in the big data environment was systematically reviewed,and its future development directions were analyzed to provide reference and support for the further development of theory and methods of the statistical analysis of big data.
基金Hunan Provincial Natural Science Foundation of China,Grant/Award Numbers:2024JJ6289,2023JJ60464,2023JJ60334Changsha City Technology Program,Grant/Award Number:kq2403120+1 种基金Climb Plan of Hunan Cancer Hospital,Grant/Award Numbers:ZX2021005,QH2023006High-Level Talent Support Program of Hunan Cancer Hospital,Grant/Award Number:20250731-1050。
文摘The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physician-centered perspective,these technologies can potentially be used to improve diagnostic precision,support individualized treatment planning,enhance longitudinal patient management,and accelerate both clinical and translational research.In this review,we synthesize the core AI methodologies most relevant to oncology-machine learning,deep learning,and large language models-and examine how they interact with established and emerging oncology data platforms.We further highlight practical use cases in clinical workflows and research pipelines,emphasizing opportunities for advancing precision cancer care while also addressing challenges associated with data heterogeneity,model generalizability,privacy protection,and real-world implementation.By underscoring the synergistic value of AI and big data,this review aims to inform the development of clinically meaningful,context-adapted strategies that promote translational innovation in both global and locally resourced healthcare environments.
文摘大孔径CT基于X线断层扫描原理的超大孔径设计,能容纳肥胖患者以及携带医疗设备的特殊患者,其凭借高清晰成像为放疗计划的制定与实施提供了关键的解剖学依据,是现代放疗“精准化”转型的重要支撑[1]。我院于2017年引进飞利浦Brilliance Big Bore CT设备,截至2024年底,已累计完成5000余例肿瘤患者的放疗定位扫描。
基金Key Research and Development Program of Hubei Province,China(Grant No.2022BAA096)Zhejiang Provincial Natural Science Foundation of China(This material is based upon work funded by Zhejiang Provincial Natural Science Foundation of China under Grant No.LR25A020002)support of the Center for Materials Analysis and Characterization,Material Characterization Lab,and Nanofabrication Lab at Hubei University。
文摘The key challenge in the preparation of perovskite solar cells is to enhance the reproducibility of PSC manufacturing,particularly by better controlling multiple high-dimensional process parameters.This study proposes a machine learning(ML)approach to efficiently predict and analyze perovskite film fabrication processes.By evaluating five classic ML algorithms on 130 experimental data sets from blade-coating parameters,the Random Forest(RF)model was identified as the most effective,enabling rapid prediction of over 100,000 parameter sets in just 10 min-equivalent to 3 years of manual experimentation.The RF model demonstrated strong predictive accuracy,with an R^(2) close to 0.8.This approach led to the identification of optimal process parameter combinations,significantly improving the reproducibility of PSCs and reducing performance variance by approximately threefold,thereby advancing the development of scalable manufacturing processes.
文摘To address the severe challenges of PM_(2.5) and ozone co-control during the"14^(th) Five-Year Plan"period and to enhance the precision and intelligence level of air environment governance,it is imperative to build an efficient comprehensive management platform for regional air quality.In this paper,the specific practice in Zibo City,Shandong Province is as an example to systematically analyze the top-level design,technical implementation,and innovative application of a comprehensive management platform for regional air quality integrating"perception monitoring,data fusion,research judgment of early warnings,analysis of sources,collaborative dispatching,and evaluation assessment".Through the construction of an"sky-air-ground"integrated three-dimensional monitoring network,the platform integrates multi-source heterogeneous environmental data,and employs big data,cloud computing,artificial intelligence,CALPUFF/CMAQ,and other numerical model technologies to achieve comprehensive perception,precise prediction,intelligent source tracing,and closed-loop management of air pollution.The platform innovatively establishes a full-process closed-loop management mechanism of"data-early warning-disposition-evaluation",and achieves a fundamental transformation from passive response to active anticipation and from experience-based judgment to data driving in environmental supervision.The application results show that this platform significantly improves the scientific decision-making ability and collaborative execution efficiency of air pollution governance in Zibo City,providing a replicable and scalable comprehensive solution for similar industrial cities to achieve the continuous improvement of air quality.