期刊文献+
共找到590篇文章
< 1 2 30 >
每页显示 20 50 100
Design of a Private Cloud Platform for Distributed Logging Big Data Based on a Unified Learning Model of Physics and Data 被引量:1
1
作者 Cheng Xi Fu Haicheng Tursyngazy Mahabbat 《Applied Geophysics》 2025年第2期499-510,560,共13页
Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of th... Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of these data has not been well stored,managed and mined.With the development of cloud computing technology,it provides a rare development opportunity for logging big data private cloud.The traditional petrophysical evaluation and interpretation model has encountered great challenges in the face of new evaluation objects.The solution research of logging big data distributed storage,processing and learning functions integrated in logging big data private cloud has not been carried out yet.To establish a distributed logging big-data private cloud platform centered on a unifi ed learning model,which achieves the distributed storage and processing of logging big data and facilitates the learning of novel knowledge patterns via the unifi ed logging learning model integrating physical simulation and data models in a large-scale functional space,thus resolving the geo-engineering evaluation problem of geothermal fi elds.Based on the research idea of“logging big data cloud platform-unifi ed logging learning model-large function space-knowledge learning&discovery-application”,the theoretical foundation of unified learning model,cloud platform architecture,data storage and learning algorithm,arithmetic power allocation and platform monitoring,platform stability,data security,etc.have been carried on analysis.The designed logging big data cloud platform realizes parallel distributed storage and processing of data and learning algorithms.The feasibility of constructing a well logging big data cloud platform based on a unifi ed learning model of physics and data is analyzed in terms of the structure,ecology,management and security of the cloud platform.The case study shows that the logging big data cloud platform has obvious technical advantages over traditional logging evaluation methods in terms of knowledge discovery method,data software and results sharing,accuracy,speed and complexity. 展开更多
关键词 Unified logging learning model logging big data private cloud machine learning
在线阅读 下载PDF
Development and validation of a stroke risk prediction model using regional healthcare big data and machine learning
2
作者 Yunxia Duan Rui Wang +6 位作者 Yumei Sun Wendi Zhu Yi Li Na Yu Yu Zhu Peng Shen Hongyu Sun 《International Journal of Nursing Sciences》 2025年第6期558-565,I0002,共9页
Objectives:This study aimed to develop and validate a stroke risk prediction model based on machine learning(ML)and regional healthcare big data,and determine whether it may improve the prediction performance compared... Objectives:This study aimed to develop and validate a stroke risk prediction model based on machine learning(ML)and regional healthcare big data,and determine whether it may improve the prediction performance compared with the conventional Logistic Regression(LR)model.Methods:This retrospective cohort study analyzed data from the CHinese Electronic health Records Research in Yinzhou(CHERRY)(2015–2021).We included adults aged 18–75 from the platform who had established records before 2015.Individuals with pre-existing stroke,key data absence,or excessive missingness(>30%)were excluded.Data on demographic,clinical measures,lifestyle factors,comorbidities,and family history of stroke were collected.Variable selection was performed in two stages:an initial screening via univariate analysis,followed by a prioritization of variables based on clinical relevance and actionability,with a focus on those that are modifiable.Stroke prediction models were developed using LR and four ML algorithms:Decision Tree(DT),Random Forest(RF),eXtreme Gradient Boosting(XGBoost),and Back Propagation Neural Network(BPNN).The dataset was split 7:3 for training and validation sets.Performance was assessed using receiver operating characteristic(ROC)curves,calibration,and confusion matrices,and the cutoff value was determined by Youden's index to classify risk groups.Results:The study cohort comprised 92,172 participants with 436 incident stroke cases(incidence rate:474/100,000 person-years).Ultimately,13 predictor variables were included.RF achieved the highest accuracy(0.935),precision(0.923),sensitivity(recall:0.947),and F1 score(0.935).Model evaluation demonstrated superior predictive performance of ML algorithms over conventional LR,with training/validation areaunderthe curve(AUC)sof0.777/0.779(LR),0.921/0.918(BPNN),0.988/0.980(RF),0.980/0.955(DT),and 0.962/0.958(XGBoost).Calibration analysis revealed a better fit for DT,LR and BPNN compared to RF and XGBoost model.Based on the optimal performance of the RF model,the ranking of factors in descending order of importance was:hypertension,age,diabetes,systolic blood pressure,waist,high-density lipoprotein Cholesterol,fasting blood glucose,physical activity,BMI,low-density lipoprotein cholesterol,total cholesterol,dietary habits,and family history of stroke.Using Youden's index as the optimal cutoff,the RF model stratified individuals into high-risk(>0.789)and low-risk(≤0.789)groups with robust discrimination.Conclusions:The ML-based prediction models demonstrated superior performance metrics compared to conventional LR and the RF is the optimal prediction model,providing an effective tool for risk stratifi cation in primary stroke prevention in community settings. 展开更多
关键词 big data Machine learning NURSING Prediction model STROKE
暂未订购
A systematic data-driven modelling framework for nonlinear distillation processes incorporating data intervals clustering and new integrated learning algorithm
3
作者 Zhe Wang Renchu He Jian Long 《Chinese Journal of Chemical Engineering》 2025年第5期182-199,共18页
The distillation process is an important chemical process,and the application of data-driven modelling approach has the potential to reduce model complexity compared to mechanistic modelling,thus improving the efficie... The distillation process is an important chemical process,and the application of data-driven modelling approach has the potential to reduce model complexity compared to mechanistic modelling,thus improving the efficiency of process optimization or monitoring studies.However,the distillation process is highly nonlinear and has multiple uncertainty perturbation intervals,which brings challenges to accurate data-driven modelling of distillation processes.This paper proposes a systematic data-driven modelling framework to solve these problems.Firstly,data segment variance was introduced into the K-means algorithm to form K-means data interval(KMDI)clustering in order to cluster the data into perturbed and steady state intervals for steady-state data extraction.Secondly,maximal information coefficient(MIC)was employed to calculate the nonlinear correlation between variables for removing redundant features.Finally,extreme gradient boosting(XGBoost)was integrated as the basic learner into adaptive boosting(AdaBoost)with the error threshold(ET)set to improve weights update strategy to construct the new integrated learning algorithm,XGBoost-AdaBoost-ET.The superiority of the proposed framework is verified by applying this data-driven modelling framework to a real industrial process of propylene distillation. 展开更多
关键词 Integrated learning algorithm data intervals clustering Feature selection Application of artificial intelligence in distillation industry data-driven modelling
在线阅读 下载PDF
The Interdisciplinary Research of Big Data and Wireless Channel: A Cluster-Nuclei Based Channel Model 被引量:27
4
作者 Jianhua Zhang 《China Communications》 SCIE CSCD 2016年第S2期14-26,共13页
Recently,internet stimulates the explosive progress of knowledge discovery in big volume data resource,to dig the valuable and hidden rules by computing.Simultaneously,the wireless channel measurement data reveals big... Recently,internet stimulates the explosive progress of knowledge discovery in big volume data resource,to dig the valuable and hidden rules by computing.Simultaneously,the wireless channel measurement data reveals big volume feature,considering the massive antennas,huge bandwidth and versatile application scenarios.This article firstly presents a comprehensive survey of channel measurement and modeling research for mobile communication,especially for 5th Generation(5G) and beyond.Considering the big data research progress,then a cluster-nuclei based model is proposed,which takes advantages of both the stochastical model and deterministic model.The novel model has low complexity with the limited number of cluster-nuclei while the cluster-nuclei has the physical mapping to real propagation objects.Combining the channel properties variation principles with antenna size,frequency,mobility and scenario dug from the channel data,the proposed model can be expanded in versatile application to support future mobile research. 展开更多
关键词 channel model big data 5G massive MIMO machine learning CLUSTER
在线阅读 下载PDF
Foundation Study on Wireless Big Data: Concept, Mining, Learning and Practices 被引量:10
5
作者 Jinkang Zhu Chen Gong +2 位作者 Sihai Zhang Ming Zhao Wuyang Zhou 《China Communications》 SCIE CSCD 2018年第12期1-15,共15页
Facing the development of future 5 G, the emerging technologies such as Internet of things, big data, cloud computing, and artificial intelligence is enhancing an explosive growth in data traffic. Radical changes in c... Facing the development of future 5 G, the emerging technologies such as Internet of things, big data, cloud computing, and artificial intelligence is enhancing an explosive growth in data traffic. Radical changes in communication theory and implement technologies, the wireless communications and wireless networks have entered a new era. Among them, wireless big data(WBD) has tremendous value, and artificial intelligence(AI) gives unthinkable possibilities. However, in the big data development and artificial intelligence application groups, the lack of a sound theoretical foundation and mathematical methods is regarded as a real challenge that needs to be solved. From the basic problem of wireless communication, the interrelationship of demand, environment and ability, this paper intends to investigate the concept and data model of WBD, the wireless data mining, the wireless knowledge and wireless knowledge learning(WKL), and typical practices examples, to facilitate and open up more opportunities of WBD research and developments. Such research is beneficial for creating new theoretical foundation and emerging technologies of future wireless communications. 展开更多
关键词 WIRELESS big data data model data MINING WIRELESS KNOWLEDGE KNOWLEDGE learning future WIRELESS communications
在线阅读 下载PDF
Modeling of Optimal Deep Learning Based Flood Forecasting Model Using Twitter Data 被引量:1
6
作者 G.Indra N.Duraipandian 《Intelligent Automation & Soft Computing》 SCIE 2023年第2期1455-1470,共16页
Aflood is a significant damaging natural calamity that causes loss of life and property.Earlier work on the construction offlood prediction models intended to reduce risks,suggest policies,reduce mortality,and limit prop... Aflood is a significant damaging natural calamity that causes loss of life and property.Earlier work on the construction offlood prediction models intended to reduce risks,suggest policies,reduce mortality,and limit property damage caused byfloods.The massive amount of data generated by social media platforms such as Twitter opens the door toflood analysis.Because of the real-time nature of Twitter data,some government agencies and authorities have used it to track natural catastrophe events in order to build a more rapid rescue strategy.However,due to the shorter duration of Tweets,it is difficult to construct a perfect prediction model for determiningflood.Machine learning(ML)and deep learning(DL)approaches can be used to statistically developflood prediction models.At the same time,the vast amount of Tweets necessitates the use of a big data analytics(BDA)tool forflood prediction.In this regard,this work provides an optimal deep learning-basedflood forecasting model with big data analytics(ODLFF-BDA)based on Twitter data.The suggested ODLFF-BDA technique intends to anticipate the existence offloods using tweets in a big data setting.The ODLFF-BDA technique comprises data pre-processing to convert the input tweets into a usable format.In addition,a Bidirectional Encoder Representations from Transformers(BERT)model is used to generate emotive contextual embed-ding from tweets.Furthermore,a gated recurrent unit(GRU)with a Multilayer Convolutional Neural Network(MLCNN)is used to extract local data and predict theflood.Finally,an Equilibrium Optimizer(EO)is used tofine-tune the hyper-parameters of the GRU and MLCNN models in order to increase prediction performance.The memory usage is pull down lesser than 3.5 MB,if its compared with the other algorithm techniques.The ODLFF-BDA technique’s performance was validated using a benchmark Kaggle dataset,and thefindings showed that it outperformed other recent approaches significantly. 展开更多
关键词 big data analytics predictive models deep learning flood prediction twitter data hyperparameter tuning
在线阅读 下载PDF
Research on a Fog Computing Architecture and BP Algorithm Application for Medical Big Data
7
作者 Baoling Qin 《Intelligent Automation & Soft Computing》 SCIE 2023年第7期255-267,共13页
Although the Internet of Things has been widely applied,the problems of cloud computing in the application of digital smart medical Big Data collection,processing,analysis,and storage remain,especially the low efficie... Although the Internet of Things has been widely applied,the problems of cloud computing in the application of digital smart medical Big Data collection,processing,analysis,and storage remain,especially the low efficiency of medical diagnosis.And with the wide application of the Internet of Things and Big Data in the medical field,medical Big Data is increasing in geometric magnitude resulting in cloud service overload,insufficient storage,communication delay,and network congestion.In order to solve these medical and network problems,a medical big-data-oriented fog computing architec-ture and BP algorithm application are proposed,and its structural advantages and characteristics are studied.This architecture enables the medical Big Data generated by medical edge devices and the existing data in the cloud service center to calculate,compare and analyze the fog node through the Internet of Things.The diagnosis results are designed to reduce the business processing delay and improve the diagnosis effect.Considering the weak computing of each edge device,the artificial intelligence BP neural network algorithm is used in the core computing model of the medical diagnosis system to improve the system computing power,enhance the medical intelligence-aided decision-making,and improve the clinical diagnosis and treatment efficiency.In the application process,combined with the characteristics of medical Big Data technology,through fog architecture design and Big Data technology integration,we could research the processing and analysis of heterogeneous data of the medical diagnosis system in the context of the Internet of Things.The results are promising:The medical platform network is smooth,the data storage space is sufficient,the data processing and analysis speed is fast,the diagnosis effect is remarkable,and it is a good assistant to doctors’treatment effect.It not only effectively solves the problem of low clinical diagnosis,treatment efficiency and quality,but also reduces the waiting time of patients,effectively solves the contradiction between doctors and patients,and improves the medical service quality and management level. 展开更多
关键词 Medical big data IOT fog computing distributed computing BP algorithm model
在线阅读 下载PDF
A Survey of Machine Learning for Big Data Processing
8
作者 Reem Almutiri Sarah Alhabeeb +1 位作者 Sarah Alhumud Rehan Ullah Khan 《Journal on Big Data》 2022年第2期97-111,共15页
Today’s world is a data-driven one,with data being produced in vast amounts as a result of the rapid growth of technology that permeates every aspect of our lives.New data processing techniques must be developed and ... Today’s world is a data-driven one,with data being produced in vast amounts as a result of the rapid growth of technology that permeates every aspect of our lives.New data processing techniques must be developed and refined over time to gain meaningful insights from this vast continuous volume of produced data in various forms.Machine learning technologies provide promising solutions and potential methods for processing large quantities of data and gaining value from it.This study conducts a literature review on the application of machine learning techniques in big data processing.It provides a general overview of machine learning algorithms and techniques,a brief introduction to big data,and a discussion of related works that have used machine learning techniques in a variety of sectors to process big amounts of data.The study also discusses the challenges and issues associated with the usage of machine learning for big data. 展开更多
关键词 Machine learning big data PROCESSING algorithmS
在线阅读 下载PDF
Deep Learning Based Optimal Multimodal Fusion Framework for Intrusion Detection Systems for Healthcare Data 被引量:1
9
作者 Phong Thanh Nguyen Vy Dang Bich Huynh +3 位作者 Khoa Dang Vo Phuong Thanh Phan Mohamed Elhoseny Dac-Nhuong Le 《Computers, Materials & Continua》 SCIE EI 2021年第3期2555-2571,共17页
Data fusion is a multidisciplinary research area that involves different domains.It is used to attain minimum detection error probability and maximum reliability with the help of data retrieved from multiple healthcar... Data fusion is a multidisciplinary research area that involves different domains.It is used to attain minimum detection error probability and maximum reliability with the help of data retrieved from multiple healthcare sources.The generation of huge quantity of data from medical devices resulted in the formation of big data during which data fusion techniques become essential.Securing medical data is a crucial issue of exponentially-pacing computing world and can be achieved by Intrusion Detection Systems(IDS).In this regard,since singularmodality is not adequate to attain high detection rate,there is a need exists to merge diverse techniques using decision-based multimodal fusion process.In this view,this research article presents a new multimodal fusion-based IDS to secure the healthcare data using Spark.The proposed model involves decision-based fusion model which has different processes such as initialization,pre-processing,Feature Selection(FS)and multimodal classification for effective detection of intrusions.In FS process,a chaotic Butterfly Optimization(BO)algorithmcalled CBOA is introduced.Though the classic BO algorithm offers effective exploration,it fails in achieving faster convergence.In order to overcome this,i.e.,to improve the convergence rate,this research work modifies the required parameters of BO algorithm using chaos theory.Finally,to detect intrusions,multimodal classifier is applied by incorporating three Deep Learning(DL)-based classification models.Besides,the concepts like Hadoop MapReduce and Spark were also utilized in this study to achieve faster computation of big data in parallel computation platform.To validate the outcome of the presented model,a series of experimentations was performed using the benchmark NSLKDDCup99 Dataset repository.The proposed model demonstrated its effective results on the applied dataset by offering the maximum accuracy of 99.21%,precision of 98.93%and detection rate of 99.59%.The results assured the betterment of the proposed model. 展开更多
关键词 big data data fusion deep learning intrusion detection bio-inspired algorithm SPARK
在线阅读 下载PDF
AutoML for calorific value prediction using a large database from the coal gasification practices in China
10
作者 Yuchao Guo Xia Liu +9 位作者 Yunfei Gao Xiaoyu Wang Lu Ding Weitong Pan Cheng Hua Yulian He Xueli Chen Zhenghua Dai Guangsuo Yu Fuchen Wang 《International Journal of Coal Science & Technology》 2025年第4期230-246,共17页
Calorific value is one of the most important properties of coal.Machine learning(ML)can be used in the prediction of calorific value to reduce experimental costs.China is one of the world’s largest coal production co... Calorific value is one of the most important properties of coal.Machine learning(ML)can be used in the prediction of calorific value to reduce experimental costs.China is one of the world’s largest coal production countries and coal occupies an important position in its national energy structure.However,ML models with a large database for the overall regions of China are still missing.Based on the extensive coal gasification practices in East China University of Science and Technology,we have built ML models with a large database for overall regions of China.An AutoML model was proposed and achieved a minimum MSE of 1.021.SHAP method was used to increase the model interpretability,and model validity was proved with literature data and additional in-house experiments.The model adaptability was discussed based on the databases of China and USA,showing that geography-specific ML models are essential.This study integrated a large coal database and AutoML method for accurate calorific value prediction and could offer key tools for Chinese coal industry. 展开更多
关键词 Coal calorific value big data Automated machine learning model interpretability model adaptability
在线阅读 下载PDF
Optimization of convolutional neural networks for predicting water pollutants using spectral data in the middle and lower reaches of the Yangtze River Basin,China
11
作者 ZHANG Guohao LI Song +3 位作者 WANG Cailing WANG Hongwei YU Tao DAI Xiaoxu 《Journal of Mountain Science》 2025年第8期2851-2869,共19页
Developing an accurate and efficient comprehensive water quality prediction model and its assessment method is crucial for the prevention and control of water pollution.Deep learning(DL),as one of the most promising t... Developing an accurate and efficient comprehensive water quality prediction model and its assessment method is crucial for the prevention and control of water pollution.Deep learning(DL),as one of the most promising technologies today,plays a crucial role in the effective assessment of water body health,which is essential for water resource management.This study models using both the original dataset and a dataset augmented with Generative Adversarial Networks(GAN).It integrates optimization algorithms(OA)with Convolutional Neural Networks(CNN)to propose a comprehensive water quality model evaluation method aiming at identifying the optimal models for different pollutants.Specifically,after preprocessing the spectral dataset,data augmentation was conducted to obtain two datasets.Then,six new models were developed on these datasets using particle swarm optimization(PSO),genetic algorithm(GA),and simulated annealing(SA)combined with CNN to simulate and forecast the concentrations of three water pollutants:Chemical Oxygen Demand(COD),Total Nitrogen(TN),and Total Phosphorus(TP).Finally,seven model evaluation methods,including uncertainty analysis,were used to evaluate the constructed models and select the optimal models for the three pollutants.The evaluation results indicate that the GPSCNN model performed best in predicting COD and TP concentrations,while the GGACNN model excelled in TN concentration prediction.Compared to existing technologies,the proposed models and evaluation methods provide a more comprehensive and rapid approach to water body prediction and assessment,offering new insights and methods for water pollution prevention and control. 展开更多
关键词 Water pollutants Convolutional neural networks data augmentation Optimization algorithms model evaluation methods Deep learning
原文传递
Modeling potential wetland distributions in China based on geographic big data and machine learning algorithms 被引量:2
12
作者 Hengxing Xiang Yanbiao Xi +5 位作者 Dehua Mao Tianyuan Xu Ming Wang Fudong Yu Kaidong Feng Zongming Wang 《International Journal of Digital Earth》 SCIE EI 2023年第1期3706-3724,共19页
Climate change and human activities have reduced the area and degraded the functions and services of wetlands in China.To protect and restore wetlands,it is urgent to predict the spatial distribution of potential wetl... Climate change and human activities have reduced the area and degraded the functions and services of wetlands in China.To protect and restore wetlands,it is urgent to predict the spatial distribution of potential wetlands.In this study,the distribution of potential wetlands in China was simulated by integrating the advantages of Google Earth Engine with geographic big data and machine learning algorithms.Based on a potential wetland database with 46,000 samples and an indicator system of 30 hydrologic,soil,vegetation,and topographic factors,a simulation model was constructed by machine learning algorithms.The accuracy of the random forest model for simulating the distribution of potential wetlands in China was good,with an area under the receiver operating characteristic curve value of 0.851.The area of potential wetlands was 332,702 km^(2),with 39.0%of potential wetlands in Northeast China.Geographic features were notable,and potential wetlands were mainly concentrated in areas with 400-600 mm precipitation,semi-hydric and hydric soils,meadow and marsh vegetation,altitude less than 700 m,and slope less than 3°.The results provide an important reference for wetland remote sensing mapping and a scientific basis for wetland management in China. 展开更多
关键词 Potential wetland distribution machine learning algorithms geographic big data China wetland geographic features
原文传递
一种采用渐进学习模式的SBS-CLearning分类算法 被引量:3
13
作者 申彦 朱玉全 宋新平 《江苏大学学报(自然科学版)》 EI CAS CSCD 北大核心 2018年第6期696-703,共8页
针对Learn++. NSE算法中多个基分类器之间相互独立、未利用前阶段学习结果辅助后续阶段学习而准确率较低的问题,借鉴人类的学习过程,优化Learn++. NSE算法内部的学习机制,转变基分类器的独立学习为渐进学习,提出了一种采用渐进学习模式... 针对Learn++. NSE算法中多个基分类器之间相互独立、未利用前阶段学习结果辅助后续阶段学习而准确率较低的问题,借鉴人类的学习过程,优化Learn++. NSE算法内部的学习机制,转变基分类器的独立学习为渐进学习,提出了一种采用渐进学习模式的SBS-CLearning分类算法.分析了Learn++. NSE算法的不足.给出了SBS-CLearning算法的步骤,该算法在前阶段基分类器的基础之上先增量学习,再完成最终的加权集成.在测试数据集上对比分析了Learn++. NSE与SBSCLearning的分类准确率.试验结果表明:SBS-CLearning算法吸收了增量学习与集成学习的优势,相比Learn++. NSE提高了分类准确率.针对SEA人工数据集,SBS-CLearning,Learn++. NSE的平均分类准确率分别为0. 982,0. 976.针对旋转棋盘真实数据集,在Constant,Sinusoidal,Pulse环境下,SBS-CLearning的平均分类准确率分别为0. 624,0. 655,0. 662,而Learn++. NSE分别为0. 593,0. 633,0. 629. 展开更多
关键词 大数据挖掘 分类算法 集成学习 增量学习 概念漂移
在线阅读 下载PDF
基于自然语言处理的职务犯罪法律文书处理与分析研究
14
作者 姜志超 杨炳文 +1 位作者 高谷刚 李林怡 《通信与信息技术》 2026年第1期7-12,30,共7页
近年来,职务犯罪案件频发,现有研究多局限于法律文本和犯罪构成分析,缺乏跨学科视角,难以揭示其特征和发展趋势。目前,专门针对职务犯罪文书处理与分析的类似系统较少,法律领域通用的数据分析系统难以处理此类文书的专业性和特殊性。因... 近年来,职务犯罪案件频发,现有研究多局限于法律文本和犯罪构成分析,缺乏跨学科视角,难以揭示其特征和发展趋势。目前,专门针对职务犯罪文书处理与分析的类似系统较少,法律领域通用的数据分析系统难以处理此类文书的专业性和特殊性。因此,借助大数据、人工智能和自然语言处理技术,分析职务犯罪案例文本,揭示犯罪规律并实现高效预防具有重要意义。本研究提出基于智能数据处理与分析的职务犯罪研究模型与算法,并构建了系统原型。通过定制化爬虫技术高效采集多平台职务犯罪文书数据。在数据预处理阶段,采用jieba分词结合深度学习序列标注技术进行清洗、分词及关键信息提取。基于Word2Vec模型将文本信息转化为数字化表达,并结合K-Means聚类算法与Llama3大语言模型挖掘关键特征,显著提升类案检索精准性。最终通过箱线图、散点图等可视化手段展示犯罪规律。实验结果表明,相较于传统方法,该模型在精确度和召回率方面分别提升了21%和9%,充分验证了Llama3在语义理解和特征提取方面的强大能力。 展开更多
关键词 职务犯罪 法律文书 大数据 自然语言处理 词向量模型 聚类算法
在线阅读 下载PDF
材料数据库的现状与未来:AI技术引领的创新应用前景
15
作者 冯建发 王畅畅 +1 位作者 苏航 宿彦京 《中国材料进展》 北大核心 2026年第2期89-101,共13页
随着人工智能技术的不断进步,材料数据库在材料科学研究中扮演着日益重要的角色。旨在探讨材料数据库如何通过与AI技术的融合,扩展其应用范围并提升其核心价值。通过文献综述的方法,系统地分析了材料数据库的当前分类,包括材料基础数据... 随着人工智能技术的不断进步,材料数据库在材料科学研究中扮演着日益重要的角色。旨在探讨材料数据库如何通过与AI技术的融合,扩展其应用范围并提升其核心价值。通过文献综述的方法,系统地分析了材料数据库的当前分类,包括材料基础数据库、生产加工数据库、应用服役数据库等,并概述了支撑技术如机器学习、深度学习、数据标准化技术的应用情况。尽管国际上材料数据库的发展呈现出智能化、网络化、资产化、去中心化的趋势,但在数据质量、数据共享、知识产权、市场运维等方面仍面临挑战。未来材料数据库的发展将受益于与新兴技术如材料数据工厂、区块链、隐私计算、AI大模型的结合,这将为新材料的研发和应用提供创新的手段和场景工具。 展开更多
关键词 材料数据库 大数据技术 AI 机器学习 大模型
在线阅读 下载PDF
大模型技术在物资储备应用的前景展望
16
作者 宁德金 王瑾 +3 位作者 王昕灵 赵子明 周园园 赖伟玲 《粮油食品科技》 北大核心 2026年第1期214-221,共8页
国家物资储备体系是国家储备体系的核心,是保障国家安全的重要组成部分,其高效、智能、安全管理直接关乎社会稳定与国家安全。在全球供应链风险加剧与国家战略安全需求升级的双重背景下,大模型技术作为人工智能领域最前沿最先进的技术代... 国家物资储备体系是国家储备体系的核心,是保障国家安全的重要组成部分,其高效、智能、安全管理直接关乎社会稳定与国家安全。在全球供应链风险加剧与国家战略安全需求升级的双重背景下,大模型技术作为人工智能领域最前沿最先进的技术代表,凭借其强大的多模态数据融合、复杂决策优化与自适应学习能力,为构建智能韧性储备体系提供了关键驱动力。本文系统性地构建大模型赋能国家物资储备管理的整体研究框架。首先,阐述了大模型的技术演进与核心优势,深度剖析其在物资储备领域的应用必然性。其次,结合大模型技术特性和储备业务痛点诉求,重点展望其在智能预测、动态优化、应急响应、智能调度、智慧化管理、风险模拟、知识共享等核心场景的赋能路径,并给出赋能涉及的关键技术要点。最后,从技术、管理等维度分析人工智能赋能物资储备面临的挑战。旨在为理论研究、技术攻关和实践探索提供方向指引,推动大模型技术与国家物资储备需求深度融合。 展开更多
关键词 物资储备 人工智能 大模型 大数据 机器学习
在线阅读 下载PDF
联邦学习驱动下串口服务器数据安全保障与同步效率研究
17
作者 陶俊 梁翀 +2 位作者 郭庆 郭力旋 喻成琛 《电子设计工程》 2026年第6期35-40,共6页
为解决企业业务持续拓展背景下串口服务器数据安全性和同步效率较低的问题,设计了一种基于联邦学习与聚合算法的数据加密和同步模型。该模型通过优化聚合算法对数据进行加密,构建联邦学习模型将全局模型权重拆分,实现数据高效同步。结... 为解决企业业务持续拓展背景下串口服务器数据安全性和同步效率较低的问题,设计了一种基于联邦学习与聚合算法的数据加密和同步模型。该模型通过优化聚合算法对数据进行加密,构建联邦学习模型将全局模型权重拆分,实现数据高效同步。结果显示,当参与方数量增至10时,所提模型的同步时间最高仅为150 s,显著低于聚类模型,证明其数据传输同步效率较高。所提出的模型性能优越,为企业在业务扩展中实现高效、安全的数据交互提供了技术支持。 展开更多
关键词 联邦学习模型 聚合算法 串口服务器 数据安全 树型联邦学习模型
在线阅读 下载PDF
基于大数据分析的水资源调度优化方法探讨
18
作者 戴儒光 梁增荣 赵珂 《科学技术创新》 2026年第1期182-185,共4页
传统水资源调度方法在复杂水文环境下面临预测精度不足与决策滞后等问题导致水资源配置效率低下,文章基于大数据分析构建了水资源调度优化模型,运用机器学习对历史水文数据深度挖掘并结合实时数据建立动态预测机制,采用多目标优化算法... 传统水资源调度方法在复杂水文环境下面临预测精度不足与决策滞后等问题导致水资源配置效率低下,文章基于大数据分析构建了水资源调度优化模型,运用机器学习对历史水文数据深度挖掘并结合实时数据建立动态预测机制,采用多目标优化算法实现精准调度。实验结果表明,该方法预测精度提升23.7%,调度响应时间缩短至15分钟内,水资源利用率提高18.5%,有效改善了水资源时空分布不均,为区域水资源可持续管理提供了技术支持。 展开更多
关键词 大数据分析 水资源调度 优化方法 机器学习 预测模型
在线阅读 下载PDF
Progress on the world's primate hotspots and coldspots:modeling ensemble super SDMs in cloud-computers based on digital citizen-science big data and 200+predictors for more sustainable conservation planning
19
作者 Moriz Steiner Falk Huettmann 《Ecological Processes》 2025年第3期82-126,共45页
Background Describing where distribution hotspots and coldspots are located is crucial for any science-based species management and governance.Thus,here we created the world's first Super Species Distribution Mode... Background Describing where distribution hotspots and coldspots are located is crucial for any science-based species management and governance.Thus,here we created the world's first Super Species Distribution Models(SDMs)including all described primate species and the best-available predictor set.These Super SDMs are conducted using an ensemble of modern Machine Learning algorithms,including Maxent,Tree Net,Random Forest,CART,CART Boosting and Bagging,and MARS with the utilization of cloud supercomputers(as an add-on option for more powerful models).For the global cold/hotspot models,we obtained global distribution data from www.GBIF.org(approx.420,000 raw occurrence records)and utilized the world's largest Open Access environmental predictor set of 201 layers.For this analysis,all occurrences have been merged into one multi-species(400+species)pixel-based analysis.Results We present the first quantified pixel-based global primate hotspot prediction for Central and Northern South America,West Africa,East Africa,Southeast Asia,Central Asia,and Southern Africa.The global primate coldspots are Antarctica,the Arctic,most temperate regions,and Oceania past the Wallace line.We additionally described all these modeled hotspots/coldspots and discussed reasons for a quantified understanding of where the world's non-human primates occur(or not).Conclusions This shows us where the focus for most future research and conservation management efforts should be,using state-of-the-art digital data indication tools with reasoning.Those areas should be considered of the highest conservation management priority,ideally following‘no killing zones'and sustainable land stewardship approaches if primates are to have a chance of survival. 展开更多
关键词 PRIMATES Species distribution modeling big data Cloud computing Machine learning(ML) Citizenscience data Open access Remote sensing
原文传递
Variogram modelling optimisation using genetic algorithm and machine learning linear regression:application for Sequential Gaussian Simulations mapping
20
作者 André William Boroh Alpha Baster Kenfack Fokem +2 位作者 Martin Luther Mfenjou Firmin Dimitry Hamat Fritz Mbounja Besseme 《Artificial Intelligence in Geosciences》 2025年第1期177-190,共14页
The objective of this study is to develop an advanced approach to variogram modelling by integrating genetic algorithms(GA)with machine learning-based linear regression,aiming to improve the accuracy and efficiency of... The objective of this study is to develop an advanced approach to variogram modelling by integrating genetic algorithms(GA)with machine learning-based linear regression,aiming to improve the accuracy and efficiency of geostatistical analysis,particularly in mineral exploration.The study combines GA and machine learning to optimise variogram parameters,including range,sill,and nugget,by minimising the root mean square error(RMSE)and maximising the coefficient of determination(R^(2)).The experimental variograms were computed and modelled using theoretical models,followed by optimisation via evolutionary algorithms.The method was applied to gravity data from the Ngoura-Batouri-Kette mining district in Eastern Cameroon,covering 141 data points.Sequential Gaussian Simulations(SGS)were employed for predictive mapping to validate simulated results against true values.Key findings show variograms with ranges between 24.71 km and 49.77 km,opti-mised RMSE and R^(2) values of 11.21 mGal^(2) and 0.969,respectively,after 42 generations of GA optimisation.Predictive mapping using SGS demonstrated that simulated values closely matched true values,with the simu-lated mean at 21.75 mGal compared to the true mean of 25.16 mGal,and variances of 465.70 mGal^(2) and 555.28 mGal^(2),respectively.The results confirmed spatial variability and anisotropies in the N170-N210 directions,consistent with prior studies.This work presents a novel integration of GA and machine learning for variogram modelling,offering an automated,efficient approach to parameter estimation.The methodology significantly enhances predictive geostatistical models,contributing to the advancement of mineral exploration and improving the precision and speed of decision-making in the petroleum and mining industries. 展开更多
关键词 Variogram modelling Genetic algorithm(GA) Machine learning Gravity data Mineral exploration
在线阅读 下载PDF
上一页 1 2 30 下一页 到第
使用帮助 返回顶部