期刊文献+
共找到201篇文章
< 1 2 11 >
每页显示 20 50 100
Heart Disease Detection by Using Machine Learning Algorithms and a Real-Time Cardiovascular Health Monitoring System 被引量:1
1
作者 Shadman Nashif Md. Rakib Raihan +1 位作者 Md. Rasedul Islam Mohammad Hasan Imam 《World Journal of Engineering and Technology》 2018年第4期854-873,共20页
Cardiovascular diseases are the most common cause of death worldwide over the last few decades in the developed as well as underdeveloped and developing countries. Early detection of cardiac diseases and continuous su... Cardiovascular diseases are the most common cause of death worldwide over the last few decades in the developed as well as underdeveloped and developing countries. Early detection of cardiac diseases and continuous supervision of clinicians can reduce the mortality rate. However, accurate detection of heart diseases in all cases and consultation of a patient for 24 hours by a doctor is not available since it requires more sapience, time and expertise. In this?study, a tentative design of a cloud-based heart disease prediction system had been proposed to detect impending heart disease using Machine learning techniques. For the accurate detection of the heart disease, an efficient machine learning technique should be used which had been derived from a distinctive analysis among several machine learning algorithms in a Java Based Open Access Data Mining Platform, WEKA. The proposed algorithm was validated using two widely used open-access database, where 10-fold cross-validation is applied in order to analyze the performance of heart disease detection. An accuracy level of 97.53% accuracy was found from the SVM algorithm along with sensitivity and specificity of 97.50% and 94.94%respectively. Moreover, to monitor the heart disease patient round-the-clock by his/her caretaker/doctor, a real-time patient monitoring system was developed and presented using Arduino, capable of sensing some real-time parameters such as body temperature, blood pressure, humidity, heartbeat. The developed system can transmit the recorded data to a central server which are updated every 10 seconds. As a result, the doctors can visualize the patient’s real-time sensor data by using the application and start live video streaming if instant medication is required. Another important feature of the proposed system was that as soon as any real-time parameter of the patient exceeds the threshold, the prescribed doctor is notified at once through GSM technology. 展开更多
关键词 data mining machine learning IoT (Internet of Things) PATIENT Monitoring System HEART DISEASE DETECTION and prediction
暂未订购
Predicting Primary School Student Dropout Risk:A Machine Learning Framework for Early Intervention
2
作者 Samuel Ocen Musitapha Katalihwa Derrick Mwanje 《Journal of Intelligent Learning Systems and Applications》 2025年第4期267-279,共13页
Student dropout in primary education is a critical global challenge with significant long-term societal and individual consequences.Early identification of at-risk students is a crucial first step towards implementing... Student dropout in primary education is a critical global challenge with significant long-term societal and individual consequences.Early identification of at-risk students is a crucial first step towards implementing effective intervention strategies.This paper presents a machine learning framework for predicting student dropout risk by leveraging historical academic,attendance,and demographic data extracted from a primary school system.We formulate the problem as a binary classification task and evaluate multiple algorithms,including Logistic Regression,Random Forest,and Gradient Boosting,to identify the most effective predictor.To address the inherent class imbalance,we employ Synthetic Minority Over-sampling Technique(SMOTE).Our results,validated via stratified 5-fold cross-validation,indicate that the Random Forest model achieved the highest performance,with a recall of 0.91±0.03,ensuring that 91%of truly at-risk students were correctly identified.Furthermore,we use SHAP(SHapley Additive exPlanations)values to provide interpretable insights into the model’s predictions,revealing that attendance rate,academic performance trends,and socio-economic proxies are the most salient features.This work demonstrates the potential of machine learning as a powerful decision-support tool for educators,enabling timely and data-driven interventions to improve student retention and completion rates. 展开更多
关键词 Educational data mining machine learning Dropout prediction Early Warning System Primary Education Explainable AI(XAI)
在线阅读 下载PDF
Hot spot temperature prediction and operating parameter estimation of racks in data center using machine learning algorithms based on simulation data 被引量:3
3
作者 Xianzhong Chen Rang Tu +2 位作者 Ming Li Xu Yang Kun Jia 《Building Simulation》 SCIE EI CSCD 2023年第11期2159-2176,共18页
In this paper,models to predict hot spot temperature and to estimate cooling air’s working parameters of racks in data centers were established using machine learning algorithms based on simulation data.First,simulat... In this paper,models to predict hot spot temperature and to estimate cooling air’s working parameters of racks in data centers were established using machine learning algorithms based on simulation data.First,simulation models of typical racks were established in computational fluid dynamics(CFD).The model was validated with field test results and results in literature,error of which was less than 3%.Then,the CFD model was used to simulate thermal environments of a typical rack considering different factors,such as servers’power,which is from 3.3 kW to 20.1 kW,cooling air’s inlet velocity,which is from 1.0 m/s to 3.0 m/s,and cooling air’s inlet temperature,which is from 16℃ to 26℃ The highest temperature in the rack,also called hot spot temperature,was selected for each case.Next,a prediction model of hot spot temperature was built using machine learning algorithms,with servers’power,cooling air’s inlet velocity and cooling air’s inlet temperature as inputs,and the hot spot temperatures as outputs.Finally,based on the prediction model,an operating parameters estimation model was established to recommend cooling air’s inlet temperatures and velocities,which can not only keep the hot spot temperature at the safety value,but are also energy saving. 展开更多
关键词 data center CFD simulation hot spot temperature machine learning algorithm prediction and estimation models
原文传递
Chimp Optimization Algorithm Based Feature Selection with Machine Learning for Medical Data Classification
4
作者 Firas Abedi Hayder M.A.Ghanimi +6 位作者 Abeer D.Algarni Naglaa F.Soliman Walid El-Shafai Ali Hashim Abbas Zahraa H.Kareem Hussein Muhi Hariz Ahmed Alkhayyat 《Computer Systems Science & Engineering》 SCIE EI 2023年第12期2791-2814,共24页
Datamining plays a crucial role in extractingmeaningful knowledge fromlarge-scale data repositories,such as data warehouses and databases.Association rule mining,a fundamental process in data mining,involves discoveri... Datamining plays a crucial role in extractingmeaningful knowledge fromlarge-scale data repositories,such as data warehouses and databases.Association rule mining,a fundamental process in data mining,involves discovering correlations,patterns,and causal structures within datasets.In the healthcare domain,association rules offer valuable opportunities for building knowledge bases,enabling intelligent diagnoses,and extracting invaluable information rapidly.This paper presents a novel approach called the Machine Learning based Association Rule Mining and Classification for Healthcare Data Management System(MLARMC-HDMS).The MLARMC-HDMS technique integrates classification and association rule mining(ARM)processes.Initially,the chimp optimization algorithm-based feature selection(COAFS)technique is employed within MLARMC-HDMS to select relevant attributes.Inspired by the foraging behavior of chimpanzees,the COA algorithm mimics their search strategy for food.Subsequently,the classification process utilizes stochastic gradient descent with a multilayer perceptron(SGD-MLP)model,while the Apriori algorithm determines attribute relationships.We propose a COA-based feature selection approach for medical data classification using machine learning techniques.This approach involves selecting pertinent features from medical datasets through COA and training machine learning models using the reduced feature set.We evaluate the performance of our approach on various medical datasets employing diverse machine learning classifiers.Experimental results demonstrate that our proposed approach surpasses alternative feature selection methods,achieving higher accuracy and precision rates in medical data classification tasks.The study showcases the effectiveness and efficiency of the COA-based feature selection approach in identifying relevant features,thereby enhancing the diagnosis and treatment of various diseases.To provide further validation,we conduct detailed experiments on a benchmark medical dataset,revealing the superiority of the MLARMCHDMS model over other methods,with a maximum accuracy of 99.75%.Therefore,this research contributes to the advancement of feature selection techniques in medical data classification and highlights the potential for improving healthcare outcomes through accurate and efficient data analysis.The presented MLARMC-HDMS framework and COA-based feature selection approach offer valuable insights for researchers and practitioners working in the field of healthcare data mining and machine learning. 展开更多
关键词 Association rule mining data classification healthcare data machine learning parameter tuning data mining feature selection MLARMC-HDMS COA stochastic gradient descent Apriori algorithm
在线阅读 下载PDF
A New Framework for Scholarship Predictor Using a Machine Learning Approach
5
作者 Bushra Kanwal Rana Saud Shoukat +3 位作者 Saif Ur Rehman Mahwish Kundi Tahani AlSaedi Abdulrahman Alahmadi 《Intelligent Automation & Soft Computing》 2024年第5期829-854,共26页
Education is the base of the survival and growth of any state,but due to resource scarcity,students,particularly at the university level,are forced into a difficult situation.Scholarships are the most significant fina... Education is the base of the survival and growth of any state,but due to resource scarcity,students,particularly at the university level,are forced into a difficult situation.Scholarships are the most significant financial aid mechanisms developed to overcome such obstacles and assist the students in continuing with their higher studies.In this study,the convoluted situation of scholarship eligibility criteria,including parental income,responsibilities,and academic achievements,is addressed.In an attempt to maximize the scholarship selection process,numerous machine learning algorithms,including Support Vector Machines,Neural Networks,K-Nearest Neighbors,and the C4.5 algorithm,were applied.The C4.5 algorithm,owing to its efficiency in the prediction of scholarship beneficiaries based on extraneous factors,was capable of predicting a phenomenal 95.62%of predictions using extensive data of a well-esteemed government sector university from Pakistan.This percentage is 4%and 15%better than the remainder of the methods tested,and it depicts the extent of the potential for the technique to enhance the scholarship selection process.The Decision Support Systems(DSS)would not only save the administrative cost but would also create a fair and transparent process in place.In a world where accessibility to education is the key,this research provides data-oriented consolidation to ensure that deserving students are helped and allowed to get the financial assistance that they need to reach higher studies and bridge the gap between the demands of the day and the institutions of intellect. 展开更多
关键词 EDUCATION data mining C4.5 algorithm decision support system scholarship guarantee machine learning
在线阅读 下载PDF
Pathways to chronic disease detection and prediction: Mapping the potential of machine learning to the pathophysiological processes while navigating ethical challenges 被引量:1
6
作者 Ebenezer Afrifa-Yamoah Eric Adua +5 位作者 Emmanuel Peprah-Yamoah Enoch O.Anto Victor Opoku-Yamoah Emmanuel Acheampong Michael J.Macartney Rashid Hashmi 《Chronic Diseases and Translational Medicine》 2025年第1期1-21,共21页
Chronic diseases such as heart disease,cancer,and diabetes are leading drivers of mortality worldwide,underscoring the need for improved efforts around early detection and prediction.The pathophysiology and management... Chronic diseases such as heart disease,cancer,and diabetes are leading drivers of mortality worldwide,underscoring the need for improved efforts around early detection and prediction.The pathophysiology and management of chronic diseases have benefitted from emerging fields in molecular biology like genomics,transcriptomics,proteomics,glycomics,and lipidomics.The complex biomarker and mechanistic data from these"omics"studies present analytical and interpretive challenges,especially for traditional statistical methods.Machine learning(ML)techniques offer considerable promise in unlocking new pathways for data-driven chronic disease risk assessment and prognosis.This review provides a comprehensive overview of state-of-the-art applications of ML algorithms for chronic disease detection and prediction across datasets,including medical imaging,genomics,wearables,and electronic health records.Specifically,we review and synthesize key studies leveraging major ML approaches ranging from traditional techniques such as logistic regression and random forests to modern deep learning neural network architectures.We consolidate existing literature to date around ML for chronic disease prediction to synthesize major trends and trajectories that may inform both future research and clinical translation efforts in this growing field.While highlighting the critical innovations and successes emerging in this space,we identify the key challenges and limitations that remain to be addressed.Finally,we discuss pathways forward toward scalable,equitable,and clinically implementable ML solutions for transforming chronic disease screening and prevention. 展开更多
关键词 big data chronic diseases disease prediction machine learning algorithms OMICs data
原文传递
Prediction of Cold and Heat Patterns Using Anthropometric Measures Based on Machine Learning
7
作者 Bum Ju Lee Jae Chul Lee +1 位作者 Jiho Nam Jong Yeol Kim 《Chinese Journal of Integrative Medicine》 SCIE CAS CSCD 2018年第1期16-23,共8页
Objective: To examine the association of body shape with cold and heat patterns, to determine which anthropometric measure is the best indicator for discriminating between the two patterns, and to investigate whether... Objective: To examine the association of body shape with cold and heat patterns, to determine which anthropometric measure is the best indicator for discriminating between the two patterns, and to investigate whether using a combination of measures can improve the predictive power to diagnose these patterns. Methods: Based on a total of 4,859 subjects (3,000 women and 1,859 men), statistical analyses using binary logistic regression were performed to assess the significance of the difference and the predictive power of each anthropometric measure, and binary logistic regression and Naive Bayes with the variable selection technique were used to assess the improvement in the predictive power of the patterns using the combined measures. Results: In women, the strongest indicators for determining the cold and heat patterns among anthropometric measures were body mass index (BMI) and rib circumference; in men, the best indicator was BMI. In experiments using a combination of measures, the values of the area under the receiver operating characteristic curve in women were 0.776 by Naive Bayes and 0.772 by logistic regression, and the values in men were 0.788 by Naive Bayes and 0.779 by logistic regression. Conclusions: Individuals with a higher BMI have a tendency toward a heat pattern in both women and men. The use of a combination of anthropometric measures can slightly improve the diagnostic accuracy. Our findings can provide fundamental information for the diagnosis of cold and heat patterns based on body shape for personalized medicine. 展开更多
关键词 cold and heat patterns personalized medicine cold syndrome predictive power machine learning data mining
原文传递
A systematic review:Detecting phishing websites using data mining models
8
作者 Dina Jibat Sarah Jamjoom +1 位作者 Qasem Abu Al-Haija Abdallah Qusef 《Intelligent and Converged Networks》 EI 2023年第4期326-341,共16页
As internet technology use is on the rise globally,phishing constitutes a considerable share of the threats that may attack individuals and organizations,leading to significant losses from personal and confidential in... As internet technology use is on the rise globally,phishing constitutes a considerable share of the threats that may attack individuals and organizations,leading to significant losses from personal and confidential information to substantial financial losses.Thus,much research has been dedicated in recent years to developing effective and robust mechanisms to enhance the ability to trace illegitimate web pages and to distinguish them from non-phishing sites as accurately as possible.Aiming to conclude whether a universally accepted model can detect phishing attempts with 100%accuracy,we conduct a systematic review of research carried out in 2018-2021 published in well-known journals published by Elsevier,IEEE,Springer,and Emerald.Those researchers studied different Data Mining(DM)algorithms,some of which created a whole new model,while others compared the performance of several algorithms.Some studies combined two or more algorithms to enhance the detection performance.Results reveal that while most algorithms achieve accuracies higher than 90%,only some specific models can achieve 100%accurate results. 展开更多
关键词 PHISHING data mining machine learning ALGORITHM CLASSIFICATION
原文传递
A Novel Approach to Heart Failure Prediction and Classification through Advanced Deep Learning Model
9
作者 Abdalla Mahgoub 《World Journal of Cardiovascular Diseases》 2023年第9期586-604,共19页
In this study, the author will investigate and utilize advanced machine learning models related to two different methodologies to determine the best and most effective way to predict individuals with heart failure and... In this study, the author will investigate and utilize advanced machine learning models related to two different methodologies to determine the best and most effective way to predict individuals with heart failure and cardiovascular diseases. The first methodology involves a list of classification machine learning algorithms, and the second methodology involves the use of a deep learning algorithm known as MLP or Multilayer Perceptrons. Globally, hospitals are dealing with cases related to cardiovascular diseases and heart failure as they are major causes of death, not only for overweight individuals but also for those who do not adopt a healthy diet and lifestyle. Often, heart failures and cardiovascular diseases can be caused by many factors, including cardiomyopathy, high blood pressure, coronary heart disease, and heart inflammation [1]. Other factors, such as irregular shocks or stress, can also contribute to heart failure or a heart attack. While these events cannot be predicted, continuous data from patients’ health can help doctors predict heart failure. Therefore, this data-driven research utilizes advanced machine learning and deep learning techniques to better analyze and manipulate the data, providing doctors with informative decision-making tools regarding a person’s likelihood of experiencing heart failure. In this paper, the author employed advanced data preprocessing and cleaning techniques. Additionally, the dataset underwent testing using two different methodologies to determine the most effective machine-learning technique for producing optimal predictions. The first methodology involved employing a list of supervised classification machine learning algorithms, including Naïve Bayes (NB), KNN, logistic regression, and the SVM algorithm. The second methodology utilized a deep learning (DL) algorithm known as Multilayer Perceptrons (MLPs). This algorithm provided the author with the flexibility to experiment with different layer sizes and activation functions, such as ReLU, logistic (sigmoid), and Tanh. Both methodologies produced optimal models with high-level accuracy rates. The first methodology involves a list of supervised machine learning algorithms, including KNN, SVM, Adaboost, Logistic Regression, Naive Bayes, and Decision Tree algorithms. They achieved accuracy rates of 86%, 89%, 89%, 81%, 79%, and 99%, respectively. The author clearly explained that Decision Tree algorithm is not suitable for the dataset at hand due to overfitting issues. Therefore, it was discarded as an optimal model to be used. However, the latter methodology (Neural Network) demonstrated the most stable and optimal accuracy, achieving over 87% accuracy while adapting well to real-life situations and requiring low computing power overall. A performance assessment and evaluation were carried out based on a confusion matrix report to demonstrate feasibility and performance. The author concluded that the performance of the model in real-life situations can advance not only the medical field of science but also mathematical concepts. Additionally, the advanced preprocessing approach behind the model can provide value to the Data Science community. The model can be further developed by employing various optimization techniques to handle even larger datasets related to heart failures. Furthermore, different neural network algorithms can be tested to explore alternative approaches and yield different results. 展开更多
关键词 Heart Disease prediction Cardiovascular Disease machine learning algorithms Lazy Predict Multilayer Perceptrons (MLPs) data Science Techniques and Analysis Deep learning Activation Functions
在线阅读 下载PDF
Prediction of Lubricant Physicochemical Properties Based on Gaussian Copula Data Expansion 被引量:1
10
作者 Feng Xin Yang Rui +1 位作者 Xie Peiyuan Xia Yanqiu 《China Petroleum Processing & Petrochemical Technology》 SCIE CAS CSCD 2024年第1期161-174,共14页
The composition of base oils affects the performance of lubricants made from them.This paper proposes a hybrid model based on gradient-boosted decision tree(GBDT)to analyze the effect of different ratios of KN4010,PAO... The composition of base oils affects the performance of lubricants made from them.This paper proposes a hybrid model based on gradient-boosted decision tree(GBDT)to analyze the effect of different ratios of KN4010,PAO40,and PriEco3000 component in a composite base oil system on the performance of lubricants.The study was conducted under small laboratory sample conditions,and a data expansion method using the Gaussian Copula function was proposed to improve the prediction ability of the hybrid model.The study also compared four optimization algorithms,sticky mushroom algorithm(SMA),genetic algorithm(GA),whale optimization algorithm(WOA),and seagull optimization algorithm(SOA),to predict the kinematic viscosity at 40℃,kinematic viscosity at 100℃,viscosity index,and oxidation induction time performance of the lubricant.The results showed that the Gaussian Copula function data expansion method improved the prediction ability of the hybrid model in the case of small samples.The SOA-GBDT hybrid model had the fastest convergence speed for the samples and the best prediction effect,with determination coefficients(R^(2))for the four indicators of lubricants reaching 0.98,0.99,0.96 and 0.96,respectively.Thus,this model can significantly reduce the model’s prediction error and has good prediction ability. 展开更多
关键词 base oil data augmentation machine learning performance prediction seagull algorithm
在线阅读 下载PDF
The Study on China’s Flu Prediction Model Based on Web Search Data 被引量:2
11
作者 Yan Bu Jinhong Bai +2 位作者 Zhuo Chen Mingjing Guo Fan Yang 《Journal of Data Analysis and Information Processing》 2018年第3期79-92,共14页
Influenza is a kind of infectious disease, which spreads quickly and widely. The outbreak of influenza has brought huge losses to society. In this paper, four major categories of flu keywords, “prevention phase”, “... Influenza is a kind of infectious disease, which spreads quickly and widely. The outbreak of influenza has brought huge losses to society. In this paper, four major categories of flu keywords, “prevention phase”, “symptom phase”, “treatment phase”, and “commonly-used phrase” were set. Python web crawler was used to obtain relevant influenza data from the National Influenza Center’s influenza surveillance weekly report and Baidu Index. The establishment of support vector regression (SVR), least absolute shrinkage and selection operator (LASSO), convolutional neural networks (CNN) prediction models through machine learning, took into account the seasonal characteristics of the influenza, also established the time series model (ARMA). The results show that, it is feasible to predict influenza based on web search data. Machine learning shows a certain forecast effect in the prediction of influenza based on web search data. In the future, it will have certain reference value in influenza prediction. The ARMA(3,0) model predicts better results and has greater generalization. Finally, the lack of research in this paper and future research directions are given. 展开更多
关键词 data mining Web SEARCH machine learning BAIDU Index INFLUENZA prediction
暂未订购
A Novel Cluster Analysis-Based Crop Dataset Recommendation Method in Precision Farming
12
作者 K.R.Naveen Kumar Husam Lahza +4 位作者 B.R.Sreenivasa Tawfeeq Shawly Ahmed A.Alsheikhy H.Arunkumar C.R.Nirmala 《Computer Systems Science & Engineering》 SCIE EI 2023年第9期3239-3260,共22页
Data mining and analytics involve inspecting and modeling large pre-existing datasets to discover decision-making information.Precision agriculture uses datamining to advance agricultural developments.Many farmers are... Data mining and analytics involve inspecting and modeling large pre-existing datasets to discover decision-making information.Precision agriculture uses datamining to advance agricultural developments.Many farmers aren’t getting the most out of their land because they don’t use precision agriculture.They harvest crops without a well-planned recommendation system.Future crop production is calculated by combining environmental conditions and management behavior,yielding numerical and categorical data.Most existing research still needs to address data preprocessing and crop categorization/classification.Furthermore,statistical analysis receives less attention,despite producing more accurate and valid results.The study was conducted on a dataset about Karnataka state,India,with crops of eight parameters taken into account,namely the minimum amount of fertilizers required,such as nitrogen,phosphorus,potassium,and pH values.The research considers rainfall,season,soil type,and temperature parameters to provide precise cultivation recommendations for high productivity.The presented algorithm converts discrete numerals to factors first,then reduces levels.Second,the algorithm generates six datasets,two fromCase-1(dataset withmany numeric variables),two from Case-2(dataset with many categorical variables),and one from Case-3(dataset with reduced factor variables).Finally,the algorithm outputs a class membership allocation based on an extended version of the K-means partitioning method with lambda estimation.The presented work produces mixed-type datasets with precisely categorized crops by organizing data based on environmental conditions,soil nutrients,and geo-location.Finally,the prepared dataset solves the classification problem,leading to a model evaluation that selects the best dataset for precise crop prediction. 展开更多
关键词 data mining crop prediction k-prototypes K-MEANS CLUSTER machine learning
在线阅读 下载PDF
ediction Model for a Good Learning Environment Using an Ensemble Approach
13
作者 S.Subha S.Baghavathi Priya 《Computer Systems Science & Engineering》 SCIE EI 2023年第3期2081-2093,共13页
This paper presents an efficient prediction model for a good learning environment using Random Forest(RF)classifier.It consists of a series of modules;data preprocessing,data normalization,data split andfinally classi... This paper presents an efficient prediction model for a good learning environment using Random Forest(RF)classifier.It consists of a series of modules;data preprocessing,data normalization,data split andfinally classification or prediction by the RF classifier.The preprocessed data is normalized using minmax normalization often used before modelfitting.As the input data or variables are measured at different scales,it is necessary to normalize them to contribute equally to the modelfitting.Then,the RF classifier is employed for course selection which is an ensemble learning method and k-fold cross-validation(k=10)is used to validate the model.The proposed Prediction Model for Course Selection(PMCS)system is considered a multi-class problem that predicts the course for a particular learner with three complexity levels,namely low,medium and high.It is operated under two modes;locally and globally.The former considers the gender of the learner and the later does not consider the gender of the learner.The database comprises the learner opinions from 75 males and 75 females per category(low,medium and high).Thus the system uses a total of 450 samples to evaluate the performance of the PMCS system.Results show that the system’s performance,while using locally i.e.,gender-wise has slightly higher performance than the global system.The RF classifier with 75 decision trees in the global system provides an average accuracy of 97.6%,whereas in the local system it is 97%(male)and 97.6%(female).The overall performance of the RF classifier with 75 trees is better than 25,50 and 100 decision trees in both local and global systems. 展开更多
关键词 machine learning ensemble learning random forest data mining prediction system
在线阅读 下载PDF
DiagData: A Tool for Generation of Fuzzy Inference System
14
作者 Silvia Maria Fonseca Silveira Massruha Raphael Fuini Riccioti Helano Povoas Lima Carlos Alberto AlvesMeira 《Journal of Environmental Science and Engineering(B)》 2012年第3期336-343,共8页
In this paper, it described the architecture of a tool called DiagData. This tool aims to use a large amount of data and information in the field of plant disease diagnostic to generate a disease predictive system. In... In this paper, it described the architecture of a tool called DiagData. This tool aims to use a large amount of data and information in the field of plant disease diagnostic to generate a disease predictive system. In this approach, techniques of data mining are used to extract knowledge from existing data. The data is extracted in the form of rules that are used in the development of a predictive intelligent system. Currently, the specification of these rules is built by an expert or data mining. When data mining on a large database is used, the number of generated rules is very complex too. The main goal of this work is minimize the rule generation time. The proposed tool, called DiagData, extracts knowledge automatically or semi-automatically from a database and uses it to build an intelligent system for disease prediction. In this work, the decision tree learning algorithm was used to generate the rules. A toolbox called Fuzzygen was used to generate a prediction system from rules generated by decision tree algorithm. The language used to implement this software was Java. The DiagData has been used in diseases prediction and diagnosis systems and in the validation of economic and environmental indicators in agricultural production systems. The validation process involved measurements and comparisons of the time spent to enter the rules by an expert with the time used to insert the same rules with the proposed tool. Thus, the tool was successfully validated, providing a reduction of time. 展开更多
关键词 prediction modelling data mining decision tree machine learning fuzzy inference system fuzzygen.
在线阅读 下载PDF
Particle Swarm Optimization: Advances, Applications, and Experimental Insights
15
作者 Laith Abualigah 《Computers, Materials & Continua》 2025年第2期1539-1592,共54页
Particle Swarm Optimization(PSO)has been utilized as a useful tool for solving intricate optimization problems for various applications in different fields.This paper attempts to carry out an update on PSO and gives a... Particle Swarm Optimization(PSO)has been utilized as a useful tool for solving intricate optimization problems for various applications in different fields.This paper attempts to carry out an update on PSO and gives a review of its recent developments and applications,but also provides arguments for its efficacy in resolving optimization problems in comparison with other algorithms.Covering six strategic areas,which include Data Mining,Machine Learning,Engineering Design,Energy Systems,Healthcare,and Robotics,the study demonstrates the versatility and effectiveness of the PSO.Experimental results are,however,used to show the strong and weak parts of PSO,and performance results are included in tables for ease of comparison.The results stress PSO’s efficiency in providing optimal solutions but also show that there are aspects that need to be improved through combination with algorithms or tuning to the parameters of the method.The review of the advantages and limitations of PSO is intended to provide academics and practitioners with a well-rounded view of the methods of employing such a tool most effectively and to encourage optimized designs of PSO in solving theoretical and practical problems in the future. 展开更多
关键词 Particle swarm optimization(PSO) optimization algorithms data mining machine learning engineer-ing design energy systems healthcare applications ROBOTICS comparative analysis algorithm performance evaluation
在线阅读 下载PDF
An improved deep learning model for soybean future price prediction with hybrid data preprocessing strategy
16
作者 Dingya CHEN Hui LIU +1 位作者 Yanfei LI Zhu DUAN 《Frontiers of Agricultural Science and Engineering》 2025年第2期208-230,共23页
The futures trading market is an important part of the financial markets and soybeans are one of the most strategically important crops in the world.How to predict soybean future price is a challenging topic being stu... The futures trading market is an important part of the financial markets and soybeans are one of the most strategically important crops in the world.How to predict soybean future price is a challenging topic being studied by many researchers.This paper proposes a novel hybrid soybean future price prediction model which includes two stages of data preprocessing and deep learning prediction.In the data preprocessing stage,futures price series are decomposed into subsequences using the ICEEMDAN(improved complete ensemble empirical mode decomposition with adaptive noise)method.The Lempel-Ziv complexity determination method was then used to identify and reconstruct high-frequency subsequences.Finally,the high frequency component is decomposed secondarily using variational mode decomposition optimized by beluga whale optimization algorithm.In the deep learning prediction stage,a deep extreme learning machine optimized by the sparrow search algorithm was used to obtain the prediction results of all subseries and reconstructs them to obtain the final soybean future price prediction results.Based on the experimental results of soybean future price markets in China,Italy,and the United States,it was found that the hybrid method proposed provides superior performance in terms of prediction accuracy and robustness. 展开更多
关键词 Deep extreme learning machine hybrid data preprocessing optimization algorithm soybean future price prediction
原文传递
基于机器学习的煤系地层TBM掘进巷道围岩强度预测 被引量:3
17
作者 丁自伟 高成登 +6 位作者 景博宇 黄兴 刘滨 胡阳 桑昊旻 徐彬 秦立学 《西安科技大学学报》 北大核心 2025年第1期49-60,共12页
为研究全断面掘进机(TBM)掘进参数与煤系地层岩体力学参数之间的互馈关系,准确、实时预测巷道围岩强度特征,基于TBM掘进过程中的现场监测,通过岩-机互馈关系分析,确定模型的输入特征参数,并建立了对应的数据库;将梯度提升决策树(GBDT)... 为研究全断面掘进机(TBM)掘进参数与煤系地层岩体力学参数之间的互馈关系,准确、实时预测巷道围岩强度特征,基于TBM掘进过程中的现场监测,通过岩-机互馈关系分析,确定模型的输入特征参数,并建立了对应的数据库;将梯度提升决策树(GBDT)、随机森林(RF)、支持向量回归(SVR)3种机器学习算法作为基学习器,线性回归(LR)算法作为元学习器,提出了一种基于Stacking集成算法的预测模型,并对比分析了Stacking集成算法与单一机器学习算法模型的预测性能。结果表明:二值判别与箱线图可有效对原始数据进行预处理;模型的主要输入特征参数为刀盘推力F、刀盘扭矩T、贯入度FPI、刀盘转速RPM、刀盘振动加速度A;Stacking模型在测试集上的拟合优度可达0.976,而均方误差、平均绝对误差、平均绝对百分误差分别仅有0.031,0.148和0.092,与其他3种模型相比,其拟合优度最高,误差指标数值最小,集成模型具有更高的预测精度,能够有效地预测煤矿TBM掘进巷道围岩点荷载强度。研究验证了Stacking模型的准确性,可为煤矿TBM掘进参数控制和巷道支护参数调整提供科学的参考依据。 展开更多
关键词 煤矿全断面掘进机 TBM掘进参数 Stacking集成算法 数据预处理 围岩强度预测
在线阅读 下载PDF
基于数据挖掘技术的稀土磁性材料研究进展
18
作者 刘丹 李渊 +2 位作者 孙若瑄 漆星 沈保根 《物理学报》 北大核心 2025年第13期81-98,共18页
稀土元素的原子结构特殊,具有内层未成对4f轨道电子多、原子磁矩高、自旋轨道耦合作用强的性质,故其电子能级极为丰富,易形成多种价态、多种配位的化合物,通常表现出特殊的磁学性质和丰富的磁畴结构,成为高新技术产业发展的关键材料.这... 稀土元素的原子结构特殊,具有内层未成对4f轨道电子多、原子磁矩高、自旋轨道耦合作用强的性质,故其电子能级极为丰富,易形成多种价态、多种配位的化合物,通常表现出特殊的磁学性质和丰富的磁畴结构,成为高新技术产业发展的关键材料.这类材料中复杂的磁结构形式、多样的磁耦合类型及多种直接或间接的磁交换作用,为开发新型功能器件提供便利的同时,也对基础研究提出了严峻挑战.随着数据挖掘技术的快速发展,大数据和人工智能的出现给研究人员提供了一个新的选择,可以高效地分析大量实验和计算数据,从而加速稀土磁性材料的研究与开发.本文围绕稀土永磁材料、稀土磁致冷材料、稀土磁致伸缩材料等,详细阐述了数据挖掘技术在其性能预测、成分与工艺优化、微观结构分析等方面的应用进展,深入探讨了当前面临的挑战,并对未来发展趋势进行展望,为推动数据挖掘技术与稀土磁性材料研究的深度融合提供理论基础. 展开更多
关键词 数据挖掘 稀土磁性材料 机器学习 性能预测
在线阅读 下载PDF
基于数据增强和优化DHKELM的短期光伏功率预测
19
作者 郭利进 马粽阳 胡晓岩 《太阳能学报》 北大核心 2025年第8期463-471,共9页
针对不同气象条件数据质量差异较大且光伏功率呈高波动性难以预测等问题,提出添加随机噪声的数据增强方法(DA)和改进的神经网络组合模型。首先利用谱聚类算法将光伏数据按不同气象条件进行分类,随后通过添加与输入同形状的随机噪声方法... 针对不同气象条件数据质量差异较大且光伏功率呈高波动性难以预测等问题,提出添加随机噪声的数据增强方法(DA)和改进的神经网络组合模型。首先利用谱聚类算法将光伏数据按不同气象条件进行分类,随后通过添加与输入同形状的随机噪声方法提升数据集的规模与质量。针对深度混合核极限学习机(DHKELM)超参数多等问题,提出融合佳点集初始化、黄金正弦更新策略、非线性扰动和最优个体自适应扰动的改进鹈鹕优化算法(IPOA)对其超参数寻优。最后以青海共和县光伏园内某电站数据为例,结果表明基于数据增强的改进鹈鹕算法优化深度混合核极限学习机(DA-IPOA-DHKELM)模型在不同天气、季节条件下预测误差最小,拟合度均能达到90%以上,改进模型预测精度高、算法适用性强。 展开更多
关键词 光伏功率 预测 聚类分析 数据增强 深度混合核极限学习机 改进算法
原文传递
一种基于AI大数据的用户满意度预测算法
20
作者 周文红 孙军亮 +2 位作者 许妍青 杨嘉忱 孙凡晰 《移动信息》 2025年第2期199-201,共3页
随着大数据技术的发展和人工智能应用的广泛化,用户满意度预测已成为运营商解决信号差和网络质量问题,减少用户投诉并获取竞争优势的关键手段.文中提出了一种结合机器学习技术的用户满意度预测算法,通过分析用户行为数据和反馈,构建了... 随着大数据技术的发展和人工智能应用的广泛化,用户满意度预测已成为运营商解决信号差和网络质量问题,减少用户投诉并获取竞争优势的关键手段.文中提出了一种结合机器学习技术的用户满意度预测算法,通过分析用户行为数据和反馈,构建了一个高精度的预测模型.通过实验验证,该模型在多个数据集上表现出优越的预测准确性和良好的泛化能力.该算法的实现,对于理解用户需求和改进服务质量具有重要意义. 展开更多
关键词 大数据 用户满意度 预测算法 机器学习
在线阅读 下载PDF
上一页 1 2 11 下一页 到第
使用帮助 返回顶部