期刊文献+
共找到12,325篇文章
< 1 2 250 >
每页显示 20 50 100
Multi-label learning algorithm with SVM based association 被引量:4
1
作者 Feng Pan Qin Danyang +3 位作者 Ji Ping Ma Jingya Zhang Yan Yang Songxiang 《High Technology Letters》 EI CAS 2019年第1期97-104,共8页
Multi-label learning is an active research area which plays an important role in machine learning. Traditional learning algorithms, however, have to depend on samples with complete labels. The existing learning algori... Multi-label learning is an active research area which plays an important role in machine learning. Traditional learning algorithms, however, have to depend on samples with complete labels. The existing learning algorithms with missing labels do not consider the relevance of labels, resulting in label estimation errors of new samples. A new multi-label learning algorithm with support vector machine(SVM) based association(SVMA) is proposed to estimate missing labels by constructing the association between different labels. SVMA will establish a mapping function to minimize the number of samples in the margin while ensuring the margin large enough as well as minimizing the misclassification probability. To evaluate the performance of SVMA in the condition of missing labels, four typical data sets are adopted with the integrity of the labels being handled manually. Simulation results show the superiority of SVMA in dealing with the samples with missing labels compared with other models in image classification. 展开更多
关键词 multi-label learning missing labels ASSOCIATION support vector machine(SVM)
在线阅读 下载PDF
Multi-Label Chinese Comments Categorization: Comparison of Multi-Label Learning Algorithms 被引量:4
2
作者 Jiahui He Chaozhi Wang +2 位作者 Hongyu Wu Leiming Yan Christian Lu 《Journal of New Media》 2019年第2期51-61,共11页
Multi-label text categorization refers to the problem of categorizing text througha multi-label learning algorithm. Text classification for Asian languages such as Chinese isdifferent from work for other languages suc... Multi-label text categorization refers to the problem of categorizing text througha multi-label learning algorithm. Text classification for Asian languages such as Chinese isdifferent from work for other languages such as English which use spaces to separate words.Before classifying text, it is necessary to perform a word segmentation operation to converta continuous language into a list of separate words and then convert it into a vector of acertain dimension. Generally, multi-label learning algorithms can be divided into twocategories, problem transformation methods and adapted algorithms. This work will usecustomer's comments about some hotels as a training data set, which contains labels for allaspects of the hotel evaluation, aiming to analyze and compare the performance of variousmulti-label learning algorithms on Chinese text classification. The experiment involves threebasic methods of problem transformation methods: Support Vector Machine, Random Forest,k-Nearest-Neighbor;and one adapted algorithm of Convolutional Neural Network. Theexperimental results show that the Support Vector Machine has better performance. 展开更多
关键词 multi-label classification Chinese text classification problem transformation adapted algorithms
在线阅读 下载PDF
Novel Apriori-Based Multi-Label Learning Algorithm by Exploiting Coupled Label Relationship 被引量:1
3
作者 Zhenwu Wang Longbing Cao 《Journal of Beijing Institute of Technology》 EI CAS 2017年第2期206-214,共9页
It is a key challenge to exploit the label coupling relationship in multi-label classification(MLC)problems.Most previous work focused on label pairwise relations,in which generally only global statistical informati... It is a key challenge to exploit the label coupling relationship in multi-label classification(MLC)problems.Most previous work focused on label pairwise relations,in which generally only global statistical information is used to analyze the coupled label relationship.In this work,firstly Bayesian and hypothesis testing methods are applied to predict the label set size of testing samples within their k nearest neighbor samples,which combines global and local statistical information,and then apriori algorithm is used to mine the label coupling relationship among multiple labels rather than pairwise labels,which can exploit the label coupling relations more accurately and comprehensively.The experimental results on text,biology and audio datasets shown that,compared with the state-of-the-art algorithm,the proposed algorithm can obtain better performance on 5 common criteria. 展开更多
关键词 multi-label classification hypothesis testing k nearest neighbor apriori algorithm label coupling
在线阅读 下载PDF
A Unified Feature Selection Framework Combining Mutual Information and Regression Optimization for Multi-Label Learning
4
作者 Hyunki Lim 《Computers, Materials & Continua》 2026年第4期1262-1281,共20页
High-dimensional data causes difficulties in machine learning due to high time consumption and large memory requirements.In particular,in amulti-label environment,higher complexity is required asmuch as the number of ... High-dimensional data causes difficulties in machine learning due to high time consumption and large memory requirements.In particular,in amulti-label environment,higher complexity is required asmuch as the number of labels.Moreover,an optimization problem that fully considers all dependencies between features and labels is difficult to solve.In this study,we propose a novel regression-basedmulti-label feature selectionmethod that integrates mutual information to better exploit the underlying data structure.By incorporating mutual information into the regression formulation,the model captures not only linear relationships but also complex non-linear dependencies.The proposed objective function simultaneously considers three types of relationships:(1)feature redundancy,(2)featurelabel relevance,and(3)inter-label dependency.These three quantities are computed usingmutual information,allowing the proposed formulation to capture nonlinear dependencies among variables.These three types of relationships are key factors in multi-label feature selection,and our method expresses them within a unified formulation,enabling efficient optimization while simultaneously accounting for all of them.To efficiently solve the proposed optimization problem under non-negativity constraints,we develop a gradient-based optimization algorithm with fast convergence.Theexperimental results on sevenmulti-label datasets show that the proposed method outperforms existingmulti-label feature selection techniques. 展开更多
关键词 feature selection multi-label learning regression model optimization mutual information
在线阅读 下载PDF
Multi-label fundus disease classification using dual-branch deep learning:an intelligent diagnosis framework inspired by traditional Chinese medicine Five Wheels theory
5
作者 Xin He Xiaohui Li +5 位作者 Jun Peng Lei Sun Dan Shu Li Xiao Qinghua Peng Xiaoxia Xiao 《Digital Chinese Medicine》 2026年第1期80-90,共11页
Objective To develop a dual-branch deep learning framework for accurate multi-label classification of fundus diseases,addressing the key limitations of insufficient complementary feature extraction and inadequate cros... Objective To develop a dual-branch deep learning framework for accurate multi-label classification of fundus diseases,addressing the key limitations of insufficient complementary feature extraction and inadequate cross-modal feature fusion in existing automated diagnostic methods.Methods The fundus multi-label classification dataset with 12 disease categories(FMLC-12)dataset was constructed by integrating complementary samples from Ocular Disease Intelligent Recognition(ODIR)and Retinal Fundus Multi-Disease Image Dataset(RFMiD),yielding 6936 fundus images across 12 retinal pathology categories,and the framework was validated on both FMLC-12 and ODIR.Inspired by the holistic multi-regional assessment principle of the Five Wheels theory in traditional Chinese medicine(TCM)ophthalmology,the dualbranch multi-label network(DBMNet)was developed as a novel framework integrating complementary visual feature extraction with pathological correlation modeling.The architecture employed a TransNeXt backbone within a dual-branch design:one branch processed redgreen-blue(RGB)images to capture color-dependent features,such as vascular patterns and lesion morphology,while the other processed grayscale-converted images to enhance subtle textural details and contrast variations.A feature interaction module(FIM)effectively integrated the multi-scale features from both branches.Comprehensive ablation studies were conducted to evaluate the contributions of the dual-branch architecture and the FIM.The performance of DBMNet was compared against four state-of-the-art methods,including EfficientNet Ensemble,transfer learning-based convolutional neural network(CNN),BFENet,and EyeDeep-Net,using mean average precision(mAP),F1-score,and Cohen's kappa coefficient.Results The dual-branch architecture improved mAP by 15.44 percentage points over the single-branch TransNeXt baseline,increasing from 34.41%to 44.24%,and the addition of FIM further boosted mAP to 49.85%.On FMLC-12,DBMNet achieved an mAP of 49.85%,a Cohen’s kappa coefficient of 62.14%,and an F1-score of 70.21%.Compared with BFENet(mAP:45.42%,kappa:46.64%,F1-score:71.34%),DBMNet outperformed it by 4.43 percentage points in mAP and 15.50 percentage points in kappa,while BFENet achieved a marginally higher F1-score.On ODIR,DBMNet achieved an F1-score of 85.50%,comparable to state-of-the-art methods.Conclusion DBMNet effectively integrates RGB and grayscale visual modalities through a dual-branch architecture,significantly improving multi-label fundus disease classification.The framework not only addresses the issue of insufficient feature fusion in existing methods but also demonstrates outstanding performance in balancing detection across both common and rare diseases,providing a promising and clinically applicable pathway for standardized,intelligent fundus disease classification. 展开更多
关键词 multi-label classification Fundus images Deep learning Dual-branch network Traditional Chinese medicine ophthalmology Five Wheels theory
在线阅读 下载PDF
Empirical tropospheric zenith wet delay models with strong generalization capability based on a robust machine learning fusion algorithm
6
作者 Jiahao Zhang Qin Liang Yunqing Huang 《Geodesy and Geodynamics》 2026年第2期211-224,共14页
Tropospheric zenith wet delay(ZWD)plays a vital role in the analysis of space geodetic observations.In recent years,machine learning methods have been increasingly applied to improve the accuracy of ZWD calculations.H... Tropospheric zenith wet delay(ZWD)plays a vital role in the analysis of space geodetic observations.In recent years,machine learning methods have been increasingly applied to improve the accuracy of ZWD calculations.However,a single machine learning model has limited generalization capabilities.To address these limitations,this study introduces a novel machine learning fusion(MLF)algorithm with stronger generalization capabilities to enhance ZWD modeling and prediction accuracy.The MLF algorithm utilizes a two-layer structure integrating extra trees(ET),backpropagation neural network(BPNN),and linear regression models.By comparing the root mean square error(RMSE)of these models,we found that both ET-based and MLF-based models outperform RF-based and BPNN-based models in terms of internal and external accuracy,across both surface meteorological data-based and blind models.The improvement in exte rnal accuracy is particularly significant in the blind models.Our re sults show that the MLF(with an RMSE of 3.93 cm)and ET(3.99 cm)models outperform the traditional GPT3model(4.07 cm),while the RF(4.21 cm)and BPNN(4.14 cm)have worse external accuracies than the GPT3 model.It is worth noting that the BPNN suffered from overfitting during external accuracy tests,which was avoided by the MLF.In summary,regardless of the availability of surface meteorological data,the MLF-based empirical models demonstrate superior internal and external accuracy compared to the other tested models in this study. 展开更多
关键词 Tropospheric zenith wet delay Machine learning Extra trees Machine learning fusion algorithm Empirical models
原文传递
Research Review of Deep Learning Algorithms for Agricultural Disease Image Classification
7
作者 Shengjiu JIANG Qian WANG 《Plant Diseases and Pests》 2026年第1期30-34,共5页
In the context of rural revitalization and the development of smart agriculture, image classification technology based on deep learning has emerged as a crucial tool for digital monitoring and intelligent prevention a... In the context of rural revitalization and the development of smart agriculture, image classification technology based on deep learning has emerged as a crucial tool for digital monitoring and intelligent prevention and control of agricultural diseases. This paper provides a systematic review of the evolutionary development of algorithms within this field. Addressing challenges such as domain drift and limited global awareness in classical convolutional neural networks (CNNs) applied to complex agricultural environments, the paper focuses on the latest advancements in vision transformers (ViT) and their hybrid architectures to enhance cross-domain robustness and fine-grained recognition capabilities. In response to the challenges posed by scarce long-tail data and limited edge computing power in real-world scenarios, the paper explores solutions related to few-shot learning and ultra-lightweight network deployment. Finally, a forward-looking analysis is presented on the application paradigms of multimodal feature fusion, vision-based large models, and explainable artificial intelligence (AI) within smart plant protection. This analysis aims to offer theoretical insights for the development of efficient and transparent intelligent diagnostic systems for agricultural diseases, thereby supporting the advancement of digital agriculture and the construction of a robust agricultural nation. 展开更多
关键词 Agricultural disease image Classification algorithm Deep learning Research Review
在线阅读 下载PDF
Flood predictions from metrics to classes by multiple machine learning algorithms coupling with clustering-deduced membership degree
8
作者 ZHAI Xiaoyan ZHANG Yongyong +5 位作者 XIA Jun ZHANG Yongqiang TANG Qiuhong SHAO Quanxi CHEN Junxu ZHANG Fan 《Journal of Geographical Sciences》 2026年第1期149-176,共28页
Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting... Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach. 展开更多
关键词 flood regime metrics class prediction machine learning algorithms hydrological model
原文传递
Classifying Syndromes in Chinese Medicine Using Multi-label Learning Algorithm with Relevant Features for Each Label 被引量:4
9
作者 徐璡 许朝霞 +5 位作者 陆萍 郭睿 燕海霞 许文杰 王忆勤 夏春明 《Chinese Journal of Integrative Medicine》 SCIE CAS CSCD 2016年第11期867-871,共5页
Objective: To develop an effective Chinese Medicine(CM) diagnostic model of coronary heart disease(CHD) and to confirm the scientific validity of CM theoretical basis from an algorithmic viewpoint. Methods: Four types... Objective: To develop an effective Chinese Medicine(CM) diagnostic model of coronary heart disease(CHD) and to confirm the scientific validity of CM theoretical basis from an algorithmic viewpoint. Methods: Four types of objective diagnostic data were collected from 835 CHD patients by using a selfdeveloped CM inquiry scale for the diagnosis of heart problems, a tongue diagnosis instrument, a ZBOX-I pulse digital collection instrument, and the sound of an attending acquisition system. These diagnostic data was analyzed and a CM diagnostic model was established using a multi-label learning algorithm(REAL). Results: REAL was employed to establish a Xin(Heart) qi deficiency, Xin yang deficiency, Xin yin deficiency, blood stasis, and phlegm five-card CM diagnostic model, which had recognition rates of 80.32%, 89.77%, 84.93%, 85.37%, and 69.90%, respectively. Conclusions: The multi-label learning method established using four diagnostic models based on mutual information feature selection yielded good recognition results. The characteristic model parameters were selected by maximizing the mutual information for each card type. The four diagnostic methods used to obtain information in CM, i.e., observation, auscultation and olfaction, inquiry, and pulse diagnosis, can be characterized by these parameters, which is consistent with CM theory. 展开更多
关键词 Chinese medicine syndrome differentiation multi-label learning algorithm
原文传递
基于Q-Learning长尾延迟优化的SSD-SMR写缓存策略研究
10
作者 刘健 章步镐 +4 位作者 方匡弛 刘宣锋 孙国道 梁荣华 梁浩然 《计算机工程》 北大核心 2026年第3期287-298,共12页
随着全球数据规模的不断增大,如何以低成本的方式有效提升数据的访问性能是存储系统面临的一项重要挑战,使用低延迟、高带宽的固态硬盘(SSD)和低成本、高存储密度的叠瓦式磁盘(SMR)来构建缓存系统,成为一种有效的解决方案。但是,SMR固... 随着全球数据规模的不断增大,如何以低成本的方式有效提升数据的访问性能是存储系统面临的一项重要挑战,使用低延迟、高带宽的固态硬盘(SSD)和低成本、高存储密度的叠瓦式磁盘(SMR)来构建缓存系统,成为一种有效的解决方案。但是,SMR固有的机械运动和多磁道堆叠的特性导致其写性能较差,SSD中的脏数据频繁写回SMR所导致的大量读-合并-写(RMW)操作可能会引起严重的长尾延迟现象。为此,基于SSD-SMR混合存储架构提出一种结合强化学习Q-Learning算法的缓存替换优化策略。通过学习SMR设备的I/O负载状况与延迟之间的经验知识来控制对SMR的写入,当SMR负载较大时,通过控制缓存中脏数据的逐出来减少SMR因写回而产生的大量RMW操作,从而优化系统在不同负载下的尾部延迟开销。将Q-Learning算法与基于数据流行度的缓存算法LRU以及SMR感知的缓存算法SAC进行结合,使用真实企业Trace和YCSB生成的模拟Trace进行测试,实验结果表明,所提方法能够有效提升现有缓存算法的性能,可以降低57.06%的平均延迟和87.49%的尾部延迟。 展开更多
关键词 Q-learning算法 I/O负载 长尾延迟 缓存替换算法 混合存储
在线阅读 下载PDF
基于随机森林与Q-learning融合的多元电力数据存储优化决策方法
11
作者 叶学顺 贾东梨 +2 位作者 周俊 唐英 贾梓豪 《科学技术与工程》 北大核心 2026年第3期1065-1074,共10页
大规模和多样的电力数据存储面临效率低和内存容量不足的瓶颈问题。数据索引和数据压缩等传统数据存储优化方法各有优劣势,如何有效应用于电力数据存储是目前研究的难点。为了解决这个问题,提出了一种融合随机森林和Q-learning的多元电... 大规模和多样的电力数据存储面临效率低和内存容量不足的瓶颈问题。数据索引和数据压缩等传统数据存储优化方法各有优劣势,如何有效应用于电力数据存储是目前研究的难点。为了解决这个问题,提出了一种融合随机森林和Q-learning的多元电力数据存储优化决策方法。该方法中的关键技术包括:首先提出了基于改进随机森林算法的存储优化策略决策模型,引入信息增益方法,综合评价数据存储时对数据库的数据访问频率、查询时间、存储速度以及数据冗余率等因素影响,做出数据直接存储、数据索引存储和数据压缩存储的存储优化方法策略决策;其次提出了基于改进Q-learning算法的数据存储算法决策模型,引入多尺度学习机制、优先经验放回机制和正负向奖励机制,决策数据索引存储时适用的索引算法以及数据压缩存储时适用的数据压缩算法。本方法有效融合了数据索引与数据压缩的技术优势,大幅提升数据存储效率并节约存储空间,为大规模多元电力数据管理提供新的解决方案。 展开更多
关键词 随机森林算法 Q-learning算法 数据存储优化方法 数据索引算法 数据压缩算法
在线阅读 下载PDF
Study on Machine Learning-based Prediction of Compressive Strength of Concrete with Different Waste Glass Powder Contents
12
作者 YU Daidong MA Yuwei +3 位作者 LI Gang WANG Aiqin HUANG Wei WANG Jingchao 《材料导报》 北大核心 2026年第6期111-125,共15页
The application and promotion of waste glass powder concrete(WGPC)cansignificantly alleviate the pressure of concrete material scarcity and environmental pollution.Compressive strength(CS)is a critical parameter for e... The application and promotion of waste glass powder concrete(WGPC)cansignificantly alleviate the pressure of concrete material scarcity and environmental pollution.Compressive strength(CS)is a critical parameter for evaluating the efficacy of WGPC.Unlike conventional testing methods,machine learning techniques offer precise and reliable predictions of concrete’s compressive strength,especially in its long-term mechanical properties.In this work,four models,namely Multiple Linear Regression(MLR),Back Propagation Neural Network(BPNN),Support Vector Regression(SVR),and Random Forest Regression(RFR)were employed.Furthermore,particle swarm optimization(PSO)algorithm and cross-validation techniques were applied to fine-tune the model parameters,striving for peak prediction performance.The results indicated that optimized models generally exhibit enhanced predictive accuracy compared to their basic counterparts.Notably,the PSO-RFR model excels among all evaluated models,showcasing superior performance on the testing dataset.It achieves a coefficient of determination(R^(2))of 0.9231,a mean absolute error(MAE)of 2.1073,and a root mean square error(RMSE)of 3.6903.When compared to experimental results,the PSO-RFR and PSO-BPNN models demonstrate exceptional predictive accuracy.Notably,the PSO-BPNN model exhibits the closest R^(2)values between its training and test sets.This close alignment of R^(2)values between the training and testing sets reflects the PSO-BPNN model’s superior generalization ability for unseen data.The findings present an efficient method for predicting concrete’s compressive strength,contributing to the sustainable development of concrete materials,and providing theoretical support for their research and application. 展开更多
关键词 waste glass powder concrete compressive strength machine learning particle swarm optimization algorithm VISUALIZATION
在线阅读 下载PDF
基于Q-Learning的多模态自适应光伏功率优化组合预测
13
作者 隗知初 杨苹 +3 位作者 周钱雨凡 陈文皓 万思洋 崔嘉雁 《电力工程技术》 北大核心 2026年第1期115-124,163,共11页
针对光伏功率序列波动性强、随机性高的问题,文中提出一种基于Q-Learning的多模态自适应光伏功率优化组合预测模型。首先,采用鲸鱼优化算法的变分模态分解方法,将原始光伏功率序列分解成不同子模态,并通过集成特征筛选模型,确定各子模... 针对光伏功率序列波动性强、随机性高的问题,文中提出一种基于Q-Learning的多模态自适应光伏功率优化组合预测模型。首先,采用鲸鱼优化算法的变分模态分解方法,将原始光伏功率序列分解成不同子模态,并通过集成特征筛选模型,确定各子模态序列最敏感的气象因素。然后,构建反向传播神经网络、双向长短期记忆网络、门控循环单元网络和时间卷积网络4种基础预测模型。考虑到不同模型对不同频率特征的子序列预测能力不同,利用Q-Learning算法自适应选择各模态对应的最优基础模型组合方式。最后,将不同子模态的预测结果叠加重构,得到最终预测结果,并利用高分辨率光伏气象功率数据集进行验证。结果证明,文中所提出的基于Q-Learning的多模态自适应光伏功率优化组合预测模型,相较于单一模型的预测误差平均绝对误差下降了16.18%,均方误差下降了17.00%。 展开更多
关键词 鲸鱼优化算法 变分模态分解 Q-learning 功率预测 组合模型 光伏发电
在线阅读 下载PDF
Development and validation of machine learningbased in-hospital mortality predictive models for acute aortic syndrome in emergency departments
14
作者 Yuanwei Fu Yilan Yang +6 位作者 Hua Zhang Daidai Wang Qiangrong Zhai Lanfang Du Nijiati Muyesai YanxiaGao Qingbian Ma 《World Journal of Emergency Medicine》 2026年第1期43-49,共7页
BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suita... BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suitable for rapid clinical application.METHODS:In this multi-center retrospective cohort study,AAS patient data from three hospitals were analyzed.The modeling cohort included data from the First Affiliated Hospital of Zhengzhou University and the People’s Hospital of Xinjiang Uygur Autonomous Region,with Peking University Third Hospital data serving as the external test set.Four machine learning algorithms—logistic regression(LR),multilayer perceptron(MLP),Gaussian naive Bayes(GNB),and random forest(RF)—were used to develop predictive models based on 34 early-accessible clinical variables.A simplifi ed model was then derived based on fi ve key variables(Stanford type,pericardial eff usion,asymmetric peripheral arterial pulsation,decreased bowel sounds,and dyspnea)via Least Absolute Shrinkage and Selection Operator(LASSO)regression to improve ED applicability.RESULTS:A total of 929 patients were included in the modeling cohort,and 210 were included in the external test set.Four machine learning models based on 34 clinical variables were developed,achieving internal and external validation AUCs of 0.85-0.90 and 0.73-0.85,respectively.The simplifi ed model incorporating fi ve key variables demonstrated internal and external validation AUCs of 0.71-0.86 and 0.75-0.78,respectively.Both models showed robust calibration and predictive stability across datasets.CONCLUSION:Both kinds of models were built based on machine learning tools,and proved to have certain prediction performance and extrapolation. 展开更多
关键词 Emergency department Acute aortic syndrome MORTALITY Predictive model Machine learning algorithmS
暂未订购
A dual attention-based deep learning model for lithology identificationwhile drilling
15
作者 Jie Chen Zhen Gui +6 位作者 Yichao Rui Xusheng Zhao Xiaokang Pan Qingfeng Wang Yuanyuan Pu Zheng Li Maoyi Liu 《Journal of Rock Mechanics and Geotechnical Engineering》 2026年第2期1177-1192,共16页
Lithology identificationwhile drilling technology can obtain rock information in real-time.However,traditional lithology identificationmodels often face limitations in feature extraction and adaptability to complex ge... Lithology identificationwhile drilling technology can obtain rock information in real-time.However,traditional lithology identificationmodels often face limitations in feature extraction and adaptability to complex geological conditions,limiting their accuracy in challenging environments.To address these challenges,a deep learning model for lithology identificationwhile drilling is proposed.The proposed model introduces a dual attention mechanism in the long short-term memory(LSTM)network,effectively enhancing the ability to capture spatial and channel dimension information.Subsequently,the crayfishoptimization algorithm(COA)is applied to optimize the model network structure,thereby enhancing its lithology identificationcapability.Laboratory test results demonstrate that the proposed model achieves 97.15%accuracy on the testing set,significantlyoutperforming the traditional support vector machine(SVM)method(81.77%).Field tests under actual drilling conditions demonstrate an average accuracy of 91.96%for the proposed model,representing a 14.31%improvement over the LSTM model alone.The proposed model demonstrates robust adaptability and generalization ability across diverse operational scenarios.This research offers reliable technical support for lithology identification while drilling. 展开更多
关键词 Lithology identificationwhile drilling Deep learning Dual attention mechanism Metaheuristic algorithm Field applications
在线阅读 下载PDF
Bearing capacity prediction of open caissons in two-layered clays using five tree-based machine learning algorithms 被引量:2
16
作者 Rungroad Suppakul Kongtawan Sangjinda +3 位作者 Wittaya Jitchaijaroen Natakorn Phuksuksakul Suraparb Keawsawasvong Peem Nuaklong 《Intelligent Geoengineering》 2025年第2期55-65,共11页
Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered so... Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered soils remains a complex challenge.This study presents a novel application of five ensemble machine(ML)algorithms-random forest(RF),gradient boosting machine(GBM),extreme gradient boosting(XGBoost),adaptive boosting(AdaBoost),and categorical boosting(CatBoost)-to predict the undrained bearing capacity factor(Nc)of circular open caissons embedded in two-layered clay on the basis of results from finite element limit analysis(FELA).The input dataset consists of 1188 numerical simulations using the Tresca failure criterion,varying in geometrical and soil parameters.The FELA was performed via OptumG2 software with adaptive meshing techniques and verified against existing benchmark studies.The ML models were trained on 70% of the dataset and tested on the remaining 30%.Their performance was evaluated using six statistical metrics:coefficient of determination(R²),mean absolute error(MAE),root mean squared error(RMSE),index of scatter(IOS),RMSE-to-standard deviation ratio(RSR),and variance explained factor(VAF).The results indicate that all the models achieved high accuracy,with R²values exceeding 97.6%and RMSE values below 0.02.Among them,AdaBoost and CatBoost consistently outperformed the other methods across both the training and testing datasets,demonstrating superior generalizability and robustness.The proposed ML framework offers an efficient,accurate,and data-driven alternative to traditional methods for estimating caisson capacity in stratified soils.This approach can aid in reducing computational costs while improving reliability in the early stages of foundation design. 展开更多
关键词 Two-layered clay Open caisson Tree-based algorithms FELA Machine learning
在线阅读 下载PDF
Robust and Efficient Federated Learning for Machinery Fault Diagnosis in Internet of Things
17
作者 Zhen Wu Hao Liu +4 位作者 Linlin Zhang Zehui Zhang Jie Wu Haibin He Bin Zhou 《Computers, Materials & Continua》 2026年第4期1051-1069,共19页
Recently,Internet ofThings(IoT)has been increasingly integrated into the automotive sector,enabling the development of diverse applications such as the Internet of Vehicles(IoV)and intelligent connected vehicles.Lever... Recently,Internet ofThings(IoT)has been increasingly integrated into the automotive sector,enabling the development of diverse applications such as the Internet of Vehicles(IoV)and intelligent connected vehicles.Leveraging IoVtechnologies,operational data fromcore vehicle components can be collected and analyzed to construct fault diagnosis models,thereby enhancing vehicle safety.However,automakers often struggle to acquire sufficient fault data to support effective model training.To address this challenge,a robust and efficient federated learning method(REFL)is constructed for machinery fault diagnosis in collaborative IoV,which can organize multiple companies to collaboratively develop a comprehensive fault diagnosis model while keeping their data locally.In the REFL,the gradient-based adversary algorithm is first introduced to the fault diagnosis field to enhance the deep learning model robustness.Moreover,the adaptive gradient processing process is designed to improve the model training speed and ensure the model accuracy under unbalance data scenarios.The proposed REFL is evaluated on non-independent and identically distributed(non-IID)real-world machinery fault dataset.Experiment results demonstrate that the REFL can achieve better performance than traditional learning methods and are promising for real industrial fault diagnosis. 展开更多
关键词 Federated learning adversary algorithm Internet of Vehicles(IoV) fault diagnosis
在线阅读 下载PDF
A review of dynamic monitoring methods for intermittent rivers:Integrating remote sensing and machine learning
18
作者 XIE Chaoshuai Lv Aifeng 《Journal of Geographical Sciences》 2026年第3期763-796,共34页
Intermittent rivers and ephemeral streams(IRES),also known as non-perennial river segments(NPRs),have garnered attention due to their significant roles in watershed hydrology and ecosystem services,especially in the c... Intermittent rivers and ephemeral streams(IRES),also known as non-perennial river segments(NPRs),have garnered attention due to their significant roles in watershed hydrology and ecosystem services,especially in the context of climate change and escalating human activities.Recent advances in machine learning(ML)techniques have significantly improved the analysis of dynamic changes in IRES.Various ML models,including random forest(RF),long short-term memory(LSTM),and U-Net,demonstrate clear advantages in processing complex hydrological data,enhancing the efficiency and accuracy of IRES extraction from remote sensing data.Furthermore,hybrid ML approaches enhance predictive performance in complex hydrological scenarios by integrating multiple algorithms.However,ML methods still face challenges,including high data dependence,computational complexity,and scalability issues with models.This review proposes an IRES monitoring framework that combines satellite data with ML algorithms,integrating remote sensing technologies such as optical imaging and synthetic aperture radar,and evaluates the advantages and limitations of different ML methods.It further highlights the potential of integrating multiple ML techniques and high-resolution remote sensing data to monitor IRES dynamics,conduct ecological assessments,and support sustainable water management,offering a scientific foundation for addressing environmental and anthropogenic pressures. 展开更多
关键词 machine learning intermittent rivers and ephemeral streams remote sensing framework algorithm selection
原文传递
A novel approach to identify the spatial characteristics of ozone-precursor sensitivity based on interpretable machine learning
19
作者 Huiling He Kaihui Zhao +6 位作者 Zibing Yuan Jin Shen Yujun Lin Shu Zhang Menglei Wang Anqi Wang Puyu Lian 《Journal of Environmental Sciences》 2026年第1期54-63,共10页
To curb the worsening tropospheric ozone(O_(3))pollution problem in China,a rapid and accurate identification of O_(3)-precursor sensitivity(OPS)is a crucial prerequisite for formulating effective contingency O_(3) po... To curb the worsening tropospheric ozone(O_(3))pollution problem in China,a rapid and accurate identification of O_(3)-precursor sensitivity(OPS)is a crucial prerequisite for formulating effective contingency O_(3) pollution control strategies.However,currently widely-used methods,such as statistical models and numerical models,exhibit inherent limitations in identifying OPS in a timely and accurate manner.In this study,we developed a novel approach to identify OPS based on eXtreme Gradient Boosting model,Shapley additive explanation(SHAP)al-gorithm,and volatile organic compound(VOC)photochemical decay adjustment,using the meteorology and speciated pollutant monitoring data as the input.By comparing the difference in SHAP values between base sce-nario and precursor reduction scenario for nitrogen oxides(NO_(x))and VOCs,OPS was divided into NO_(x)-limited,VOCs-limited and transition regime.Using the long-lasting O_(3) pollution episode in the autumn of 2022 at the Guangdong-Hong Kong-Macao Greater Bay Area(GBA)as an example,we demonstrated large spatiotemporal heterogeneities of OPS over the GBA,which were generally shifted from NO_(x)-limited to VOCs-limited from September to October and more inclined to be VOCs-limited at the central and NO_(x)-limited in the peripheral areas.This study developed an innovative OPS identification method by comparing the difference in SHAP value before and after precursor emission reduction.Our method enables the accurate identification of OPS in the time scale of seconds,thereby providing a state-of-the-art tool for the rapid guidance of spatial-specific O_(3) control strategies. 展开更多
关键词 O_(3)-precursor sensitivity Machine learning Extreme gradient boosting model Shapley algorithm Greater bay area
原文传递
Rapid pathologic grading-based diagnosis of esophageal squamous cell carcinoma via Raman spectroscopy and a deep learning algorithm 被引量:1
20
作者 Xin-Ying Yu Jian Chen +2 位作者 Lian-Yu Li Feng-En Chen Qiang He 《World Journal of Gastroenterology》 2025年第14期32-46,共15页
BACKGROUND Esophageal squamous cell carcinoma is a major histological subtype of esophageal cancer.Many molecular genetic changes are associated with its occurrence.Raman spectroscopy has become a new method for the e... BACKGROUND Esophageal squamous cell carcinoma is a major histological subtype of esophageal cancer.Many molecular genetic changes are associated with its occurrence.Raman spectroscopy has become a new method for the early diagnosis of tumors because it can reflect the structures of substances and their changes at the molecular level.AIM To detect alterations in Raman spectral information across different stages of esophageal neoplasia.METHODS Different grades of esophageal lesions were collected,and a total of 360 groups of Raman spectrum data were collected.A 1D-transformer network model was proposed to handle the task of classifying the spectral data of esophageal squamous cell carcinoma.In addition,a deep learning model was applied to visualize the Raman spectral data and interpret their molecular characteristics.RESULTS A comparison among Raman spectral data with different pathological grades and a visual analysis revealed that the Raman peaks with significant differences were concentrated mainly at 1095 cm^(-1)(DNA,symmetric PO,and stretching vibration),1132 cm^(-1)(cytochrome c),1171 cm^(-1)(acetoacetate),1216 cm^(-1)(amide III),and 1315 cm^(-1)(glycerol).A comparison among the training results of different models revealed that the 1Dtransformer network performed best.A 93.30%accuracy value,a 96.65%specificity value,a 93.30%sensitivity value,and a 93.17%F1 score were achieved.CONCLUSION Raman spectroscopy revealed significantly different waveforms for the different stages of esophageal neoplasia.The combination of Raman spectroscopy and deep learning methods could significantly improve the accuracy of classification. 展开更多
关键词 Raman spectroscopy Esophageal neoplasia Early diagnosis Deep learning algorithm Rapid pathologic grading
暂未订购
上一页 1 2 250 下一页 到第
使用帮助 返回顶部