期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
Enhanced semi-supervised learning for top gas flow state classification to optimize emission and production in blast ironmaking furnaces
1
作者 Song Liu Qiqi Li +3 位作者 Qing Ye Zhiwei Zhao Dianyu E Shibo Kuang 《International Journal of Minerals,Metallurgy and Materials》 2026年第1期204-216,共13页
Automated classification of gas flow states in blast furnaces using top-camera imagery typically demands a large volume of labeled data,whose manual annotation is both labor-intensive and cost-prohibitive.To mitigate ... Automated classification of gas flow states in blast furnaces using top-camera imagery typically demands a large volume of labeled data,whose manual annotation is both labor-intensive and cost-prohibitive.To mitigate this challenge,we present an enhanced semi-supervised learning approach based on the Mean Teacher framework,incorporating a novel feature loss module to maximize classification performance with limited labeled samples.The model studies show that the proposed model surpasses both the baseline Mean Teacher model and fully supervised method in accuracy.Specifically,for datasets with 20%,30%,and 40%label ratios,using a single training iteration,the model yields accuracies of 78.61%,82.21%,and 85.2%,respectively,while multiple-cycle training iterations achieves 82.09%,81.97%,and 81.59%,respectively.Furthermore,scenario-specific training schemes are introduced to support diverse deployment need.These findings highlight the potential of the proposed technique in minimizing labeling requirements and advancing intelligent blast furnace diagnostics. 展开更多
关键词 blast furnace gas flow state semi-supervised learning mean teacher feature loss
在线阅读 下载PDF
Enhanced Multimodal Sentiment Analysis via Integrated Spatial Position Encoding and Fusion Embedding
2
作者 Chenquan Gan Xu Liu +3 位作者 Yu Tang Xianrong Yu Qingyi Zhu Deepak Kumar Jain 《Computers, Materials & Continua》 2025年第12期5399-5421,共23页
Multimodal sentiment analysis aims to understand emotions from text,speech,and video data.However,current methods often overlook the dominant role of text and suffer from feature loss during integration.Given the vary... Multimodal sentiment analysis aims to understand emotions from text,speech,and video data.However,current methods often overlook the dominant role of text and suffer from feature loss during integration.Given the varying importance of each modality across different contexts,a central and pressing challenge in multimodal sentiment analysis lies in maximizing the use of rich intra-modal features while minimizing information loss during the fusion process.In response to these critical limitations,we propose a novel framework that integrates spatial position encoding and fusion embedding modules to address these issues.In our model,text is treated as the core modality,while speech and video features are selectively incorporated through a unique position-aware fusion process.The spatial position encoding strategy preserves the internal structural information of speech and visual modalities,enabling the model to capture localized intra-modal dependencies that are often overlooked.This design enhances the richness and discriminative power of the fused representation,enabling more accurate and context-aware sentiment prediction.Finally,we conduct comprehensive evaluations on two widely recognized standard datasets in the field—CMU-MOSI and CMU-MOSEI to validate the performance of the proposed model.The experimental results demonstrate that our model exhibits good performance and effectiveness for sentiment analysis tasks. 展开更多
关键词 Multimodal sentiment analysis spatial position encoding fusion embedding feature loss reduction
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部