期刊文献+
共找到191篇文章
< 1 2 10 >
每页显示 20 50 100
Text-Image Feature Fine-Grained Learning for Joint Multimodal Aspect-Based Sentiment Analysis
1
作者 Tianzhi Zhang Gang Zhou +4 位作者 Shuang Zhang Shunhang Li Yepeng Sun Qiankun Pi Shuo Liu 《Computers, Materials & Continua》 SCIE EI 2025年第1期279-305,共27页
Joint Multimodal Aspect-based Sentiment Analysis(JMASA)is a significant task in the research of multimodal fine-grained sentiment analysis,which combines two subtasks:Multimodal Aspect Term Extraction(MATE)and Multimo... Joint Multimodal Aspect-based Sentiment Analysis(JMASA)is a significant task in the research of multimodal fine-grained sentiment analysis,which combines two subtasks:Multimodal Aspect Term Extraction(MATE)and Multimodal Aspect-oriented Sentiment Classification(MASC).Currently,most existing models for JMASA only perform text and image feature encoding from a basic level,but often neglect the in-depth analysis of unimodal intrinsic features,which may lead to the low accuracy of aspect term extraction and the poor ability of sentiment prediction due to the insufficient learning of intra-modal features.Given this problem,we propose a Text-Image Feature Fine-grained Learning(TIFFL)model for JMASA.First,we construct an enhanced adjacency matrix of word dependencies and adopt graph convolutional network to learn the syntactic structure features for text,which addresses the context interference problem of identifying different aspect terms.Then,the adjective-noun pairs extracted from image are introduced to enable the semantic representation of visual features more intuitive,which addresses the ambiguous semantic extraction problem during image feature learning.Thereby,the model performance of aspect term extraction and sentiment polarity prediction can be further optimized and enhanced.Experiments on two Twitter benchmark datasets demonstrate that TIFFL achieves competitive results for JMASA,MATE and MASC,thus validating the effectiveness of our proposed methods. 展开更多
关键词 multimodal sentiment analysis aspect-based sentiment analysis feature fine-grained learning graph convolutional network adjective-noun pairs
在线阅读 下载PDF
Enhanced Multimodal Sentiment Analysis via Integrated Spatial Position Encoding and Fusion Embedding
2
作者 Chenquan Gan Xu Liu +3 位作者 Yu Tang Xianrong Yu Qingyi Zhu Deepak Kumar Jain 《Computers, Materials & Continua》 2025年第12期5399-5421,共23页
Multimodal sentiment analysis aims to understand emotions from text,speech,and video data.However,current methods often overlook the dominant role of text and suffer from feature loss during integration.Given the vary... Multimodal sentiment analysis aims to understand emotions from text,speech,and video data.However,current methods often overlook the dominant role of text and suffer from feature loss during integration.Given the varying importance of each modality across different contexts,a central and pressing challenge in multimodal sentiment analysis lies in maximizing the use of rich intra-modal features while minimizing information loss during the fusion process.In response to these critical limitations,we propose a novel framework that integrates spatial position encoding and fusion embedding modules to address these issues.In our model,text is treated as the core modality,while speech and video features are selectively incorporated through a unique position-aware fusion process.The spatial position encoding strategy preserves the internal structural information of speech and visual modalities,enabling the model to capture localized intra-modal dependencies that are often overlooked.This design enhances the richness and discriminative power of the fused representation,enabling more accurate and context-aware sentiment prediction.Finally,we conduct comprehensive evaluations on two widely recognized standard datasets in the field—CMU-MOSI and CMU-MOSEI to validate the performance of the proposed model.The experimental results demonstrate that our model exhibits good performance and effectiveness for sentiment analysis tasks. 展开更多
关键词 multimodal sentiment analysis spatial position encoding fusion embedding feature loss reduction
在线阅读 下载PDF
Aspect-Level Sentiment Analysis of Bi-Graph Convolutional Networks Based on Enhanced Syntactic Structural Information
3
作者 Junpeng Hu Yegang Li 《Journal of Computer and Communications》 2025年第1期72-89,共18页
Aspect-oriented sentiment analysis is a meticulous sentiment analysis task that aims to analyse the sentiment polarity of specific aspects. Most of the current research builds graph convolutional networks based on dep... Aspect-oriented sentiment analysis is a meticulous sentiment analysis task that aims to analyse the sentiment polarity of specific aspects. Most of the current research builds graph convolutional networks based on dependent syntactic trees, which improves the classification performance of the models to some extent. However, the technical limitations of dependent syntactic trees can introduce considerable noise into the model. Meanwhile, it is difficult for a single graph convolutional network to aggregate both semantic and syntactic structural information of nodes, which affects the final sentence classification. To cope with the above problems, this paper proposes a bi-channel graph convolutional network model. The model introduces a phrase structure tree and transforms it into a hierarchical phrase matrix. The adjacency matrix of the dependent syntactic tree and the hierarchical phrase matrix are combined as the initial matrix of the graph convolutional network to enhance the syntactic information. The semantic information feature representations of the sentences are obtained by the graph convolutional network with a multi-head attention mechanism and fused to achieve complementary learning of dual-channel features. Experimental results show that the model performs well and improves the accuracy of sentiment classification on three public benchmark datasets, namely Rest14, Lap14 and Twitter. 展开更多
关键词 aspect-level sentiment analysis sentiment Knowledge Multi-Head Attention Mechanism Graph Convolutional Networks
在线阅读 下载PDF
Multimodal sentiment analysis for social media contents during public emergencies 被引量:1
4
作者 Tao Fan Hao Wang +2 位作者 Peng Wu Chen Ling Milad Taleby Ahvanooey 《Journal of Data and Information Science》 CSCD 2023年第3期61-87,共27页
Purpose:Nowadays,public opinions during public emergencies involve not only textual contents but also contain images.However,the existing works mainly focus on textual contents and they do not provide a satisfactory a... Purpose:Nowadays,public opinions during public emergencies involve not only textual contents but also contain images.However,the existing works mainly focus on textual contents and they do not provide a satisfactory accuracy of sentiment analysis,lacking the combination of multimodal contents.In this paper,we propose to combine texts and images generated in the social media to perform sentiment analysis.Design/methodology/approach:We propose a Deep Multimodal Fusion Model(DMFM),which combines textual and visual sentiment analysis.We first train word2vec model on a large-scale public emergency corpus to obtain semantic-rich word vectors as the input of textual sentiment analysis.BiLSTM is employed to generate encoded textual embeddings.To fully excavate visual information from images,a modified pretrained VGG16-based sentiment analysis network is used with the best-performed fine-tuning strategy.A multimodal fusion method is implemented to fuse textual and visual embeddings completely,producing predicted labels.Findings:We performed extensive experiments on Weibo and Twitter public emergency datasets,to evaluate the performance of our proposed model.Experimental results demonstrate that the DMFM provides higher accuracy compared with baseline models.The introduction of images can boost the performance of sentiment analysis during public emergencies.Research limitations:In the future,we will test our model in a wider dataset.We will also consider a better way to learn the multimodal fusion information.Practical implications:We build an efficient multimodal sentiment analysis model for the social media contents during public emergencies.Originality/value:We consider the images posted by online users during public emergencies on social platforms.The proposed method can present a novel scope for sentiment analysis during public emergencies and provide the decision support for the government when formulating policies in public emergencies. 展开更多
关键词 Public emergency multimodal sentiment analysis Social platform Textual sentiment analysis Visual sentiment analysis
在线阅读 下载PDF
A Robust Framework for Multimodal Sentiment Analysis with Noisy Labels Generated from Distributed Data Annotation 被引量:1
5
作者 Kai Jiang Bin Cao Jing Fan 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第6期2965-2984,共20页
Multimodal sentiment analysis utilizes multimodal data such as text,facial expressions and voice to detect people’s attitudes.With the advent of distributed data collection and annotation,we can easily obtain and sha... Multimodal sentiment analysis utilizes multimodal data such as text,facial expressions and voice to detect people’s attitudes.With the advent of distributed data collection and annotation,we can easily obtain and share such multimodal data.However,due to professional discrepancies among annotators and lax quality control,noisy labels might be introduced.Recent research suggests that deep neural networks(DNNs)will overfit noisy labels,leading to the poor performance of the DNNs.To address this challenging problem,we present a Multimodal Robust Meta Learning framework(MRML)for multimodal sentiment analysis to resist noisy labels and correlate distinct modalities simultaneously.Specifically,we propose a two-layer fusion net to deeply fuse different modalities and improve the quality of the multimodal data features for label correction and network training.Besides,a multiple meta-learner(label corrector)strategy is proposed to enhance the label correction approach and prevent models from overfitting to noisy labels.We conducted experiments on three popular multimodal datasets to verify the superiority of ourmethod by comparing it with four baselines. 展开更多
关键词 Distributed data collection multimodal sentiment analysis meta learning learn with noisy labels
在线阅读 下载PDF
Leveraging Vision-Language Pre-Trained Model and Contrastive Learning for Enhanced Multimodal Sentiment Analysis
6
作者 Jieyu An Wan Mohd Nazmee Wan Zainon Binfen Ding 《Intelligent Automation & Soft Computing》 SCIE 2023年第8期1673-1689,共17页
Multimodal sentiment analysis is an essential area of research in artificial intelligence that combines multiple modes,such as text and image,to accurately assess sentiment.However,conventional approaches that rely on... Multimodal sentiment analysis is an essential area of research in artificial intelligence that combines multiple modes,such as text and image,to accurately assess sentiment.However,conventional approaches that rely on unimodal pre-trained models for feature extraction from each modality often overlook the intrinsic connections of semantic information between modalities.This limitation is attributed to their training on unimodal data,and necessitates the use of complex fusion mechanisms for sentiment analysis.In this study,we present a novel approach that combines a vision-language pre-trained model with a proposed multimodal contrastive learning method.Our approach harnesses the power of transfer learning by utilizing a vision-language pre-trained model to extract both visual and textual representations in a unified framework.We employ a Transformer architecture to integrate these representations,thereby enabling the capture of rich semantic infor-mation in image-text pairs.To further enhance the representation learning of these pairs,we introduce our proposed multimodal contrastive learning method,which leads to improved performance in sentiment analysis tasks.Our approach is evaluated through extensive experiments on two publicly accessible datasets,where we demonstrate its effectiveness.We achieve a significant improvement in sentiment analysis accuracy,indicating the supe-riority of our approach over existing techniques.These results highlight the potential of multimodal sentiment analysis and underscore the importance of considering the intrinsic semantic connections between modalities for accurate sentiment assessment. 展开更多
关键词 multimodal sentiment analysis vision–language pre-trained model contrastive learning sentiment classification
在线阅读 下载PDF
GLAMSNet:A Gated-Linear Aspect-Aware Multimodal Sentiment Network with Alignment Supervision and External Knowledge Guidance
7
作者 Dan Wang Zhoubin Li +1 位作者 Yuze Xia Zhenhua Yu 《Computers, Materials & Continua》 2025年第12期5823-5845,共23页
Multimodal Aspect-Based Sentiment Analysis(MABSA)aims to detect sentiment polarity toward specific aspects by leveraging both textual and visual inputs.However,existing models suffer from weak aspectimage alignment,mo... Multimodal Aspect-Based Sentiment Analysis(MABSA)aims to detect sentiment polarity toward specific aspects by leveraging both textual and visual inputs.However,existing models suffer from weak aspectimage alignment,modality imbalance dominated by textual signals,and limited reasoning for implicit or ambiguous sentiments requiring external knowledge.To address these issues,we propose a unified framework named Gated-Linear Aspect-Aware Multimodal Sentiment Network(GLAMSNet).First of all,an input encoding module is employed to construct modality-specific and aspect-aware representations.Subsequently,we introduce an image–aspect correlation matching module to provide hierarchical supervision for visual-textual alignment.Building upon these components,we further design a Gated-Linear Aspect-Aware Fusion(GLAF)module to enhance aspect-aware representation learning by adaptively filtering irrelevant textual information and refining semantic alignment under aspect guidance.Additionally,an External Language Model Knowledge-Guided mechanism is integrated to incorporate sentimentaware prior knowledge from GPT-4o,enabling robust semantic reasoning especially under noisy or ambiguous inputs.Experimental studies conducted based on Twitter-15 and Twitter-17 datasets demonstrate that the proposed model outperforms most state-of-the-art methods,achieving 79.36%accuracy and 74.72%F1-score,and 74.31%accuracy and 72.01%F1-score,respectively. 展开更多
关键词 sentiment analysis multimodal aspect-based sentiment analysis cross-modal alignment multimodal sentiment classification large language model
在线阅读 下载PDF
Aspect-Level Sentiment Analysis Based on Deep Learning
8
作者 Mengqi Zhang Jiazhao Chai +2 位作者 Jianxiang Cao Jialing Ji Tong Yi 《Computers, Materials & Continua》 SCIE EI 2024年第3期3743-3762,共20页
In recent years,deep learning methods have developed rapidly and found application in many fields,including natural language processing.In the field of aspect-level sentiment analysis,deep learning methods can also gr... In recent years,deep learning methods have developed rapidly and found application in many fields,including natural language processing.In the field of aspect-level sentiment analysis,deep learning methods can also greatly improve the performance of models.However,previous studies did not take into account the relationship between user feature extraction and contextual terms.To address this issue,we use data feature extraction and deep learning combined to develop an aspect-level sentiment analysis method.To be specific,we design user comment feature extraction(UCFE)to distill salient features from users’historical comments and transform them into representative user feature vectors.Then,the aspect-sentence graph convolutional neural network(ASGCN)is used to incorporate innovative techniques for calculating adjacency matrices;meanwhile,ASGCN emphasizes capturing nuanced semantics within relationships among aspect words and syntactic dependency types.Afterward,three embedding methods are devised to embed the user feature vector into the ASGCN model.The empirical validations verify the effectiveness of these models,consistently surpassing conventional benchmarks and reaffirming the indispensable role of deep learning in advancing sentiment analysis methodologies. 展开更多
关键词 aspect-level sentiment analysis deep learning graph convolutional neural network user features syntactic dependency tree
在线阅读 下载PDF
Aspect-Level Sentiment Analysis Incorporating Semantic and Syntactic Information
9
作者 Jiachen Yang Yegang Li +2 位作者 Hao Zhang Junpeng Hu Rujiang Bai 《Journal of Computer and Communications》 2024年第1期191-207,共17页
Aiming at the problem that existing models in aspect-level sentiment analysis cannot fully and effectively utilize sentence semantic and syntactic structure information, this paper proposes a graph neural network-base... Aiming at the problem that existing models in aspect-level sentiment analysis cannot fully and effectively utilize sentence semantic and syntactic structure information, this paper proposes a graph neural network-based aspect-level sentiment classification model. Self-attention, aspectual word multi-head attention and dependent syntactic relations are fused and the node representations are enhanced with graph convolutional networks to enable the model to fully learn the global semantic and syntactic structural information of sentences. Experimental results show that the model performs well on three public benchmark datasets Rest14, Lap14, and Twitter, improving the accuracy of sentiment classification. 展开更多
关键词 aspect-level sentiment analysis Attentional Mechanisms Dependent Syntactic Trees Graph Convolutional Neural Networks
在线阅读 下载PDF
Improving Targeted Multimodal Sentiment Classification with Semantic Description of Images 被引量:1
10
作者 Jieyu An Wan Mohd Nazmee Wan Zainon Zhang Hao 《Computers, Materials & Continua》 SCIE EI 2023年第6期5801-5815,共15页
Targeted multimodal sentiment classification(TMSC)aims to identify the sentiment polarity of a target mentioned in a multimodal post.The majority of current studies on this task focus on mapping the image and the text... Targeted multimodal sentiment classification(TMSC)aims to identify the sentiment polarity of a target mentioned in a multimodal post.The majority of current studies on this task focus on mapping the image and the text to a high-dimensional space in order to obtain and fuse implicit representations,ignoring the rich semantic information contained in the images and not taking into account the contribution of the visual modality in the multimodal fusion representation,which can potentially influence the results of TMSC tasks.This paper proposes a general model for Improving Targeted Multimodal Sentiment Classification with Semantic Description of Images(ITMSC)as a way to tackle these issues and improve the accu-racy of multimodal sentiment analysis.Specifically,the ITMSC model can automatically adjust the contribution of images in the fusion representation through the exploitation of semantic descriptions of images and text similarity relations.Further,we propose a target-based attention module to capture the target-text relevance,an image-based attention module to capture the image-text relevance,and a target-image matching module based on the former two modules to properly align the target with the image so that fine-grained semantic information can be extracted.Our experimental results demonstrate that our model achieves comparable performance with several state-of-the-art approaches on two multimodal sentiment datasets.Our findings indicate that incorporating semantic descriptions of images can enhance our understanding of multimodal content and lead to improved sentiment analysis performance. 展开更多
关键词 Targeted sentiment analysis multimodal sentiment classification visual sentiment textual sentiment social media
在线阅读 下载PDF
A Multimodal Sentiment Analysis Method Based on Multi-Granularity Guided Fusion
11
作者 Zilin Zhang Yan Liu +3 位作者 Jia Liu Senbao Hou Yuping Zhang Chenyuan Wang 《Computers, Materials & Continua》 2026年第2期1228-1241,共14页
With the growing demand formore comprehensive and nuanced sentiment understanding,Multimodal Sentiment Analysis(MSA)has gained significant traction in recent years and continues to attract widespread attention in the ... With the growing demand formore comprehensive and nuanced sentiment understanding,Multimodal Sentiment Analysis(MSA)has gained significant traction in recent years and continues to attract widespread attention in the academic community.Despite notable advances,existing approaches still face critical challenges in both information modeling and modality fusion.On one hand,many current methods rely heavily on encoders to extract global features from each modality,which limits their ability to capture latent fine-grained emotional cues within modalities.On the other hand,prevailing fusion strategies often lack mechanisms to model semantic discrepancies across modalities and to adaptively regulate modality interactions.To address these limitations,we propose a novel framework for MSA,termed Multi-Granularity Guided Fusion(MGGF).The proposed framework consists of three core components:(i)Multi-Granularity Feature Extraction Module,which simultaneously captures both global and local emotional features within each modality,and integrates them to construct richer intra-modal representations;(ii)Cross-ModalGuidance Learning Module(CMGL),which introduces a cross-modal scoring mechanism to quantify the divergence and complementarity betweenmodalities.These scores are then used as guiding signals to enable the fusion strategy to adaptively respond to scenarios of modality agreement or conflict;(iii)Cross-Modal Fusion Module(CMF),which learns the semantic dependencies among modalities and facilitates deep-level emotional feature interaction,thereby enhancing sentiment prediction with complementary information.We evaluate MGGF on two benchmark datasets:MVSA-Single and MVSA-Multiple.Experimental results demonstrate that MGGF outperforms the current state-of-the-art model CLMLF on MVSA-Single by achieving a 2.32% improvement in F1 score.On MVSA-Multiple,it surpasses MGNNS with a 0.26% increase in accuracy.These results substantiate the effectiveness ofMGGFin addressing two major limitations of existing methods—insufficient intra-modal fine-grained sentiment modeling and inadequate cross-modal semantic fusion. 展开更多
关键词 multimodal sentiment analysis cross-modal fusion cross-modal guided learning
在线阅读 下载PDF
Multimodal sentiment analysis based on contrastive learning and cross-modal guided fusion
12
作者 Liu Huanqi Sun Jingtao +1 位作者 Zhang Fengling Hou Wenyan 《The Journal of China Universities of Posts and Telecommunications》 2025年第4期18-33,共16页
Multimodal sentiment analysis,which integrates text,speech,and image modalities,has emerged as a prominent research direction in artificial intelligence for precise emotion assessment.However,current techniques experi... Multimodal sentiment analysis,which integrates text,speech,and image modalities,has emerged as a prominent research direction in artificial intelligence for precise emotion assessment.However,current techniques experience difficulties in efficiently managing redundancy and inconsistency across features from different modalities,compromising sentiment analysis accuracy.Additionally,while the analysis of intraclass emotional features has garnered substantial attention,studies of interclass relationships have been neglected.To address these challenges,a multimodal sentiment analysis method based on contrastive learning and cross-modal guided fusion(CLCGF)is proposed.This method encodes text and images to derive latent representations and employs a cross-modal guided module with sparse attention mechanisms to effectively integrate textual and visual features,thereby mitigating redundancy issues within each modality's features.In addition to the sentiment classification task,a supervised contrastive learning task is incorporated to aid the model in learning effective features from multimodal data related to emotions.To assess the efficacy of the CLCGF method,experiments were conducted on three public datasets:MVSA-Single,MVSA-Multiple and HFM.The experimental results indicate that CLCGF significantly improves sentiment analysis accuracy compared with traditional methods. 展开更多
关键词 multimodal sentiment analysis sparse attention contrastive learning
原文传递
跨模态特征增强与层次化MLP通信的多模态情感分析
13
作者 王旭阳 马瑾 《广西师范大学学报(自然科学版)》 北大核心 2026年第1期91-101,共11页
在多模态情感分析任务中,由于非语言模态信息利用不充分、跨模态交互缺乏细粒度关联建模以及层次化语义融合机制不完善,导致不同模态之间的情感信息难以实现有效融合。为此,本文提出一种跨模态特征增强与层次化MLP通信的多模态情感分析... 在多模态情感分析任务中,由于非语言模态信息利用不充分、跨模态交互缺乏细粒度关联建模以及层次化语义融合机制不完善,导致不同模态之间的情感信息难以实现有效融合。为此,本文提出一种跨模态特征增强与层次化MLP通信的多模态情感分析方法。该方法构建渐进式融合架构,首先通过跨模态注意力机制增强非语言模态信息,捕捉多对多的跨模态细粒度交互;继而使用层次化MLP通信模块,在模态融合维度与时间建模维度上分别设计并行与堆叠的MLP模块,实现水平与垂直方向的层次化特征交互,有效提升情感理解的准确性与表达能力。实验结果表明,本文模型在CMU-MOSI上,Acc2和F_(1)值较次优模型分别提升0.89和0.77个百分点,在CMU-MOSEI上对比实验各项指标均优于基准模型,Acc2、F_(1)值分别达到86.34%、86.25%。 展开更多
关键词 多模态 情感分析 跨模态注意力 层次化MLP通信 门控单元
在线阅读 下载PDF
基于提示学习和注意力机制的多模态方面级情感分析
14
作者 曾桢 廖茂熙 +2 位作者 杨楠 罗燕 杨超 《科学技术与工程》 北大核心 2026年第2期708-716,共9页
针对现有多模态方面级情感分析模型存在不同模态之间关联性不强,难以有效利用它们之间相关信息的问题,提出了一种基于提示学习和注意力机制的多模态方面级情感分析模型。首先,在文本和图像特征分别输入多层Transformer层之前,对文本特... 针对现有多模态方面级情感分析模型存在不同模态之间关联性不强,难以有效利用它们之间相关信息的问题,提出了一种基于提示学习和注意力机制的多模态方面级情感分析模型。首先,在文本和图像特征分别输入多层Transformer层之前,对文本特征添加提示信息,并通过耦合函数投影到图像特征中;其次,使用跨模态注意力机制进行特征融合,并通过双向门控循环单元(bidirectional gated recurrent unit,BiGRU)过滤融合过程中的无关信息;然后,构建了一个由方面信息引导的注意力模块,以加强对方面信息的关注度;最后,通过全连接层和softmax分类层得到情感分类结果。该模型分别在公开的Twitter-15和Twitter-17数据集上进行实验,结果显示与多个基线模型相比,所提模型的准确率和F1均有所提升,证明了所提模型在多模态方面级情感分析任务中的有效性。 展开更多
关键词 方面级情感分析 多模态 提示学习 特征提取 注意力机制 自然语言处理
在线阅读 下载PDF
Modal Interactive Feature Encoder for Multimodal Sentiment Analysis
15
作者 Xiaowei Zhao Jie Zhou Xiujuan Xu 《国际计算机前沿大会会议论文集》 EI 2023年第2期285-303,共19页
Multimodal Sentiment analysis refers to analyzing emotions in infor-mation carriers containing multiple modalities.To better analyze the features within and between modalities and solve the problem of incomplete multi... Multimodal Sentiment analysis refers to analyzing emotions in infor-mation carriers containing multiple modalities.To better analyze the features within and between modalities and solve the problem of incomplete multimodal feature fusion,this paper proposes a multimodal sentiment analysis model MIF(Modal Interactive Feature Encoder For Multimodal Sentiment Analysis).First,the global features of three modalities are obtained through unimodal feature extraction networks.Second,the inter-modal interactive feature encoder and the intra-modal interactive feature encoder extract similarity features between modal-ities and intra-modal special features separately.Finally,unimodal special features and the interaction information between modalities are decoded to get the fusion features and predict sentimental polarity results.We conduct extensive experi-ments on three public multimodal datasets,including one in Chinese and two in English.The results show that the performance of our approach is significantly improved compared with benchmark models. 展开更多
关键词 multimodal sentiment analysis Modal Interaction Feature ENCODER
原文传递
基于全景语义和多层次特征融合的方面级多模态情感分析
16
作者 张洋 胡慧君 刘茂福 《计算机工程与科学》 北大核心 2026年第2期341-352,共12页
目前,方面级多模态情感分析在相关任务中面临中文数据集匮乏与类别分布不均衡的问题。传统模型在处理情感信息时常忽视词语的局部依赖性,导致全局语义理解不足,难以准确定位情感信息。此外,多模态信息融合过程中难以有效筛选和过滤无关... 目前,方面级多模态情感分析在相关任务中面临中文数据集匮乏与类别分布不均衡的问题。传统模型在处理情感信息时常忽视词语的局部依赖性,导致全局语义理解不足,难以准确定位情感信息。此外,多模态信息融合过程中难以有效筛选和过滤无关信息,影响情感分类的准确性。为解决这些问题,构建了高质量多模态中文数据集WAMSA,并提出了一种基于全景语义和多层次特征融合的方面级多模态情感分析模型PSMFF。该模型通过全景语义网络模块,将文本特征与语义扩展信息相结合,利用GCN和图编码器捕捉细粒度和粗粒度的语义特征;多层次特征融合模块则通过局部引导提取相关图像特征,利用Transformer增强后,再与文本特征进行全局引导融合,生成丰富的多模态表征。实验结果表明,PSMFF模型在3个数据集上的表现优于多种基线模型。 展开更多
关键词 方面级多模态情感分析 WAMSA数据集 全景语义网络 多层次特征融合
在线阅读 下载PDF
A comprehensive survey on multimodal sentiment analysis:Techniques,models,and applications
17
作者 Heming Zhang 《Advances in Engineering Innovation》 2024年第7期47-52,共6页
Multimodal sentiment analysis(MSA)is an evolving field that integrates information from multiple modalities such as text,audio,and visual data to analyze and interpret human emotions and sentiments.This review provide... Multimodal sentiment analysis(MSA)is an evolving field that integrates information from multiple modalities such as text,audio,and visual data to analyze and interpret human emotions and sentiments.This review provides an extensive survey of the current state of multimodal sentiment analysis,highlighting fundamental concepts,popular datasets,techniques,models,challenges,applications,and future trends.By examining existing research and methodologies,this paper aims to present a cohesive understanding of MSA,Multimodal sentiment analysis(MSA)integrates data from text,audio,and visual sources,each contributing unique insights that enhance the overall understanding of sentiment.Textual data provides explicit content and context,audio data captures the emotional tone through speech characteristics,and visual data offers cues from facial expressions and body language.Despite these strengths,MSA faces limitations such as data integration challenges,computational complexity,and the scarcity of annotated multimodal datasets.Future directions include the development of advanced fusion techniques,real-time processing capabilities,and explainable AI models.These advancements will enable more accurate and robust sentiment analysis,improve user experiences,and enhance applications in human-computer interaction,healthcare,and social media analysis.By addressing these challenges and leveraging diverse data sources,MSA has the potential to revolutionize sentiment analysis and drive positive outcomes across various domains. 展开更多
关键词 multimodal sentiment analysis Natural Language Processing Emotion Recognition Data Fusion Techniques Deep Learning Models
在线阅读 下载PDF
基于集成学习与多模态大语言模型的图文情感分析方法
18
作者 王宁 武芳宇 +2 位作者 赵宇轩 张百灵 庞超逸 《计算机工程与应用》 北大核心 2026年第3期153-162,共10页
提出了一种融合集成学习与多模态大语言模型(multimodal large language models,MLLMs)的图文情感分析方法。针对图文情感分析中类别不平衡与跨模态情感不一致等关键挑战,设计了EMSAN(ensemble multimodal sentiment analysis network)... 提出了一种融合集成学习与多模态大语言模型(multimodal large language models,MLLMs)的图文情感分析方法。针对图文情感分析中类别不平衡与跨模态情感不一致等关键挑战,设计了EMSAN(ensemble multimodal sentiment analysis network)框架。该框架采用主辅模型结构,将在完整数据集上训练的主模型与在平衡子集上优化的辅助模型相结合,实现对各情感类别的精准识别。在特征学习方面,EMSAN采用两阶段策略增强情感特征:利用多模态大语言模型生成高质量的图像描述,缩小视觉与文本模态间的语义差距;引入一致性对比学习机制,通过对比文本和视觉特征的差异,强化跨模态情感的一致性表达,获得更为精细的特征。通过在平衡和不平衡数据集上的学习,EMSAN在保持数据自然分布的同时,有效缓解了类别不平衡问题。多个公共基准数据集上的实验结果表明,提出的方法取得了显著的性能提升。 展开更多
关键词 集成学习 多模态大语言模型 图文情感分析
在线阅读 下载PDF
跨模态不一致感知下双视角交互融合的多模态情感分析
19
作者 卜韵阳 齐彬廷 卜凡亮 《计算机科学》 北大核心 2026年第1期187-194,共8页
在社交媒体上,人们的评论通常会描述对应图像中的某一情感区域,图像和文本之间是具有对应信息的。以往的大多数多模态情感分析方法只是从单一视角探索图像和文本的相互影响,捕获图像区域和文本单词的对应关系,导致结果不是最优的。此外... 在社交媒体上,人们的评论通常会描述对应图像中的某一情感区域,图像和文本之间是具有对应信息的。以往的大多数多模态情感分析方法只是从单一视角探索图像和文本的相互影响,捕获图像区域和文本单词的对应关系,导致结果不是最优的。此外,社交媒体上的数据具有强烈的个人主观性,数据中的情感是多维和复杂的,导致出现了图像和文本情感一致性弱的数据。针对上述问题,提出了一种跨模态不一致感知下双视角交互融合的多模态情感分析模型。一方面,从全局和局部两种视角对图文特征进行跨模态交互,提供更全面、准确的情感分析,从而提升模型的表现和应用效果。另一方面,计算图文特征的不一致分数,用于代表图文不一致程度,以此来动态调控单模态表示和多模态表示的最终情感特征的权重,从而提高模型的鲁棒性。在MVSA-Single和MVSA-Multiple两个公共数据集上进行广泛实验,结果证明所提出的多模态情感分析模型与现有基线模型相比F1值分别提高0.59个百分点和0.39个百分点,具有有效性和优越性。 展开更多
关键词 多模态情感分析 跨模态不一致感知 双视角交互融合 动态调控 跨模态交互
在线阅读 下载PDF
基于H-GEM模型的多模态情感分析
20
作者 杨新航 王晶晶 +1 位作者 陈思宇 田宏 《计算机系统应用》 2026年第3期59-68,共10页
传统多模态情感分析方法在特征拼接和融合中易产生信息冗余,难以捕捉细粒度复杂情感特征,在模态缺失和跨域迁移场景下鲁棒性不足.同时,现有混合专家(MoE)方法大多为单层结构,专家分工不明确,存在功能重叠和泛化性欠佳的问题.本文提出一... 传统多模态情感分析方法在特征拼接和融合中易产生信息冗余,难以捕捉细粒度复杂情感特征,在模态缺失和跨域迁移场景下鲁棒性不足.同时,现有混合专家(MoE)方法大多为单层结构,专家分工不明确,存在功能重叠和泛化性欠佳的问题.本文提出一种分层自适应混合专家模型H-GEM(hierarchical gated expert mixture).通过构建3层分级专家体系:模态专家层提炼模态特征;融合与抽象专家层自适应选择融合策略;情感极性专家层进行细粒度建模.同时引入信息论与判别性约束提升专家选择的语义区分性和稀疏性.通过分层门控实现逐级决策,保证专家差异化分工与跨任务建模.在CMU-MOSI和CMU-MOSEI数据集上的实验结果表明, H-GEM在一系列指标上均优于基线模型.与单层MoE架构相比,显著降低的路由熵表明其能够有效缓解专家冗余问题.该模型在低资源和模态缺失复杂任务中表现出更高的鲁棒性,展现出良好的应用潜力. 展开更多
关键词 多模态情感分析 分层门控机制 混合专家模型 互信息约束 鲁棒性
在线阅读 下载PDF
上一页 1 2 10 下一页 到第
使用帮助 返回顶部