期刊文献+
共找到184篇文章
< 1 2 10 >
每页显示 20 50 100
Text-Image Feature Fine-Grained Learning for Joint Multimodal Aspect-Based Sentiment Analysis
1
作者 Tianzhi Zhang Gang Zhou +4 位作者 Shuang Zhang Shunhang Li Yepeng Sun Qiankun Pi Shuo Liu 《Computers, Materials & Continua》 SCIE EI 2025年第1期279-305,共27页
Joint Multimodal Aspect-based Sentiment Analysis(JMASA)is a significant task in the research of multimodal fine-grained sentiment analysis,which combines two subtasks:Multimodal Aspect Term Extraction(MATE)and Multimo... Joint Multimodal Aspect-based Sentiment Analysis(JMASA)is a significant task in the research of multimodal fine-grained sentiment analysis,which combines two subtasks:Multimodal Aspect Term Extraction(MATE)and Multimodal Aspect-oriented Sentiment Classification(MASC).Currently,most existing models for JMASA only perform text and image feature encoding from a basic level,but often neglect the in-depth analysis of unimodal intrinsic features,which may lead to the low accuracy of aspect term extraction and the poor ability of sentiment prediction due to the insufficient learning of intra-modal features.Given this problem,we propose a Text-Image Feature Fine-grained Learning(TIFFL)model for JMASA.First,we construct an enhanced adjacency matrix of word dependencies and adopt graph convolutional network to learn the syntactic structure features for text,which addresses the context interference problem of identifying different aspect terms.Then,the adjective-noun pairs extracted from image are introduced to enable the semantic representation of visual features more intuitive,which addresses the ambiguous semantic extraction problem during image feature learning.Thereby,the model performance of aspect term extraction and sentiment polarity prediction can be further optimized and enhanced.Experiments on two Twitter benchmark datasets demonstrate that TIFFL achieves competitive results for JMASA,MATE and MASC,thus validating the effectiveness of our proposed methods. 展开更多
关键词 multimodal sentiment analysis aspect-based sentiment analysis feature fine-grained learning graph convolutional network adjective-noun pairs
在线阅读 下载PDF
Multimodal sentiment analysis for social media contents during public emergencies 被引量:1
2
作者 Tao Fan Hao Wang +2 位作者 Peng Wu Chen Ling Milad Taleby Ahvanooey 《Journal of Data and Information Science》 CSCD 2023年第3期61-87,共27页
Purpose:Nowadays,public opinions during public emergencies involve not only textual contents but also contain images.However,the existing works mainly focus on textual contents and they do not provide a satisfactory a... Purpose:Nowadays,public opinions during public emergencies involve not only textual contents but also contain images.However,the existing works mainly focus on textual contents and they do not provide a satisfactory accuracy of sentiment analysis,lacking the combination of multimodal contents.In this paper,we propose to combine texts and images generated in the social media to perform sentiment analysis.Design/methodology/approach:We propose a Deep Multimodal Fusion Model(DMFM),which combines textual and visual sentiment analysis.We first train word2vec model on a large-scale public emergency corpus to obtain semantic-rich word vectors as the input of textual sentiment analysis.BiLSTM is employed to generate encoded textual embeddings.To fully excavate visual information from images,a modified pretrained VGG16-based sentiment analysis network is used with the best-performed fine-tuning strategy.A multimodal fusion method is implemented to fuse textual and visual embeddings completely,producing predicted labels.Findings:We performed extensive experiments on Weibo and Twitter public emergency datasets,to evaluate the performance of our proposed model.Experimental results demonstrate that the DMFM provides higher accuracy compared with baseline models.The introduction of images can boost the performance of sentiment analysis during public emergencies.Research limitations:In the future,we will test our model in a wider dataset.We will also consider a better way to learn the multimodal fusion information.Practical implications:We build an efficient multimodal sentiment analysis model for the social media contents during public emergencies.Originality/value:We consider the images posted by online users during public emergencies on social platforms.The proposed method can present a novel scope for sentiment analysis during public emergencies and provide the decision support for the government when formulating policies in public emergencies. 展开更多
关键词 Public emergency multimodal sentiment analysis Social platform Textual sentiment analysis Visual sentiment analysis
在线阅读 下载PDF
A Robust Framework for Multimodal Sentiment Analysis with Noisy Labels Generated from Distributed Data Annotation 被引量:1
3
作者 Kai Jiang Bin Cao Jing Fan 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第6期2965-2984,共20页
Multimodal sentiment analysis utilizes multimodal data such as text,facial expressions and voice to detect people’s attitudes.With the advent of distributed data collection and annotation,we can easily obtain and sha... Multimodal sentiment analysis utilizes multimodal data such as text,facial expressions and voice to detect people’s attitudes.With the advent of distributed data collection and annotation,we can easily obtain and share such multimodal data.However,due to professional discrepancies among annotators and lax quality control,noisy labels might be introduced.Recent research suggests that deep neural networks(DNNs)will overfit noisy labels,leading to the poor performance of the DNNs.To address this challenging problem,we present a Multimodal Robust Meta Learning framework(MRML)for multimodal sentiment analysis to resist noisy labels and correlate distinct modalities simultaneously.Specifically,we propose a two-layer fusion net to deeply fuse different modalities and improve the quality of the multimodal data features for label correction and network training.Besides,a multiple meta-learner(label corrector)strategy is proposed to enhance the label correction approach and prevent models from overfitting to noisy labels.We conducted experiments on three popular multimodal datasets to verify the superiority of ourmethod by comparing it with four baselines. 展开更多
关键词 Distributed data collection multimodal sentiment analysis meta learning learn with noisy labels
在线阅读 下载PDF
Leveraging Vision-Language Pre-Trained Model and Contrastive Learning for Enhanced Multimodal Sentiment Analysis
4
作者 Jieyu An Wan Mohd Nazmee Wan Zainon Binfen Ding 《Intelligent Automation & Soft Computing》 SCIE 2023年第8期1673-1689,共17页
Multimodal sentiment analysis is an essential area of research in artificial intelligence that combines multiple modes,such as text and image,to accurately assess sentiment.However,conventional approaches that rely on... Multimodal sentiment analysis is an essential area of research in artificial intelligence that combines multiple modes,such as text and image,to accurately assess sentiment.However,conventional approaches that rely on unimodal pre-trained models for feature extraction from each modality often overlook the intrinsic connections of semantic information between modalities.This limitation is attributed to their training on unimodal data,and necessitates the use of complex fusion mechanisms for sentiment analysis.In this study,we present a novel approach that combines a vision-language pre-trained model with a proposed multimodal contrastive learning method.Our approach harnesses the power of transfer learning by utilizing a vision-language pre-trained model to extract both visual and textual representations in a unified framework.We employ a Transformer architecture to integrate these representations,thereby enabling the capture of rich semantic infor-mation in image-text pairs.To further enhance the representation learning of these pairs,we introduce our proposed multimodal contrastive learning method,which leads to improved performance in sentiment analysis tasks.Our approach is evaluated through extensive experiments on two publicly accessible datasets,where we demonstrate its effectiveness.We achieve a significant improvement in sentiment analysis accuracy,indicating the supe-riority of our approach over existing techniques.These results highlight the potential of multimodal sentiment analysis and underscore the importance of considering the intrinsic semantic connections between modalities for accurate sentiment assessment. 展开更多
关键词 multimodal sentiment analysis vision–language pre-trained model contrastive learning sentiment classification
在线阅读 下载PDF
Improving Targeted Multimodal Sentiment Classification with Semantic Description of Images 被引量:1
5
作者 Jieyu An Wan Mohd Nazmee Wan Zainon Zhang Hao 《Computers, Materials & Continua》 SCIE EI 2023年第6期5801-5815,共15页
Targeted multimodal sentiment classification(TMSC)aims to identify the sentiment polarity of a target mentioned in a multimodal post.The majority of current studies on this task focus on mapping the image and the text... Targeted multimodal sentiment classification(TMSC)aims to identify the sentiment polarity of a target mentioned in a multimodal post.The majority of current studies on this task focus on mapping the image and the text to a high-dimensional space in order to obtain and fuse implicit representations,ignoring the rich semantic information contained in the images and not taking into account the contribution of the visual modality in the multimodal fusion representation,which can potentially influence the results of TMSC tasks.This paper proposes a general model for Improving Targeted Multimodal Sentiment Classification with Semantic Description of Images(ITMSC)as a way to tackle these issues and improve the accu-racy of multimodal sentiment analysis.Specifically,the ITMSC model can automatically adjust the contribution of images in the fusion representation through the exploitation of semantic descriptions of images and text similarity relations.Further,we propose a target-based attention module to capture the target-text relevance,an image-based attention module to capture the image-text relevance,and a target-image matching module based on the former two modules to properly align the target with the image so that fine-grained semantic information can be extracted.Our experimental results demonstrate that our model achieves comparable performance with several state-of-the-art approaches on two multimodal sentiment datasets.Our findings indicate that incorporating semantic descriptions of images can enhance our understanding of multimodal content and lead to improved sentiment analysis performance. 展开更多
关键词 Targeted sentiment analysis multimodal sentiment classification visual sentiment textual sentiment social media
在线阅读 下载PDF
Enhanced Multimodal Sentiment Analysis via Integrated Spatial Position Encoding and Fusion Embedding
6
作者 Chenquan Gan Xu Liu +3 位作者 Yu Tang Xianrong Yu Qingyi Zhu Deepak Kumar Jain 《Computers, Materials & Continua》 2025年第12期5399-5421,共23页
Multimodal sentiment analysis aims to understand emotions from text,speech,and video data.However,current methods often overlook the dominant role of text and suffer from feature loss during integration.Given the vary... Multimodal sentiment analysis aims to understand emotions from text,speech,and video data.However,current methods often overlook the dominant role of text and suffer from feature loss during integration.Given the varying importance of each modality across different contexts,a central and pressing challenge in multimodal sentiment analysis lies in maximizing the use of rich intra-modal features while minimizing information loss during the fusion process.In response to these critical limitations,we propose a novel framework that integrates spatial position encoding and fusion embedding modules to address these issues.In our model,text is treated as the core modality,while speech and video features are selectively incorporated through a unique position-aware fusion process.The spatial position encoding strategy preserves the internal structural information of speech and visual modalities,enabling the model to capture localized intra-modal dependencies that are often overlooked.This design enhances the richness and discriminative power of the fused representation,enabling more accurate and context-aware sentiment prediction.Finally,we conduct comprehensive evaluations on two widely recognized standard datasets in the field—CMU-MOSI and CMU-MOSEI to validate the performance of the proposed model.The experimental results demonstrate that our model exhibits good performance and effectiveness for sentiment analysis tasks. 展开更多
关键词 multimodal sentiment analysis spatial position encoding fusion embedding feature loss reduction
在线阅读 下载PDF
End-to-end aspect category sentiment analysis based on type graph convolutional networks
7
作者 邵清 ZHANG Wenshuang WANG Shaojun 《High Technology Letters》 EI CAS 2023年第3期325-334,共10页
For the existing aspect category sentiment analysis research,most of the aspects are given for sentiment extraction,and this pipeline method is prone to error accumulation,and the use of graph convolutional neural net... For the existing aspect category sentiment analysis research,most of the aspects are given for sentiment extraction,and this pipeline method is prone to error accumulation,and the use of graph convolutional neural network for aspect category sentiment analysis does not fully utilize the dependency type information between words,so it cannot enhance feature extraction.This paper proposes an end-to-end aspect category sentiment analysis(ETESA)model based on type graph convolutional networks.The model uses the bidirectional encoder representation from transformers(BERT)pretraining model to obtain aspect categories and word vectors containing contextual dynamic semantic information,which can solve the problem of polysemy;when using graph convolutional network(GCN)for feature extraction,the fusion operation of word vectors and initialization tensor of dependency types can obtain the importance values of different dependency types and enhance the text feature representation;by transforming aspect category and sentiment pair extraction into multiple single-label classification problems,aspect category and sentiment can be extracted simultaneously in an end-to-end way and solve the problem of error accumulation.Experiments are tested on three public datasets,and the results show that the ETESA model can achieve higher Precision,Recall and F1 value,proving the effectiveness of the model. 展开更多
关键词 aspect-based sentiment analysis(ABSA) bidirectional encoder representation from transformers(BERT) type graph convolutional network(TGCN) aspect category and senti-ment pair extraction
在线阅读 下载PDF
GLAMSNet:A Gated-Linear Aspect-Aware Multimodal Sentiment Network with Alignment Supervision and External Knowledge Guidance
8
作者 Dan Wang Zhoubin Li +1 位作者 Yuze Xia Zhenhua Yu 《Computers, Materials & Continua》 2025年第12期5823-5845,共23页
Multimodal Aspect-Based Sentiment Analysis(MABSA)aims to detect sentiment polarity toward specific aspects by leveraging both textual and visual inputs.However,existing models suffer from weak aspectimage alignment,mo... Multimodal Aspect-Based Sentiment Analysis(MABSA)aims to detect sentiment polarity toward specific aspects by leveraging both textual and visual inputs.However,existing models suffer from weak aspectimage alignment,modality imbalance dominated by textual signals,and limited reasoning for implicit or ambiguous sentiments requiring external knowledge.To address these issues,we propose a unified framework named Gated-Linear Aspect-Aware Multimodal Sentiment Network(GLAMSNet).First of all,an input encoding module is employed to construct modality-specific and aspect-aware representations.Subsequently,we introduce an image–aspect correlation matching module to provide hierarchical supervision for visual-textual alignment.Building upon these components,we further design a Gated-Linear Aspect-Aware Fusion(GLAF)module to enhance aspect-aware representation learning by adaptively filtering irrelevant textual information and refining semantic alignment under aspect guidance.Additionally,an External Language Model Knowledge-Guided mechanism is integrated to incorporate sentimentaware prior knowledge from GPT-4o,enabling robust semantic reasoning especially under noisy or ambiguous inputs.Experimental studies conducted based on Twitter-15 and Twitter-17 datasets demonstrate that the proposed model outperforms most state-of-the-art methods,achieving 79.36%accuracy and 74.72%F1-score,and 74.31%accuracy and 72.01%F1-score,respectively. 展开更多
关键词 sentiment analysis multimodal aspect-based sentiment analysis cross-modal alignment multimodal sentiment classification large language model
在线阅读 下载PDF
跨模态特征增强与层次化MLP通信的多模态情感分析
9
作者 王旭阳 马瑾 《广西师范大学学报(自然科学版)》 北大核心 2026年第1期91-101,共11页
在多模态情感分析任务中,由于非语言模态信息利用不充分、跨模态交互缺乏细粒度关联建模以及层次化语义融合机制不完善,导致不同模态之间的情感信息难以实现有效融合。为此,本文提出一种跨模态特征增强与层次化MLP通信的多模态情感分析... 在多模态情感分析任务中,由于非语言模态信息利用不充分、跨模态交互缺乏细粒度关联建模以及层次化语义融合机制不完善,导致不同模态之间的情感信息难以实现有效融合。为此,本文提出一种跨模态特征增强与层次化MLP通信的多模态情感分析方法。该方法构建渐进式融合架构,首先通过跨模态注意力机制增强非语言模态信息,捕捉多对多的跨模态细粒度交互;继而使用层次化MLP通信模块,在模态融合维度与时间建模维度上分别设计并行与堆叠的MLP模块,实现水平与垂直方向的层次化特征交互,有效提升情感理解的准确性与表达能力。实验结果表明,本文模型在CMU-MOSI上,Acc2和F_(1)值较次优模型分别提升0.89和0.77个百分点,在CMU-MOSEI上对比实验各项指标均优于基准模型,Acc2、F_(1)值分别达到86.34%、86.25%。 展开更多
关键词 多模态 情感分析 跨模态注意力 层次化MLP通信 门控单元
在线阅读 下载PDF
Fusing Syntactic Structure Information and Lexical Semantic Information for End-to-End Aspect-Based Sentiment Analysis 被引量:3
10
作者 Yong Bie Yan Yang Yiling Zhang 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2023年第2期230-243,共14页
The aspect-based sentiment analysis(ABSA)consists of two subtasksaspect term extraction and aspect sentiment prediction.Most methods conduct the ABSA task by handling the subtasks in a pipeline manner,whereby problems... The aspect-based sentiment analysis(ABSA)consists of two subtasksaspect term extraction and aspect sentiment prediction.Most methods conduct the ABSA task by handling the subtasks in a pipeline manner,whereby problems in performance and real application emerge.In this study,we propose an end-to-end ABSA model,namely,SSi-LSi,which fuses the syntactic structure information and the lexical semantic information,to address the limitation that existing end-to-end methods do not fully exploit the text information.Through two network branches,the model extracts syntactic structure information and lexical semantic information,which integrates the part of speech,sememes,and context,respectively.Then,on the basis of an attention mechanism,the model further realizes the fusion of the syntactic structure information and the lexical semantic information to obtain higher quality ABSA results,in which way the text information is fully used.Subsequent experiments demonstrate that the SSi-LSi model has certain advantages in using different text information. 展开更多
关键词 deep learning natural language processing aspect-based sentiment analysis graph convolutional
原文传递
A Multitask Multiview Neural Network for End-to-End Aspect-Based Sentiment Analysis 被引量:5
11
作者 Yong Bie Yan Yang 《Big Data Mining and Analytics》 EI 2021年第3期195-207,共13页
The aspect-based sentiment analysis(ABSA) consists of two subtasks—aspect term extraction and aspect sentiment prediction. Existing methods deal with both subtasks one by one in a pipeline manner, in which there lies... The aspect-based sentiment analysis(ABSA) consists of two subtasks—aspect term extraction and aspect sentiment prediction. Existing methods deal with both subtasks one by one in a pipeline manner, in which there lies some problems in performance and real application. This study investigates the end-to-end ABSA and proposes a novel multitask multiview network(MTMVN) architecture. Specifically, the architecture takes the unified ABSA as the main task with the two subtasks as auxiliary tasks. Meanwhile, the representation obtained from the branch network of the main task is regarded as the global view, whereas the representations of the two subtasks are considered two local views with different emphases. Through multitask learning, the main task can be facilitated by additional accurate aspect boundary information and sentiment polarity information. By enhancing the correlations between the views under the idea of multiview learning, the representation of the global view can be optimized to improve the overall performance of the model. The experimental results on three benchmark datasets show that the proposed method exceeds the existing pipeline methods and end-to-end methods, proving the superiority of our MTMVN architecture. 展开更多
关键词 deep learning multitask learning multiview learning natural language processing aspect-based sentiment analysis
原文传递
Aspect-Guided Multi-Graph Convolutional Networks for Aspect-based Sentiment Analysis
12
作者 Yong Wang Ningchuang Yang +1 位作者 Duoqian Miao Qiuyi Chen 《Data Intelligence》 EI 2024年第3期771-791,共21页
The Aspect-Based Sentiment Analysis(ABSA)task is designed to judge the sentiment polarity of a particular aspect in a review.Recent studies have proved that GCN can capture syntactic and semantic features from depende... The Aspect-Based Sentiment Analysis(ABSA)task is designed to judge the sentiment polarity of a particular aspect in a review.Recent studies have proved that GCN can capture syntactic and semantic features from dependency graphs generated by dependency trees and semantic graphs generated by Multi-headed self-attention(MHSA).However,these approaches do not highlight the sentiment information associated with aspect in the syntactic and semantic graphs.We propose the Aspect-Guided Multi-Graph Convolutional Networks(AGGCN)for Aspect-Based Sentiment Classification.Specifically,we reconstruct two kinds of graphs,changing the weight of the dependency graph by distance from aspect and improving the semantic graph by Aspect-guided MHSA.For interactive learning of syntax and semantics,we dynamically fuse syntactic and semantic diagrams to generate syntactic-semantic graphs to learn emotional features jointly.In addition,Multi-dropout is added to solve the overftting of AGGCN in training.The experimental results on extensive datasets show that our model AGGCN achieves particularly advanced results and validates the effectiveness of the model. 展开更多
关键词 Graph convolutional networks aspect-based sentiment analysis Multi-headed attention BERT encoder
原文传递
Modal Interactive Feature Encoder for Multimodal Sentiment Analysis
13
作者 Xiaowei Zhao Jie Zhou Xiujuan Xu 《国际计算机前沿大会会议论文集》 EI 2023年第2期285-303,共19页
Multimodal Sentiment analysis refers to analyzing emotions in infor-mation carriers containing multiple modalities.To better analyze the features within and between modalities and solve the problem of incomplete multi... Multimodal Sentiment analysis refers to analyzing emotions in infor-mation carriers containing multiple modalities.To better analyze the features within and between modalities and solve the problem of incomplete multimodal feature fusion,this paper proposes a multimodal sentiment analysis model MIF(Modal Interactive Feature Encoder For Multimodal Sentiment Analysis).First,the global features of three modalities are obtained through unimodal feature extraction networks.Second,the inter-modal interactive feature encoder and the intra-modal interactive feature encoder extract similarity features between modal-ities and intra-modal special features separately.Finally,unimodal special features and the interaction information between modalities are decoded to get the fusion features and predict sentimental polarity results.We conduct extensive experi-ments on three public multimodal datasets,including one in Chinese and two in English.The results show that the performance of our approach is significantly improved compared with benchmark models. 展开更多
关键词 multimodal sentiment analysis Modal Interaction Feature ENCODER
原文传递
A comprehensive survey on multimodal sentiment analysis:Techniques,models,and applications
14
作者 Heming Zhang 《Advances in Engineering Innovation》 2024年第7期47-52,共6页
Multimodal sentiment analysis(MSA)is an evolving field that integrates information from multiple modalities such as text,audio,and visual data to analyze and interpret human emotions and sentiments.This review provide... Multimodal sentiment analysis(MSA)is an evolving field that integrates information from multiple modalities such as text,audio,and visual data to analyze and interpret human emotions and sentiments.This review provides an extensive survey of the current state of multimodal sentiment analysis,highlighting fundamental concepts,popular datasets,techniques,models,challenges,applications,and future trends.By examining existing research and methodologies,this paper aims to present a cohesive understanding of MSA,Multimodal sentiment analysis(MSA)integrates data from text,audio,and visual sources,each contributing unique insights that enhance the overall understanding of sentiment.Textual data provides explicit content and context,audio data captures the emotional tone through speech characteristics,and visual data offers cues from facial expressions and body language.Despite these strengths,MSA faces limitations such as data integration challenges,computational complexity,and the scarcity of annotated multimodal datasets.Future directions include the development of advanced fusion techniques,real-time processing capabilities,and explainable AI models.These advancements will enable more accurate and robust sentiment analysis,improve user experiences,and enhance applications in human-computer interaction,healthcare,and social media analysis.By addressing these challenges and leveraging diverse data sources,MSA has the potential to revolutionize sentiment analysis and drive positive outcomes across various domains. 展开更多
关键词 multimodal sentiment analysis Natural Language Processing Emotion Recognition Data Fusion Techniques Deep Learning Models
在线阅读 下载PDF
跨模态不一致感知下双视角交互融合的多模态情感分析
15
作者 卜韵阳 齐彬廷 卜凡亮 《计算机科学》 北大核心 2026年第1期187-194,共8页
在社交媒体上,人们的评论通常会描述对应图像中的某一情感区域,图像和文本之间是具有对应信息的。以往的大多数多模态情感分析方法只是从单一视角探索图像和文本的相互影响,捕获图像区域和文本单词的对应关系,导致结果不是最优的。此外... 在社交媒体上,人们的评论通常会描述对应图像中的某一情感区域,图像和文本之间是具有对应信息的。以往的大多数多模态情感分析方法只是从单一视角探索图像和文本的相互影响,捕获图像区域和文本单词的对应关系,导致结果不是最优的。此外,社交媒体上的数据具有强烈的个人主观性,数据中的情感是多维和复杂的,导致出现了图像和文本情感一致性弱的数据。针对上述问题,提出了一种跨模态不一致感知下双视角交互融合的多模态情感分析模型。一方面,从全局和局部两种视角对图文特征进行跨模态交互,提供更全面、准确的情感分析,从而提升模型的表现和应用效果。另一方面,计算图文特征的不一致分数,用于代表图文不一致程度,以此来动态调控单模态表示和多模态表示的最终情感特征的权重,从而提高模型的鲁棒性。在MVSA-Single和MVSA-Multiple两个公共数据集上进行广泛实验,结果证明所提出的多模态情感分析模型与现有基线模型相比F1值分别提高0.59个百分点和0.39个百分点,具有有效性和优越性。 展开更多
关键词 多模态情感分析 跨模态不一致感知 双视角交互融合 动态调控 跨模态交互
在线阅读 下载PDF
基于对比学习与大语言模型增强的多模态方面级情感分析模型
16
作者 余传明 蒋展 孙邹驰 《现代情报》 北大核心 2026年第2期77-90,共14页
[目的/意义]针对多模态方面级情感分析(MABSA)研究领域存在的数据稀疏、数据不平衡等问题,探索大语言模型在MABSA任务中的应用和性能效果。[方法/过程]本文提出一种基于大语言模型数据增强和HiLo注意力对比学习的多模态方面级情感分析模... [目的/意义]针对多模态方面级情感分析(MABSA)研究领域存在的数据稀疏、数据不平衡等问题,探索大语言模型在MABSA任务中的应用和性能效果。[方法/过程]本文提出一种基于大语言模型数据增强和HiLo注意力对比学习的多模态方面级情感分析模型HLCL-GLM4。该模型调用ChatGLM4-Flash进行数据增强,采用Faster R-CNN和BART词嵌入分别获取文本和图像模态特征,将图像特征通过HiLo注意力机制进行建模,并使用一种自监督的对比学习策略进行模态特征学习和融合,提升样本多样性和情感语义的丰富性。[结果/结论]实验结果表明,HLCL-GLM4在Twitter-15和Twitter-17数据集上均取得了优异的性能。具体地,相较于最优基线模型,HLCL-GLM4在Twitter-15数据集的F1值提升1.6%,在Twitter-17数据集的F1值提升0.8%。 展开更多
关键词 多模态方面级情感分析 对比学习 大语言模型 提示工程 数据增强
在线阅读 下载PDF
Multi-Model Fusion Framework Using Deep Learning for Visual-Textual Sentiment Classification
17
作者 Israa K.Salman Al-Tameemi Mohammad-Reza Feizi-Derakhshi +1 位作者 Saeed Pashazadeh Mohammad Asadpour 《Computers, Materials & Continua》 SCIE EI 2023年第8期2145-2177,共33页
Multimodal Sentiment Analysis(SA)is gaining popularity due to its broad application potential.The existing studies have focused on the SA of single modalities,such as texts or photos,posing challenges in effectively h... Multimodal Sentiment Analysis(SA)is gaining popularity due to its broad application potential.The existing studies have focused on the SA of single modalities,such as texts or photos,posing challenges in effectively handling social media data with multiple modalities.Moreover,most multimodal research has concentrated on merely combining the two modalities rather than exploring their complex correlations,leading to unsatisfactory sentiment classification results.Motivated by this,we propose a new visualtextual sentiment classification model named Multi-Model Fusion(MMF),which uses a mixed fusion framework for SA to effectively capture the essential information and the intrinsic relationship between the visual and textual content.The proposed model comprises three deep neural networks.Two different neural networks are proposed to extract the most emotionally relevant aspects of image and text data.Thus,more discriminative features are gathered for accurate sentiment classification.Then,a multichannel joint fusion modelwith a self-attention technique is proposed to exploit the intrinsic correlation between visual and textual characteristics and obtain emotionally rich information for joint sentiment classification.Finally,the results of the three classifiers are integrated using a decision fusion scheme to improve the robustness and generalizability of the proposed model.An interpretable visual-textual sentiment classification model is further developed using the Local Interpretable Model-agnostic Explanation model(LIME)to ensure the model’s explainability and resilience.The proposed MMF model has been tested on four real-world sentiment datasets,achieving(99.78%)accuracy on Binary_Getty(BG),(99.12%)on Binary_iStock(BIS),(95.70%)on Twitter,and(79.06%)on the Multi-View Sentiment Analysis(MVSA)dataset.These results demonstrate the superior performance of our MMF model compared to single-model approaches and current state-of-the-art techniques based on model evaluation criteria. 展开更多
关键词 sentiment analysis multimodal classification deep learning joint fusion decision fusion INTERPRETABILITY
在线阅读 下载PDF
基于联合交互注意力的图文情感分析方法 被引量:1
18
作者 胡慧君 丁子毅 +1 位作者 张耀峰 刘茂福 《北京航空航天大学学报》 北大核心 2025年第7期2262-2270,共9页
社交媒体中的图文情感对于引导舆论走向具有重要意义,越来越受到自然语言处理(NLP)领域的广泛关注。当前,社交媒体图文情感分析的研究对象主要为单幅图像文本对,针对无时序性及多样性的图集文本对的研究相对较少,为有效挖掘图集中图像... 社交媒体中的图文情感对于引导舆论走向具有重要意义,越来越受到自然语言处理(NLP)领域的广泛关注。当前,社交媒体图文情感分析的研究对象主要为单幅图像文本对,针对无时序性及多样性的图集文本对的研究相对较少,为有效挖掘图集中图像与文本之间情感一致性信息,提出基于联合交互注意力的图文情感分析(SA-JIA)方法。该方法使用RoBERTa和双向门控循环单元(Bi-GRU)来提取文本表达特征,使用ResNet50获取图像视觉特征,利用联合注意力来找到图文情感信息表达一致的显著区域,获得新的文本和图像视觉特征,采用交互注意力关注模态间的特征交互,并进行多模态特征融合,进而完成情感分类任务。在IsTS-CN数据集和CCIR20-YQ数据集上进行了实验验证,结果表明:所提方法能够提升社交媒体图文情感分析的性能。 展开更多
关键词 社交媒体 图文情感分析 联合注意力 交互注意力 多模态融合
原文传递
基于元优化特征解耦的多模态跨域情感分析算法 被引量:1
19
作者 贾熹滨 李宸 +4 位作者 王珞 张沐晨 刘潇健 张旸旸 温家凯 《计算机研究与发展》 北大核心 2025年第11期2697-2709,共13页
多模态情感分析旨在利用多模态点评等数据识别用户情感倾向.为实现存在域偏移的跨域应用,常用无监督领域自适应方法.然而,该类方法着重于领域不变特征提取,忽略了目标领域特定特征的重要作用.为此,提出基于元优化的领域不变及领域特定... 多模态情感分析旨在利用多模态点评等数据识别用户情感倾向.为实现存在域偏移的跨域应用,常用无监督领域自适应方法.然而,该类方法着重于领域不变特征提取,忽略了目标领域特定特征的重要作用.为此,提出基于元优化的领域不变及领域特定特征解耦网络.首先,通过嵌入情感适配器对预训练大模型微调,建立图文融合情感特征编码器.进而,构建基于因子分解的特征解耦模块,分别利用领域对抗及领域分类、协同独立性约束,实现知识可传递的领域不变特征编码的同时,提取领域特定特征以增强目标域情感分类性能.为保证特征解耦与情感分类的总体优化方向一致性,提出基于元学习的元优化训练策略,实现情感分析网络的协同优化.基于MVSA和Yelp数据集构建的双向情感迁移任务的对比实验表明,较之其他先进的图文情感迁移算法,所提算法于双向情感迁移任务的精确率、召回率和F1值3项评价指标均取得了优异的性能. 展开更多
关键词 多模态情感分析 无监督领域自适应 跨领域情感分类 特征解耦 元优化
在线阅读 下载PDF
基于多模态信息融合的中文隐式情感分析 被引量:4
20
作者 张换香 李梦云 张景 《计算机工程与应用》 北大核心 2025年第2期179-190,共12页
隐式情感表达中缺乏显式情感词,给隐式情感分析带来一定的挑战。为有效解决此问题,借助外部信息是有效解决隐式情感分析的方法之一。与现有的主要借助单一文本信息的研究不同,提出一种融合多模态信息(包括语音和视频)的隐式情感分析方... 隐式情感表达中缺乏显式情感词,给隐式情感分析带来一定的挑战。为有效解决此问题,借助外部信息是有效解决隐式情感分析的方法之一。与现有的主要借助单一文本信息的研究不同,提出一种融合多模态信息(包括语音和视频)的隐式情感分析方法。通过从语音中提取音调、强度等声学特征,以及从视频中捕捉面部表情等视觉特征,辅助理解隐式情感。利用BiLSTM网络挖掘各单模态内部的上下文信息;结合多头互注意力机制分别捕捉与文本相关的语音和视觉特征,并通过迭代优化,减少非文本模态的低阶冗余信息。此外,通过设计以文本为中心的交叉注意融合模块,强化隐式文本特征表示,并处理模态间的异质性,增强隐式情感分析的综合性能。在CMUMOSI、CMU-MOSEI、MUMETA数据集上的实验结果表明,所提出的模型优于其他基线模型。这种针对隐式情感分析的多模态处理策略,充分利用语音和视觉外部知识,更全面、准确地捕捉隐式情感表达,有效提升了隐式情感分析的准确率。 展开更多
关键词 隐式情感分析 深度神经网络 多模态 注意力机制 特征融合
在线阅读 下载PDF
上一页 1 2 10 下一页 到第
使用帮助 返回顶部