期刊文献+
共找到8篇文章
< 1 >
每页显示 20 50 100
MMCSD:Multi-Modal Knowledge Graph Completion Based on Super-Resolution and Detailed Description Generation
1
作者 Huansha Wang Ruiyang Huang +2 位作者 Qinrang Liu Shaomei Li Jianpeng Zhang 《Computers, Materials & Continua》 2025年第4期761-783,共23页
Multi-modal knowledge graph completion(MMKGC)aims to complete missing entities or relations in multi-modal knowledge graphs,thereby discovering more previously unknown triples.Due to the continuous growth of data and ... Multi-modal knowledge graph completion(MMKGC)aims to complete missing entities or relations in multi-modal knowledge graphs,thereby discovering more previously unknown triples.Due to the continuous growth of data and knowledge and the limitations of data sources,the visual knowledge within the knowledge graphs is generally of low quality,and some entities suffer from the issue of missing visual modality.Nevertheless,previous studies of MMKGC have primarily focused on how to facilitate modality interaction and fusion while neglecting the problems of low modality quality and modality missing.In this case,mainstream MMKGC models only use pre-trained visual encoders to extract features and transfer the semantic information to the joint embeddings through modal fusion,which inevitably suffers from problems such as error propagation and increased uncertainty.To address these problems,we propose a Multi-modal knowledge graph Completion model based on Super-resolution and Detailed Description Generation(MMCSD).Specifically,we leverage a pre-trained residual network to enhance the resolution and improve the quality of the visual modality.Moreover,we design multi-level visual semantic extraction and entity description generation,thereby further extracting entity semantics from structural triples and visual images.Meanwhile,we train a variational multi-modal auto-encoder and utilize a pre-trained multi-modal language model to complement the missing visual features.We conducted experiments on FB15K-237 and DB13K,and the results showed that MMCSD can effectively perform MMKGC and achieve state-of-the-art performance. 展开更多
关键词 multi-modal knowledge graph knowledge graph completion multi-modal fusion
在线阅读 下载PDF
Multi-modal knowledge graph inference via media convergence and logic rule
2
作者 Feng Lin Dongmei Li +5 位作者 Wenbin Zhang Dongsheng Shi Yuanzhou Jiao Qianzhong Chen Yiying Lin Wentao Zhu 《CAAI Transactions on Intelligence Technology》 SCIE EI 2024年第1期211-221,共11页
Media convergence works by processing information from different modalities and applying them to different domains.It is difficult for the conventional knowledge graph to utilise multi-media features because the intro... Media convergence works by processing information from different modalities and applying them to different domains.It is difficult for the conventional knowledge graph to utilise multi-media features because the introduction of a large amount of information from other modalities reduces the effectiveness of representation learning and makes knowledge graph inference less effective.To address the issue,an inference method based on Media Convergence and Rule-guided Joint Inference model(MCRJI)has been pro-posed.The authors not only converge multi-media features of entities but also introduce logic rules to improve the accuracy and interpretability of link prediction.First,a multi-headed self-attention approach is used to obtain the attention of different media features of entities during semantic synthesis.Second,logic rules of different lengths are mined from knowledge graph to learn new entity representations.Finally,knowledge graph inference is performed based on representing entities that converge multi-media features.Numerous experimental results show that MCRJI outperforms other advanced baselines in using multi-media features and knowledge graph inference,demonstrating that MCRJI provides an excellent approach for knowledge graph inference with converged multi-media features. 展开更多
关键词 logic rule media convergence multi-modal knowledge graph inference representation learning
在线阅读 下载PDF
DCRL-KG: Distributed Multi-Modal Knowledge Graph Retrieval Platform Based on Collaborative Representation Learning
3
作者 Leilei Li Yansheng Fu +6 位作者 Dongjie Zhu Xiaofang Li Yundong Sun Jianrui Ding Mingrui Wu Ning Cao Russell Higgs 《Intelligent Automation & Soft Computing》 SCIE 2023年第6期3295-3307,共13页
The knowledge graph with relational abundant information has been widely used as the basic data support for the retrieval platforms.Image and text descriptions added to the knowledge graph enrich the node information,... The knowledge graph with relational abundant information has been widely used as the basic data support for the retrieval platforms.Image and text descriptions added to the knowledge graph enrich the node information,which accounts for the advantage of the multi-modal knowledge graph.In the field of cross-modal retrieval platforms,multi-modal knowledge graphs can help to improve retrieval accuracy and efficiency because of the abundant relational infor-mation provided by knowledge graphs.The representation learning method is sig-nificant to the application of multi-modal knowledge graphs.This paper proposes a distributed collaborative vector retrieval platform(DCRL-KG)using the multi-modal knowledge graph VisualSem as the foundation to achieve efficient and high-precision multimodal data retrieval.Firstly,use distributed technology to classify and store the data in the knowledge graph to improve retrieval efficiency.Secondly,this paper uses BabelNet to expand the knowledge graph through multi-ple filtering processes and increase the diversification of information.Finally,this paper builds a variety of retrieval models to achieve the fusion of retrieval results through linear combination methods to achieve high-precision language retrieval and image retrieval.The paper uses sentence retrieval and image retrieval experi-ments to prove that the platform can optimize the storage structure of the multi-modal knowledge graph and have good performance in multi-modal space. 展开更多
关键词 multi-modal retrieval distributed storage knowledge graph
在线阅读 下载PDF
Multi-Modal Pre-Synergistic Fusion Entity Alignment Based on Mutual Information Strategy Optimization
4
作者 Huayu Li Xinxin Chen +3 位作者 Lizhuang Tan Konstantin I.Kostromitin Athanasios V.Vasilakos Peiying Zhang 《Computers, Materials & Continua》 2025年第11期4133-4153,共21页
To address the challenge of missing modal information in entity alignment and to mitigate information loss or bias arising frommodal heterogeneity during fusion,while also capturing shared information acrossmodalities... To address the challenge of missing modal information in entity alignment and to mitigate information loss or bias arising frommodal heterogeneity during fusion,while also capturing shared information acrossmodalities,this paper proposes a Multi-modal Pre-synergistic Entity Alignmentmodel based on Cross-modalMutual Information Strategy Optimization(MPSEA).The model first employs independent encoders to process multi-modal features,including text,images,and numerical values.Next,a multi-modal pre-synergistic fusion mechanism integrates graph structural and visual modal features into the textual modality as preparatory information.This pre-fusion strategy enables unified perception of heterogeneous modalities at the model’s initial stage,reducing discrepancies during the fusion process.Finally,using cross-modal deep perception reinforcement learning,the model achieves adaptive multilevel feature fusion between modalities,supporting learningmore effective alignment strategies.Extensive experiments on multiple public datasets show that the MPSEA method achieves gains of up to 7% in Hits@1 and 8.2% in MRR on the FBDB15K dataset,and up to 9.1% in Hits@1 and 7.7% in MRR on the FBYG15K dataset,compared to existing state-of-the-art methods.These results confirm the effectiveness of the proposed model. 展开更多
关键词 knowledge graph multi-modal entity alignment feature fusion pre-synergistic fusion
在线阅读 下载PDF
多模态知识图谱补全方法综述
5
作者 王雪 张丽萍 +2 位作者 闫盛 李娜 张学飞 《计算机应用》 北大核心 2026年第2期341-353,共13页
传统知识图谱(KG)虽然为网络中的信息提供了一种统一的且机器可理解的表示方式,但在处理多模态应用时逐渐暴露出局限性。为了应对这些局限性,研究者提出多模态知识图谱(MMKG)作为有效解决方案。然而,KG引入多模态数据后广泛存在模态融... 传统知识图谱(KG)虽然为网络中的信息提供了一种统一的且机器可理解的表示方式,但在处理多模态应用时逐渐暴露出局限性。为了应对这些局限性,研究者提出多模态知识图谱(MMKG)作为有效解决方案。然而,KG引入多模态数据后广泛存在模态融合不充分和推理困难的问题,这制约了MMKG的应用和发展。而多模态知识图谱补全(MMKGC)技术不仅能够在构建阶段充分融合跨模态信息,还能够在构建完成阶段预测缺失的链接,从而解决在模态融合和推理时遇到的问题。因此,对MMKG方法进行综述。首先,详尽阐述MMKGC的基本概述以及常用的基准数据集和评价指标;其次,将现有方法分为针对MMKG构建阶段的融合任务和构建完成阶段的推理任务,前者聚焦于关键技术如实体对齐和实体链接,后者则涵盖关系推理、信息缺失补全及多模态扩展这3类技术;再次,详细介绍了各类MMKGC方法,并分析它们的特点;最后,分析MMKGC方法面临的问题与挑战并总结前面的内容。 展开更多
关键词 多模态数据 多模态知识图谱 多模态知识图谱补全 实体对齐 关系推理
在线阅读 下载PDF
Prospects on mangrove knowledge services in the smart era:From plant atlas to knowledge graphs
6
作者 Zhi-Wei HOU Wenlong JING +3 位作者 Cheng-Zhi QIN Ji YANG Qing XIA Xiaoling YIN 《Science China Earth Sciences》 2025年第1期111-127,共17页
Mangroves are crucial to the ecological security of the Earth and human well-being.Their management,conservation,and restoration are of great importance and necessitate the support of spatio-temporal information and m... Mangroves are crucial to the ecological security of the Earth and human well-being.Their management,conservation,and restoration are of great importance and necessitate the support of spatio-temporal information and multidisciplinary knowledge such as biology and ecology.Traditional knowledge services such as plant atlas provide illustrated textual knowledge of mangroves.However,this kind of service is oriented to information retrieval and is incapable of effectively mining and utilizing fragmented knowledge from multi-source heterogeneous data,facing the problem of“massive data,rare knowledge”.Knowledge graphs are capable of extracting,organizing,and fusing the knowledge contained in massive data into semantic networks that can be understood and computed by computers.They provide a solution for the realization of intelligent knowledge services.Focusing on the urgent need for mangrove knowledge acquisition,formal representation,and intelligent services,this paper proposes a research prospect on mangrove knowledge graphs and knowledge services.We first analyze the similarities and differences between various domain-specific concepts of Tupu.On this basis,we define the mangrove knowledge graph as a large-scale knowledge base that integrates multi-disciplinary knowledge and spatio-temporal information with mangrove ecosystems as the core.Then,we propose a research framework for mangrove knowledge services that can realize the transformation from multi-modal data to intelligent knowledge services,including multiple research levels such as ubiquitous data sensing and aggregation,knowledge organization and graph construction,and intelligent mangrove knowledge services.Subsequently,the methods and workflow for constructing mangrove knowledge graphs are introduced.Finally,we discuss the challenges and possible future directions of mangrove knowledge services in the smart era,including the construction of a mangrove knowledge system that integrates the domain-specific characteristics and spatio-temporal features of mangroves,the exploration of knowledge extraction and fusion methods supported by large language models,and the development of intelligent knowledge applications for typical scenarios. 展开更多
关键词 MANGROVE Tupu knowledge graph multi-modal knowledge service
原文传递
Visual Entity Linking via Multi-modal Learning 被引量:4
7
作者 Qiushuo Zheng Hao Wen +1 位作者 Meng Wang Guilin Qi 《Data Intelligence》 EI 2022年第1期1-19,共19页
Existing visual scene understanding methods mainly focus on identifying coarse-grained concepts about the visual objects and their relationships,largely neglecting fine-grained scene understanding.In fact,many data-dr... Existing visual scene understanding methods mainly focus on identifying coarse-grained concepts about the visual objects and their relationships,largely neglecting fine-grained scene understanding.In fact,many data-driven applications on the Web(e.g.,news-reading and e-shopping)require accurate recognition of much less coarse concepts as entities and proper linking them to a knowledge graph(KG),which can take their performance to the next level.In light of this,in this paper,we identify a new research task:visual entity linking for fine-grained scene understanding.To accomplish the task,we first extract features of candidate entities from different modalities,i.e.,visual features,textual features,and KG features.Then,we design a deep modal-attention neural network-based learning-to-rank method which aggregates all features and maps visual objects to the entities in KG.Extensive experimental results on the newly constructed dataset show that our proposed method is effective as it significantly improves the accuracy performance from 66.46%to 83.16%compared with baselines. 展开更多
关键词 knowledge graph multi-modal learning Entity linking Learning to rank knowledge graph representation
原文传递
Faster Zero-shot Multi-modal Entity Linking via Visual-Linguistic Representation 被引量:2
8
作者 Qiushuo Zheng Hao Wen +2 位作者 Meng Wang Guilin Qi Chaoyu Bai 《Data Intelligence》 EI 2022年第3期493-508,共16页
Multi-modal entity linking plays a crucial role in a wide range of knowledge-based modal-fusion tasks, i.e., multi-modal retrieval and multi-modal event extraction. We introduce the new ZEro-shot Multi-modal Entity Li... Multi-modal entity linking plays a crucial role in a wide range of knowledge-based modal-fusion tasks, i.e., multi-modal retrieval and multi-modal event extraction. We introduce the new ZEro-shot Multi-modal Entity Linking(ZEMEL) task, the format is similar to multi-modal entity linking, but multi-modal mentions are linked to unseen entities in the knowledge graph, and the purpose of zero-shot setting is to realize robust linking in highly specialized domains. Simultaneously, the inference efficiency of existing models is low when there are many candidate entities. On this account, we propose a novel model that leverages visuallinguistic representation through the co-attentional mechanism to deal with the ZEMEL task, considering the trade-off between performance and efficiency of the model. We also build a dataset named ZEMELD for the new task, which contains multi-modal data resources collected from Wikipedia, and we annotate the entities as ground truth. Extensive experimental results on the dataset show that our proposed model is effective as it significantly improves the precision from 68.93% to 82.62% comparing with baselines in the ZEMEL task. 展开更多
关键词 knowledge graph multi-modal Learning Poly Encoders
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部