• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
成果搜索

author:

Wang, Jingbin (Wang, Jingbin.) [1] | Ke, XiFan (Ke, XiFan.) [2] | Zhang, FuYuan (Zhang, FuYuan.) [3] | Wu, YuWei (Wu, YuWei.) [4] | Zhang, SiRui (Zhang, SiRui.) [5] | Guo, Kun (Guo, Kun.) [6]

Indexed by:

EI

Abstract:

The extrapolation task in the temporal knowledge graph has received increasing attention from scholars due to its wide range of practical application scenarios. At present, recurrent neural networks are currently widely used in temporal knowledge graph completion techniques. These networks are employed to depict the sequential pattern of entities and relations. However, as the sequence lengthens, some critical early information may become diluted. Prediction errors ensue in the completion task as a result. Furthermore, it is observed that existing temporal knowledge graph completion methods fail to account for the topological structure of relations, which leads to relation representations with essentially little distinction across different timestamps. In order to tackle the previously mentioned concern, our research introduces a Temporal Knowledge Graph Completion Method utilizing Sequence-Focus Patterns Representation Learning (SFP). This method contains two patterns: the Focus pattern and the Sequential pattern. In the SFP model, we developed a novel graph attention network called ConvGAT. This network efficiently distinguishes and extracts complex relation information, thereby enhancing the accuracy of entity representations that are aggregated in the Focus pattern and Sequential pattern. Furthermore we proposed RelGAT, a graph attention network that simulates the topological structure of relations. This enhances the precision of relation representations and facilitates the differentiation between relation embeddings generated at various timestamps in the Focus pattern. Utilizing a time-aware attention mechanism, the Focus pattern extracts vital information at particular timestamps in order to amplify the data that the Sequential pattern dilutes. On five distinct benchmark datasets, SFP significantly outperforms the baseline, according to a comprehensive series of experiments. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2025.

Keyword:

Graph embeddings Graph neural networks Graph theory Knowledge graph Recurrent neural networks

Community:

  • [ 1 ] [Wang, Jingbin]College of Computer and Data Science, Fuzhou University, Fujian; 350108, China
  • [ 2 ] [Ke, XiFan]College of Computer and Data Science, Fuzhou University, Fujian; 350108, China
  • [ 3 ] [Zhang, FuYuan]College of Computer and Data Science, Fuzhou University, Fujian; 350108, China
  • [ 4 ] [Wu, YuWei]College of Computer and Data Science, Fuzhou University, Fujian; 350108, China
  • [ 5 ] [Zhang, SiRui]College of Computer and Data Science, Fuzhou University, Fujian; 350108, China
  • [ 6 ] [Guo, Kun]College of Computer and Data Science, Fuzhou University, Fujian; 350108, China

Reprint 's Address:

  • [guo, kun]college of computer and data science, fuzhou university, fujian; 350108, china;;

Show more details

Related Keywords:

Related Article:

Source :

Applied Intelligence

ISSN: 0924-669X

Year: 2025

Issue: 6

Volume: 55

3 . 4 0 0

JCR@2023

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count:

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 0

Affiliated Colleges:

Online/Total:55/10057265
Address:FZU Library(No.2 Xuyuan Road, Fuzhou, Fujian, PRC Post Code:350116) Contact Us:0591-22865326
Copyright:FZU Library Technical Support:Beijing Aegean Software Co., Ltd. 闽ICP备05005463号-1