Peng Ye
2024
Prompt-fused Framework for Inductive Logical Query Answering
Zezhong Xu
|
Wen Zhang
|
Peng Ye
|
Lei Liang
|
Huajun Chen
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Answering logical queries on knowledge graphs (KG) poses a significant challenge for machine reasoning. The primary obstacle in this task stems from the inherent incompleteness of KGs. Existing research has predominantly focused on addressing the issue of missing edges in KGs, thereby neglecting another aspect of incompleteness: the emergence of new entities. Furthermore, most of the existing methods tend to reason over each logical operator separately, rather than comprehensively analyzing the query as a whole during the reasoning process. In this paper, we propose a query-aware prompt-fused framework named Pro-QE, which could incorporate existing query embedding methods and address the embedding of emerging entities through contextual information aggregation. Additionally, a query prompt, which is generated by encoding the symbolic query, is introduced to gather information relevant to the query from a holistic perspective. To evaluate the efficacy of our model in the inductive setting, we introduce two new challenging benchmarks. Experimental results demonstrate that our model successfully handles the issue of unseen entities in logical queries. Furthermore, the ablation study confirms the efficacy of the aggregator and prompt components.
2022
Ruleformer: Context-aware Rule Mining over Knowledge Graph
Zezhong Xu
|
Peng Ye
|
Hui Chen
|
Meng Zhao
|
Huajun Chen
|
Wen Zhang
Proceedings of the 29th International Conference on Computational Linguistics
Rule mining is an effective approach for reasoning over knowledge graph (KG). Existing works mainly concentrate on mining rules. However, there might be several rules that could be applied for reasoning for one relation, and how to select appropriate rules for completion of different triples has not been discussed. In this paper, we propose to take the context information into consideration, which helps select suitable rules for the inference tasks. Based on this idea, we propose a transformer-based rule mining approach, Ruleformer. It consists of two blocks: 1) an encoder extracting the context information from subgraph of head entities with modified attention mechanism, and 2) a decoder which aggregates the subgraph information from the encoder output and generates the probability of relations for each step of reasoning. The basic idea behind Ruleformer is regarding rule mining process as a sequence to sequence task. To make the subgraph a sequence input to the encoder and retain the graph structure, we devise a relational attention mechanism in Transformer. The experiment results show the necessity of considering these information in rule mining task and the effectiveness of our model.
2012
A Random Forest System Combination Approach for Error Detection in Digital Dictionaries
Michael Bloodgood
|
Peng Ye
|
Paul Rodrigues
|
David Zajic
|
David Doermann
Proceedings of the Workshop on Innovative Hybrid Approaches to the Processing of Textual Data
Search
Co-authors
- Zezhong Xu 2
- Huajun Chen 2
- Wen Zhang 2
- Michael Bloodgood 1
- Paul Rodrigues 1
- show all...