Biao Hu
2025
Partial Order-centered Hyperbolic Representation Learning for Few-shot Relation Extraction
Biao Hu
|
Zhen Huang
|
Minghao Hu
|
Pinglv Yang
|
Peng Qiao
|
Yong Dou
|
Zhilin Wang
Proceedings of the 31st International Conference on Computational Linguistics
Prototype network-based methods have made substantial progress in few-shot relation extraction (FSRE) by enhancing relation prototypes with relation descriptions. However, the distribution of relations and instances in distinct representation spaces isolates the constraints of relations on instances, making relation prototypes biased. In this paper, we propose an end-to-end partial order-centered hyperbolic representation learning (PO-HRL) framework, which imposes the constraints of relations on instances by modeling partial order in hyperbolic space, so as to effectively learn the distribution of instance representations. Specifically, we develop the hyperbolic supervised contrastive learning based on Lorentzian cosine similarity to align representations of relations and instances, and model the partial order by constraining instances to reside within the Lorentzian entailment cone of their respective relation. Experiments on three benchmark datasets show that PO-HRL outperforms the strong baselines, especially in 1-shot settings lacking relation descriptions.
2022
Adaptive Threshold Selective Self-Attention for Chinese NER
Biao Hu
|
Zhen Huang
|
Minghao Hu
|
Ziwen Zhang
|
Yong Dou
Proceedings of the 29th International Conference on Computational Linguistics
Recently, Transformer has achieved great success in Chinese named entity recognition (NER) owing to its good parallelism and ability to model long-range dependencies, which utilizes self-attention to encode context. However, the fully connected way of self-attention may scatter the attention distribution and allow some irrelevant character information to be integrated, leading to entity boundaries being misidentified. In this paper, we propose a data-driven Adaptive Threshold Selective Self-Attention (ATSSA) mechanism that aims to dynamically select the most relevant characters to enhance the Transformer architecture for Chinese NER. In ATSSA, the attention score threshold of each query is automatically generated, and characters with attention score higher than the threshold are selected by the query while others are discarded, so as to address irrelevant attention integration. Experiments on four benchmark Chinese NER datasets show that the proposed ATSSA brings 1.68 average F1 score improvements to the baseline model and achieves state-of-the-art performance.
Search
Fix data
Co-authors
- Yong Dou 2
- Minghao Hu 2
- Zhen Huang 2
- Peng Qiao 1
- Zhilin Wang 1
- show all...