Learning contextual representations of citations via graph transformer
- Authors
- Jeon, H.-J.; Choi, G.-S.; Cho, S.-Y.; Lee, H.; Ko, H.Y.; Jung, J.J.; Lee, O.J.; Yi, M.-Y.
- Issue Date
- Oct-2021
- Publisher
- CEUR-WS
- Keywords
- Citation Context; Citation Network; Graph Transformer; Network Embedding; Positional Embedding
- Citation
- CEUR Workshop Proceedings, v.3026, pp 150 - 158
- Pages
- 9
- Journal Title
- CEUR Workshop Proceedings
- Volume
- 3026
- Start Page
- 150
- End Page
- 158
- URI
- https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/52794
- ISSN
- 1613-0073
- Abstract
- This study aims at representing the citation based on the citation context extracted from the citation network. Researchers cite papers for various purposes to describe their arguments in a logical structure. Thus, citations have different roles depending on what structure they are cited in the paper. In this paper, we first present a definition of the citation context and initialize the embedding vector based on the citation order and location. Then, based on the graph transformer model, we learn contextual citation embeddings. To represent citation context, we consider the following three parts: (i) textual features of paper, (ii) positional features of the citation context, and (iii) structural features of the citation network by applying the self-attention mechanism. © 2021 CEUR-WS. All rights reserved.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Software > School of Computer Science and Engineering > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/52794)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.