DG‐based SPO tuple recognition using self‐attention M‐Bi‐LSTM
This study proposes a dependency grammar‐based self‐attention multilayered bidirectional long short‐term memory (DG‐M‐Bi‐LSTM) model for subject–predicate–object (SPO) tuple recognition from natural language (NL) sentences. To add recent knowledge to the knowledge base autonomously, it is essential...
Saved in:
Published in: | ETRI journal Vol. 44; no. 3; pp. 438 - 449 |
---|---|
Main Author: | |
Format: | Journal Article |
Language: | English |
Published: |
Electronics and Telecommunications Research Institute (ETRI)
01-06-2022
한국전자통신연구원 |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | This study proposes a dependency grammar‐based self‐attention multilayered bidirectional long short‐term memory (DG‐M‐Bi‐LSTM) model for subject–predicate–object (SPO) tuple recognition from natural language (NL) sentences. To add recent knowledge to the knowledge base autonomously, it is essential to extract knowledge from numerous NL data. Therefore, this study proposes a high‐accuracy SPO tuple recognition model that requires a small amount of learning data to extract knowledge from NL sentences. The accuracy of SPO tuple recognition using DG‐M‐Bi‐LSTM is compared with that using NL‐based self‐attention multilayered bidirectional LSTM, DG‐based bidirectional encoder representations from transformers (BERT), and NL‐based BERT to evaluate its effectiveness. The DG‐M‐Bi‐LSTM model achieves the best results in terms of recognition accuracy for extracting SPO tuples from NL sentences even if it has fewer deep neural network (DNN) parameters than BERT. In particular, its accuracy is better than that of BERT when the learning data are limited. Additionally, its pretrained DNN parameters can be applied to other domains because it learns the structural relations in NL sentences. |
---|---|
Bibliography: | Funding information Electronics and Telecommunications Research Institute, Grant/Award Number: 21ZS1100 https://doi.org/10.4218/etrij.2020-0460 |
ISSN: | 1225-6463 2233-7326 |
DOI: | 10.4218/etrij.2020-0460 |