DG‐based SPO tuple recognition using self‐attention M‐Bi‐LSTM

This study proposes a dependency grammar‐based self‐attention multilayered bidirectional long short‐term memory (DG‐M‐Bi‐LSTM) model for subject–predicate–object (SPO) tuple recognition from natural language (NL) sentences. To add recent knowledge to the knowledge base autonomously, it is essential...

Full description

Saved in:
Bibliographic Details
Published in:ETRI journal Vol. 44; no. 3; pp. 438 - 449
Main Author: Jung, Joon‐young
Format: Journal Article
Language:English
Published: Electronics and Telecommunications Research Institute (ETRI) 01-06-2022
한국전자통신연구원
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This study proposes a dependency grammar‐based self‐attention multilayered bidirectional long short‐term memory (DG‐M‐Bi‐LSTM) model for subject–predicate–object (SPO) tuple recognition from natural language (NL) sentences. To add recent knowledge to the knowledge base autonomously, it is essential to extract knowledge from numerous NL data. Therefore, this study proposes a high‐accuracy SPO tuple recognition model that requires a small amount of learning data to extract knowledge from NL sentences. The accuracy of SPO tuple recognition using DG‐M‐Bi‐LSTM is compared with that using NL‐based self‐attention multilayered bidirectional LSTM, DG‐based bidirectional encoder representations from transformers (BERT), and NL‐based BERT to evaluate its effectiveness. The DG‐M‐Bi‐LSTM model achieves the best results in terms of recognition accuracy for extracting SPO tuples from NL sentences even if it has fewer deep neural network (DNN) parameters than BERT. In particular, its accuracy is better than that of BERT when the learning data are limited. Additionally, its pretrained DNN parameters can be applied to other domains because it learns the structural relations in NL sentences.
Bibliography:Funding information
Electronics and Telecommunications Research Institute, Grant/Award Number: 21ZS1100
https://doi.org/10.4218/etrij.2020-0460
ISSN:1225-6463
2233-7326
DOI:10.4218/etrij.2020-0460