Relation Classification via Recurrent Neural Network with Attention and Tensor Layers
Relation classification is a crucial component in many Natural Language Processing (NLP) systems. In this paper, we propose a novel bidirectional recurrent neural network architecture (using Long Short-Term Memory, LSTM, cells) for relation classification, with an attention layer for organizing the...
Saved in:
Main Authors: | Runyan Zhang, Fanrong Meng, Yong Zhou, Bing Liu |
---|---|
Format: | Article |
Language: | English |
Published: |
Tsinghua University Press
2018-09-01
|
Series: | Big Data Mining and Analytics |
Subjects: | |
Online Access: | https://www.sciopen.com/article/10.26599/BDMA.2018.9020022 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
A Topological Approach to Enhancing Consistency in Machine Learning via Recurrent Neural Networks
by: Muhammed Adil Yatkin, et al.
Published: (2025-01-01) -
Survey of FPGA based recurrent neural network accelerator
by: Chen GAO, et al.
Published: (2019-08-01) -
Model of the malicious traffic classification based on hypergraph neural network
by: Wenbo ZHAO, et al.
Published: (2023-10-01) -
Secure Image Reconstruction using Deep Learning-based Autoencoder with Integrated Encryption Layers
by: Wurood Abd Ali
Published: (2024-12-01) -
Deep BiLSTM Attention Model for Spatial and Temporal Anomaly Detection in Video Surveillance
by: Sarfaraz Natha, et al.
Published: (2025-01-01)