TY - JOUR
T1 - Self-attention-based conditional random fields latent variables model for sequence labeling
AU - Shao, Yinan
AU - Lin, Jerry Chun Wei
AU - Srivastava, Gautam
AU - Jolfaei, Alireza
AU - Guo, Dongdong
AU - Hu, Yi
PY - 2021/5
Y1 - 2021/5
N2 - To process data like text and speech, Natural Language Processing (NLP) is a valuable tool. As on of NLP's upstream tasks, sequence labeling is a vital part of NLP through techniques like text classification, machine translation, and sentiment analysis. In this paper, our focus is on sequence labeling where we assign semantic labels within input sequences. We present two novel frameworks, namely SA-CRFLV-I and SA-CRFLV-II, that use latent variables within random fields. These frameworks make use of an encoding schema in the form of a latent variable to be able to capture the latent structure in the observed data. SA-CRFLV-I shows the best performance at the sentence level whereas SA-CRFLV-II works best at the word level. In our in-depth experimental results, we compare our frameworks with 4 well-known sequence prediction methodologies which include NER, reference parsing, chunking as well as POS tagging. The proposed frameworks are shown to have better performance in terms of many well-known metrics.
AB - To process data like text and speech, Natural Language Processing (NLP) is a valuable tool. As on of NLP's upstream tasks, sequence labeling is a vital part of NLP through techniques like text classification, machine translation, and sentiment analysis. In this paper, our focus is on sequence labeling where we assign semantic labels within input sequences. We present two novel frameworks, namely SA-CRFLV-I and SA-CRFLV-II, that use latent variables within random fields. These frameworks make use of an encoding schema in the form of a latent variable to be able to capture the latent structure in the observed data. SA-CRFLV-I shows the best performance at the sentence level whereas SA-CRFLV-II works best at the word level. In our in-depth experimental results, we compare our frameworks with 4 well-known sequence prediction methodologies which include NER, reference parsing, chunking as well as POS tagging. The proposed frameworks are shown to have better performance in terms of many well-known metrics.
KW - Big data
KW - Encoding schema
KW - Latent CRF
KW - Natural language processing
KW - Sequence labeling
KW - VQA
UR - http://www.scopus.com/inward/record.url?scp=85101381633&partnerID=8YFLogxK
U2 - 10.1016/j.patrec.2021.02.008
DO - 10.1016/j.patrec.2021.02.008
M3 - Article
AN - SCOPUS:85101381633
SN - 0167-8655
VL - 145
SP - 157
EP - 164
JO - Pattern Recognition Letters
JF - Pattern Recognition Letters
ER -