سیویلیکا را در شبکه های اجتماعی دنبال نمایید.

Neural Architecture for Persian Named Entity Recognition

Publish Year: 1397
Type: Conference paper
Language: English
View: 525
این Paper فقط به صورت چکیده توسط دبیرخانه ارسال شده است و فایل کامل قابل دریافت نیست. برای یافتن Papers دارای فایل کامل، از بخش [جستجوی مقالات فارسی] اقدام فرمایید.

نسخه کامل این Paper ارائه نشده است و در دسترس نمی باشد

Export:

Link to this Paper:

Document National Code:

SPIS04_039

Index date: 6 May 2019

Neural Architecture for Persian Named Entity Recognition abstract

Named entity recognition task is challenging problem specially for languages with low resources of annotated corpora such as Persian. In this paper, we use deep learning approaches to obtain high performance results without using any specific hand-made features and gazetteers. In the recent years, LSTMs have been observed as the most effective solution for sequence prediction problems like NER. For this purpose, we use LSTM-CRF approach for Persian NER and combine with the pre-trained word embedding. For achieving language independent vector representation, we use character-based representation. By using bidirectional LSTM, word and character level features automatically detect and the need for most of hand-engineered features eliminates. This model obtains 86.55% F1 score that is the highest score in compare to previous results in Persian language. We also apply this architecture for dataset in [2] and achieve 84.23% F1 score, published result for this dataset is showed 77.45%F1 score.

Neural Architecture for Persian Named Entity Recognition authors