دانلود مقاله ISI انگلیسی شماره 156671
ترجمه فارسی عنوان مقاله

چارچوب برای پیش آموزش بخش های پنهان محدوده تصادفی شرطی و گسترش آن به شبکه های حافظه طولانی مدت

عنوان انگلیسی
A Framework for pre-training hidden-unit conditional random fields and its extension to long short term memory networks
کد مقاله سال انتشار تعداد صفحات مقاله انگلیسی
156671 2017 16 صفحه PDF
منبع

Publisher : Elsevier - Science Direct (الزویر - ساینس دایرکت)

Journal : Computer Speech & Language, Volume 46, November 2017, Pages 311-326

پیش نمایش مقاله
پیش نمایش مقاله  چارچوب برای پیش آموزش بخش های پنهان محدوده تصادفی شرطی و گسترش آن به شبکه های حافظه طولانی مدت

چکیده انگلیسی

In this paper, we introduce a simple unsupervised framework for pre-training hidden-unit conditional random fields (HUCRFs), i.e., learning initial parameter estimates for HUCRFs prior to supervised training.Our framework exploits the model structure of HUCRFs to make effective use of unlabeled data from the same domain or labeled data from a different domain. The key idea is to use the separation of HUCRF parameters between observations and labels: this allows us to pre-train observation parameters independently of label parameters. Pre-training is achieved by creating pseudo-labels from such resources. In the case of unlabeled data, we cluster observations and use the resulting clusters as pseudo-labels. Observation parameters can be trained on these resources and then transferred to initialize the supervised training process on the target labeled data. Experiments on various sequence labeling tasks demonstrate that the proposed pre-training method consistently yields significant improvement in performance. The core idea could be extended to other learning techniques including deep learning. We applied the proposed technique to recurrent neural networks (RNN) with long short term memory (LSTM) architecture and obtained similar gains.