메뉴 건너뛰기
.. 내서재 .. 알림
소속 기관/학교 인증
인증하면 논문, 학술자료 등을  무료로 열람할 수 있어요.
한국대학교, 누리자동차, 시립도서관 등 나의 기관을 확인해보세요
(국내 대학 90% 이상 구독 중)
로그인 회원가입 고객센터 ENG
주제분류

추천
검색

논문 기본 정보

자료유형
학위논문
저자정보

김현 (포항공과대학교, 포항공과대학교 일반대학원)

지도교수
이종혁
발행연도
2018
저작권
포항공과대학교 논문은 저작권에 의해 보호받습니다.

이용수0

표지
AI에게 요청하기
추천
검색

이 논문의 연구 히스토리 (2)

초록· 키워드

오류제보하기
Translation quality estimation is an important issue in machine translation, which has been attracting increasing interest from researchers, aiming at finding a good estimator for the quality of machine translation output.
The common approach for quality estimation is to treat the problem as a supervised regression/classification task using a quality-annotated noisy parallel corpus, called quality estimation data, as training data. However, the available size of quality estimation data remains small, due to the too expensive cost of creating such data. In addition, most conventional quality estimation approaches rely on manually designed features to model nonlinear relationships between feature vectors and corresponding quality labels.
To overcome these problems, this thesis proposes a novel neural network architecture for quality estimation task -- called the Predictor-Estimator -- that considers word prediction as an additional pre-task. The major component of the proposed neural architecture is a word prediction model based on a modified neural machine translation model -- a probabilistic model for predicting a target word conditioned on all the other source and target contexts. The underlying assumption is that the word prediction model is highly related to quality estimation models, and is therefore able to transfer useful knowledge to quality estimation tasks. Our proposed quality estimation method trains the following two types of neural models: 1) Predictor: Neural word prediction model trained from parallel corpora and 2) Estimator: Neural quality estimation model trained from quality estimation data. To transfer word prediction task to quality estimation task, we generate quality estimation feature vectors from the word prediction model and feed them into the quality estimation model.
Our proposed Predictor-Estimator has a potential problem: how to learn efficiently two-stage Predictor-Estimator model, because the Predictor-Estimator model is comprised of stacked main tasks (word prediction and quality estimation tasks) and parallel multi-level sub-tasks (sentence, word and phrase-level quality estimation sub-tasks). This thesis also proposes an efficient multi-task learning method for Predictor-Estimator model.
The experimental results on WMT17 quality estimation datasets indicated that our proposed method has potential as it achieves state-of-the-art performances for all subtasks (sentence, word, and phrase levels).

목차

I. Introduction
1.1 Problem statement
1.2 Issues and contributions
1.3 Thesis outline
II. Background
2.1 Translation quality estimation task
2.2 Evaluation measures
2.3 Notation
III. Related Work
3.1 Feature engineering based approach
3.2 Neural network based approach
IV. Predictor-Estimator Architecture
4.1 Overview
4.2 Predictor: Neural word prediction model
4.3 Extraction of quality estimation feature vectors
4.4 Estimator: Neural QE model
4.4.1 Sentence level
4.4.2 Word level
4.4.3 Phrase level
4.5 Experimentation
4.5.1 Experimental settings
4.5.2 QE Methods
4.5.3 Comparison with QEFV types
4.5.4 Effects of word predictor: prediction-based QEFVs vs. wordbembedding-based QEFVs
4.5.5 Influence of word predictor on quality estimator
4.6 Conclusion
V. Improving Predictor-Estimator using Multi-task Learning and Ensemble Methods
5.1 Motivation
5.2 Improvements on Predictor-Estimator
5.2.1 Weight updates
5.2.2 Multimodality
5.3 Multi-task learning methods for Predictor-Estimator
5.3.1 (Single-level) Stack propagation
5.3.2 Multi-level sub-task learning
5.3.3 Combination: Multi-level stack propagation
5.4 Experimentation
5.4.1 Experimental settings
5.4.2 Comparison with learning methods
5.4.3 Ensembles of multiple instances
5.5 Conclusion
VI. Summary of Thesis and Future Works
6.1 Summary
6.2 Future directions
Summary (in Korean)
References

최근 본 자료

전체보기

댓글(0)

0